Arcee-ai

Arcee AI: Virtuoso Large

Virtuoso‑Large is Arcee's top‑tier general‑purpose LLM at 72 B parameters, tuned to tackle cross‑domain reasoning, creative writing and enterprise QA. Unlike many 70 B peers, it retains the 128 k context inherited from Qwen 2.5, letting it ingest books, codebases or financial filings wholesale. Training blended DeepSeek R1 distillation, multi‑epoch supervised fine‑tuning and a final DPO/RLHF alignment stage, yielding strong performance on BIG‑Bench‑Hard, GSM‑8K and long‑context Needle‑In‑Haystack tests. Enterprises use Virtuoso‑Large as the "fallback" brain in Conductor pipelines when other SLMs flag low confidence. Despite its size, aggressive KV‑cache optimizations keep first‑token latency in the low‑second range on 8× H100 nodes, making it a practical production‑grade powerhouse.

Context Window
131K tokens
Parameters
72B
Input Price
$0.75/1M
Output Price
$1.20/1M
Price Tier
standard
Provider
Arcee-ai

How to Use Arcee AI: Virtuoso Large

With CoreAI, you can start chatting with Arcee AI: Virtuoso Large instantly — no separate subscription needed. CoreAI bundles access to Arcee AI: Virtuoso Large along with 300+ other AI models from Arcee-ai and other providers like OpenAI, Anthropic, Google, Meta, and more.

  1. Download the CoreAI app for iOS, Android, or use the Web App
  2. Select Arcee AI: Virtuoso Large from the model selector
  3. Start chatting, comparing, or creating with AI

More Arcee-ai Models

Arcee-ai

Arcee AI: Trinity Large Preview (free)

Trinity-Large-Preview is a frontier-scale open-weight language model from Arcee, built as a 400B-parameter sparse Mixture-of-Experts with 13B active p
131K budget
Arcee-ai

Arcee AI: Trinity Mini (free)

Trinity Mini is a 26B-parameter (3B active) sparse mixture-of-experts language model featuring 128 experts with 8 active per token. Engineered for eff
131K budget
Arcee-ai

Arcee AI: Trinity Mini

Trinity Mini is a 26B-parameter (3B active) sparse mixture-of-experts language model featuring 128 experts with 8 active per token. Engineered for eff
131K budget
Arcee-ai

Arcee AI: Spotlight

Spotlight is a 7‑billion‑parameter vision‑language model derived from Qwen 2.5‑VL and fine‑tuned by Arcee AI for tight image‑text grounding tasks. It
131K budget
Arcee-ai

Arcee AI: Maestro Reasoning

Maestro Reasoning is Arcee's flagship analysis model: a 32 B‑parameter derivative of Qwen 2.5‑32 B tuned with DPO and chain‑of‑thought RL for step‑by‑
131K standard
Arcee-ai

Arcee AI: Coder Large

Coder‑Large is a 32 B‑parameter offspring of Qwen 2.5‑Instruct that has been further trained on permissively‑licensed GitHub, CodeSearchNet and synthe
33K budget

Try Arcee AI: Virtuoso Large Now

Chat with Arcee AI: Virtuoso Large and 300+ other AI models — all in one app.

Download App → Try on Web App