Arcee AI: Coder Large

Text input Text output
Author's Description

Coder‑Large is a 32 B‑parameter offspring of Qwen 2.5‑Instruct that has been further trained on permissively‑licensed GitHub, CodeSearchNet and synthetic bug‑fix corpora. It supports a 32k context window, enabling multi‑file refactoring or long diff review in a single call, and understands 30‑plus programming languages with special attention to TypeScript, Go and Terraform. Internal benchmarks show 5–8 pt gains over CodeLlama‑34 B‑Python on HumanEval and competitive BugFix scores thanks to a reinforcement pass that rewards compilable output. The model emits structured explanations alongside code blocks by default, making it suitable for educational tooling as well as production copilot scenarios. Cost‑wise, Together AI prices it well below proprietary incumbents, so teams can scale interactive coding without runaway spend.

Key Specifications
Cost
$$$
Context
32K
Parameters
32B (Rumoured)
Released
May 05, 2025
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Logit Bias Stop Min P Top P Max Tokens Frequency Penalty Temperature Presence Penalty
Performance Summary

Arcee AI: Coder Large, a 32B-parameter model derived from Qwen 2.5-Instruct, demonstrates strong performance in coding-related tasks and general reliability. It consistently ranks among the fastest models, achieving the 86th percentile across seven benchmarks, and offers competitive pricing, placing in the 54th percentile. Notably, the model exhibits exceptional reliability with a 100% success rate across all benchmarks, indicating minimal technical failures. In terms of specific benchmark performance, Coder Large achieved perfect accuracy in Hallucinations (Baseline), showcasing its ability to appropriately acknowledge uncertainty. It also performed well in Ethics (99.0% accuracy) and General Knowledge (96.0% accuracy). Its core strength lies in coding, where it scored 81.0% accuracy, further supported by its specialized training on GitHub, CodeSearchNet, and bug-fix corpora. While its Instruction Following (56.6% accuracy) and Email Classification (95.0% accuracy) scores are moderate, its Reasoning capabilities are solid at 72.0% accuracy. The model's ability to emit structured explanations alongside code blocks makes it particularly suitable for educational and copilot scenarios.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.5
Completion $0.8

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
Together
Together | arcee-ai/coder-large 32K $0.5 / 1M tokens $0.8 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Strategy Free Executions Accuracy Cost Duration
Other Models by arcee-ai