HOT
Sharecaster
No Result
View All Result
Sharecaster
No Result
View All Result
Advertisement Banner
Home Technology

DeepCoder-14B matches OpenAI’s o3-mini performance with a smaller footprint

Sharecaster by Sharecaster
April 9, 2025
in Technology
383 16
0
DeepCoder-14B matches OpenAI’s o3-mini performance with a smaller footprint
548
SHARES
2.5k
VIEWS
Share on FacebookShare on Twitter



summary
Summary

Agentica and Together AI release DeepCoder-14B, a new open-source language model designed for code generation.

The model aims to deliver similar performance to closed systems like OpenAI’s o3-mini, but with a smaller footprint. According to benchmark tests on LiveCodeBench, DeepCoder-14B performs at the same level as o3-mini while potentially requiring less computing power to run.

Scatter plot: LiveCodeBench Pass@1 accuracy vs model size, DeepCoder competes with large models with a small number of parameters.
Benchmark results show DeepCoder matching advanced models o3-mini (low) and o1, despite using fewer parameters. | Image: Together AI

Share

Recommend our article

Together AI developed a technique called “one-off pipelining” that reportedly cuts training time in half. The process runs training, reward calculation, and sampling in parallel, with each training iteration requiring over 1,000 separate tests. Training ran for two and a half weeks on 32 Nvidia H100 GPUs.

The training data combined 24,000 programming problems from three key sources: TACO Verified (7,500 problems), PrimeIntellects SYNTHETIC-1 (16,000 problems), and LiveCodeBench (600 problems). Each problem needed at least five test cases and a verified solution. Popular datasets like KodCode and LeetCode didn’t make the cut, either for being too simple or having inadequate test coverage.

Ad

THE DECODER Newsletter

The most important AI news straight to your inbox.

✓ Weekly

✓ Free

✓ Cancel at any time

In addition, Together AI implemented what they call a “sparse outcome reward” system – the model only receives positive feedback when its code passes all test cases. For problems with many tests, it focuses on the 15 most challenging ones.

During development, the team gradually increased the model’s context window from 16,000 to 32,000 tokens. The results improved steadily: 54 percent accuracy with 16,000 tokens, 58 percent with 32,000 tokens, and finally reaching 60.6 percent at 64,000 tokens.

Line diagram: development of rewards, response length and context window during training of the
Training metrics visualization showing rewards, response length, and context window growth from 16K to 32K tokens during model development. | Image: Together AI

This scaling ability distinguishes DeepCoder from its foundation model, DeepSeek-R1-Distill-Qwen-14B, which doesn’t show similar improvements with larger context windows. As training progressed, the model’s average response length grew from 8,000 to 17,500 tokens.

Full open-source release planned

Beyond coding, the model shows strong mathematical reasoning skills. It achieved 73.8 percent accuracy on AIME2024 problems, a 4.1 percent improvement over its base model.

While OpenAI recently announced it would share model weights for an upcoming reasoning system, Together AI is making everything available to the open-source community – including code, training data, logs, and system optimizations.

Recommendation

Google Deepmind's new PEER architecture uses a million tiny experts to boost AI efficiency

Google Deepmind's new PEER architecture uses a million tiny experts to boost AI efficiency



Source link

Advertisement Banner
Sharecaster

Sharecaster

Amplify Your Content. Broadcast multimedia posts to every social platform—all from one profile.

Trending

Technology

Google boosts AI coding and video skills with Gemini 2.5 Pro

23 hours ago
What it means for the ChatGPT future
Technology

What it means for the ChatGPT future

2 days ago
Manus AI: how to get started and what it can do
Technology

Manus AI: how to get started and what it can do

3 days ago
Google’s AI model beats 29-year-old video game, CEO Sundar Pichai says: What …
Technology

Google’s AI model beats 29-year-old video game, CEO Sundar Pichai says: What …

4 days ago
OpenAI Launches BrowseComp to Benchmark AI Agents’ Web Search and Deep Research Skills
Technology

OpenAI Launches BrowseComp to Benchmark AI Agents’ Web Search and Deep Research Skills

5 days ago
Sharecaster

We bring you the best Premium WordPress Themes that perfect for news, magazine, personal blog, etc. Check our landing page for details.

Follow Us

Recent News

Google boosts AI coding and video skills with Gemini 2.5 Pro

May 8, 2025
What it means for the ChatGPT future

What it means for the ChatGPT future

May 7, 2025

Categories

  • Animals
  • Buzz
  • Celebs
  • Life
  • Tech
  • Technology
  • Video

Tags

Art Entertainment Funny Health News Split Post Viral
  • About
  • Advertise
  • Privacy & Policy
  • Contact Us

© 2025 JNews - Premium WordPress news & magazine theme by Jegtheme.

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • Home
  • Contact Us
  • Amplify
    • Multi-Platform Sharing
    • Content Optimization
    • Viral Boost Tools
  • Analytics+
    • Audience Insights
    • Engagement Metrics
    • Performance Reports
  • Broadcast Studio
    • Story Creator
    • Live Streaming
    • Multimedia Posts
  • How It Works
    • FAQ
    • Case Studies
    • Tutorials
  • Pricing
  • Profile Hub
    • Unified Dashboard
    • Custom Branding
    • Account Management
  • Sync Everywhere
    • Platform Integrations
    • One-Click Scheduling
    • Auto-Post to All Networks
  • Technology

© 2025 JNews - Premium WordPress news & magazine theme by Jegtheme.

Go to mobile version
Skip to toolbar
  • About WordPress
    • WordPress.org
    • Documentation
    • Learn WordPress
    • Support
    • Feedback
  • Log In
  • AMP
    • View AMP version
    • Get support