Together AI adds tool calling, reasoning traces, and vision-language fine-tuning to its platform, with 6x throughput gains for 100B+ parameter models. (Read MoreTogether AI adds tool calling, reasoning traces, and vision-language fine-tuning to its platform, with 6x throughput gains for 100B+ parameter models. (Read More

Together AI Upgrades Fine-Tuning Platform With Vision and Reasoning Support

2026/03/19 02:27
3 min read
For feedback or concerns regarding this content, please contact us at crypto.news@mexc.com

Together AI Upgrades Fine-Tuning Platform With Vision and Reasoning Support

Joerg Hiller Mar 18, 2026 18:27

Together AI adds tool calling, reasoning traces, and vision-language fine-tuning to its platform, with 6x throughput gains for 100B+ parameter models.

Together AI Upgrades Fine-Tuning Platform With Vision and Reasoning Support

Together AI rolled out a major expansion to its fine-tuning service on March 18, adding native support for tool calling, reasoning traces, and vision-language models—capabilities that address persistent pain points for teams building production AI systems.

The update arrives as the company reportedly negotiates a funding round that would value it at $7.5 billion, more than doubling its $3.3 billion valuation from its February 2025 Series B.

What's Actually New

The platform now handles three categories of fine-tuning that previously required fragmented workarounds:

Tool calling gets end-to-end support using OpenAI-compatible schemas. The system validates that every tool call in training data matches declared functions before training begins—a safeguard against the hallucinated parameters and schema mismatches that plague agentic workflows.

Reasoning fine-tuning allows teams to train models on domain-specific thinking traces using a dedicated reasoning_content field. This matters because reasoning formats vary wildly across model families, making consistent training difficult without standardization.

Vision-language fine-tuning supports hybrid datasets mixing image-text and text-only examples. By default, the vision encoder stays frozen while language layers update, though teams can enable joint training when visual pattern recognition needs improvement.

Infrastructure Upgrades

Beyond new capabilities, Together AI claims significant performance gains from optimizing its training stack for mixture-of-experts architectures. The company integrated SonicMoE kernels that overlap memory operations with computation, plus custom CUDA kernels for loss computation.

Results vary by model size: smaller models see roughly 2x throughput improvements, while larger architectures like Kimi-K2 hit 6x gains. The platform now handles datasets up to 100GB and models exceeding 100 billion parameters.

New models available for fine-tuning include Qwen 3.5 variants (up to 397B parameters), Kimi K2 and K2.5, and GLM-4.6 and 4.7.

Practical Additions

The update includes cost estimation before job execution and live progress tracking with dynamic completion estimates—features that sound basic but prevent the budget surprises that make experimentation risky.

XY.AI Labs, cited by Together AI as a customer example, reported moving from weekly to daily iteration cycles while cutting costs 2-3x and improving accuracy from 77% to 87% using the platform's fine-tuning and deployment APIs.

Market Context

The timing aligns with a surge in AI infrastructure spending. Startup funding in the AI sector hit $220 billion in the first two months of 2026, per recent reports, with much of that capital flowing toward training and inference infrastructure.

Together AI positions itself as an alternative to building in-house AI infrastructure, offering access to over 200 open-source models through its platform. The company's pitch—removing infrastructure complexity so teams can focus on product development—now extends to increasingly sophisticated post-training workflows that were previously the domain of well-resourced research labs.

Image source: Shutterstock
  • together ai
  • ai infrastructure
  • fine-tuning
  • machine learning
  • enterprise ai
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact crypto.news@mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.
Tags:

KAIO Global Debut

KAIO Global DebutKAIO Global Debut

Enjoy 0-fee KAIO trading and tap into the RWA boom