20B小モデルの検索能力がGPT-5とOpusに追いつく:ベクトルデータベースChromaがオープンソースエージェント検索モデルContext-1

BlockBeatNews

According to 1M AI News monitoring, the open-source vector database Chroma has released Context-1, a 20 billion parameter intelligent search model specifically designed for multi-turn retrieval tasks. The model weights are open-sourced under the Apache 2.0 license, and the code for the synthetic data generation pipeline has been made public simultaneously.

The positioning of Context-1 is as a retrieval subagent: it does not directly answer questions but returns a set of supporting documents for downstream reasoning models through multi-turn searches. The core technology is “self-editing context,” which means the model actively discards irrelevant document fragments during the search process, making space for subsequent searches within a limited context window to avoid performance degradation caused by context bloat.

Training is divided into two phases: first, using large models like Kimi K2.5 to generate SFT trajectories for supervised fine-tuning warm-up, followed by training on over 8,000 synthetic tasks through reinforcement learning (based on the CISPO algorithm). The reward design employs a curriculum mechanism, encouraging broad exploration in the early stages with a focus on recall, and gradually shifting towards precision to encourage selective retention in the later stages. The base model is gpt-oss-20b, adapted using LoRA, and runs on B200 with MXFP4 quantization during inference, achieving a throughput of 400-500 token/s.

On Chroma’s self-built four domain benchmarks (web, finance, law, email) and public benchmarks (BrowseComp-Plus, SealQA, FRAMES, HotpotQA), the 4-way parallel version of Context-1 is on par with or close to cutting-edge models like GPT-5.2, Opus 4.5, Sonnet 4.5 in the “final answer hit rate” metric, for example, achieving 0.96 on BrowseComp-Plus (Opus 4.5 is 0.87, GPT-5.2 is 0.82), while the cost and latency are only a fraction of the latter. Notably, this model is trained only on web, legal, and financial data, yet it also shows significant improvements in the email domain, which was not part of the training, demonstrating the cross-domain transferability of its search capabilities.

免責事項:このページの情報は第三者から提供される場合があり、Gateの見解または意見を代表するものではありません。このページに表示される内容は参考情報のみであり、いかなる金融、投資、または法律上の助言を構成するものではありません。Gateは情報の正確性または完全性を保証せず、当該情報の利用に起因するいかなる損失についても責任を負いません。仮想資産への投資は高いリスクを伴い、大きな価格変動の影響を受けます。投資元本の全額を失う可能性があります。関連するリスクを十分に理解したうえで、ご自身の財務状況およびリスク許容度に基づき慎重に判断してください。詳細は免責事項をご参照ください。
コメント
0/400
コメントなし