Overview
- ByteDance uploaded Seed-OSS-36B Base, Base-woSyn, and Instruct variants to Hugging Face under the Apache-2.0 license, enabling free commercial use.
- The 36B-parameter design spans 64 layers with a 155,000-token vocabulary, native 512,000-token context, and a controllable thinking budget for reasoning length.
- ByteDance-reported results claim open-source state-of-the-art for the Instruct model, including 91.7 on AIME24, 67.4 on LiveCodeBench v6, and 94.6 on RULER at 128K.
- Documentation covers deployment via Transformers and vLLM, with 4-bit and 8-bit quantization options to reduce memory requirements.
- The company says Seed-OSS-36B matches or surpasses similar-sized models from Alibaba Cloud, Google and OpenAI, with training reportedly using about 12 trillion tokens.