Frontier AI is now on Cerebras. This week we are launching Qwen3-235 Alibaba Group's flagship reasoning model that rivals ChatGPT and Claude. In classic Cerebras style, we run the model at 1,500 tokens/second. That means reasoning time goes from 60 seconds on GPUs to just 0.6 seconds. For enterprise customers, we're enabling them with 131K context, which enables production-grade code generation. Qwen3-235B will be available for all to try later this week. Meanwhile, try Qwen3-32B for free and experience the power of Cerebras Inference at https://lnkd.in/gyFuyinr
Cerebras Systems
Computer Hardware
Sunnyvale, California 72,614 followers
AI insights, faster! We're a computer systems company dedicated to accelerating deep learning.
About us
Cerebras Systems is accelerating the future of generative AI. We’re a team of pioneering computer architects, deep learning researchers, and engineers building a new class of AI supercomputers from the ground up. Our flagship system, Cerebras CS-3, is powered by the Wafer Scale Engine 3—the world’s largest and fastest AI processor. CS-3s are effortlessly clustered to create the largest AI supercomputers on Earth, while abstracting away the complexity of traditional distributed computing. From sub-second inference speeds to breakthrough training performance, Cerebras makes it easier to build and deploy state-of-the-art AI—from proprietary enterprise models to open-source projects downloaded millions of times. Here’s what makes our platform different: 🔦 Sub-second reasoning – Instant intelligence and real-time responsiveness, even at massive scale ⚡ Blazing-fast inference – Up to 100x performance gains over traditional AI infrastructure 🧠 Agentic AI in action – Models that can plan, act, and adapt autonomously 🌍 Scalable infrastructure – Built to move from prototype to global deployment without friction Cerebras solutions are available in the Cerebras Cloud or on-prem, serving leading enterprises, research labs, and government agencies worldwide. 👉 Learn more: www.cerebras.ai Join us: https://cerebras.net/careers/
- Website
-
http://www.cerebras.ai
External link for Cerebras Systems
- Industry
- Computer Hardware
- Company size
- 201-500 employees
- Headquarters
- Sunnyvale, California
- Type
- Privately Held
- Founded
- 2016
- Specialties
- artificial intelligence, deep learning, natural language processing, and inference
Products
Locations
-
Primary
1237 E Arques Ave
Sunnyvale, California 94085, US
-
150 King St W
Toronto, Ontario M5H 1J9, CA
-
Tokyo, JP
-
Bangalore, IN
Employees at Cerebras Systems
Updates
-
Cerebras Systems reposted this
🚀 Qwen3-Coder 480B is now LIVE on Cerebras + 🔥 Announcing Cerebras Code monthly plans Running on Cerebras Code, Qwen3-Coder screams at 2K tokens / sec (20x faster than GPU) with a 131K token context, no proprietary IDE lock-in, and no weekly limits! 👀 TLDR •Frontier coding accuracy on par with Claude 4 Sonnet per SWE bench •20x faster at 33% lower cost as compared to Claude 4 Sonnet •Open weight freedom – full control and abilitiy toplug into any OpenAI-compatible IDE (Cursor, Continue.dev, Cline, RooCode… you name it) •Easily accessible through two new monthly plans with no weekly limits: 👉 Cerebras Code Pro for indie developers ($50/mo) 👉 Cerebras Code Max for power users with 5x rate limits ($200/mo) 👩💻 Get your free API key, drop it into your favorite editor, and ship faster than ever: https://lnkd.in/gMdnXyE9
-
🚀 Qwen3-Coder 480B is now LIVE on Cerebras + 🔥 Announcing Cerebras Code monthly plans Running on Cerebras Code, Qwen3-Coder screams at 2K tokens / sec (20x faster than GPU) with a 131K token context, no proprietary IDE lock-in, and no weekly limits! 👀 TLDR •Frontier coding accuracy on par with Claude 4 Sonnet per SWE bench •20x faster at 33% lower cost as compared to Claude 4 Sonnet •Open weight freedom – full control and abilitiy toplug into any OpenAI-compatible IDE (Cursor, Continue.dev, Cline, RooCode… you name it) •Easily accessible through two new monthly plans with no weekly limits: 👉 Cerebras Code Pro for indie developers ($50/mo) 👉 Cerebras Code Max for power users with 5x rate limits ($200/mo) 👩💻 Get your free API key, drop it into your favorite editor, and ship faster than ever: https://lnkd.in/gMdnXyE9
-
🚀 Qwen-3‑235B‑A22B‑Thinking‑2507 is now live! This model is smarter and faster than DeepSeek R1, per Artificial Analysis ✅ Key Facts: - World’s fastest frontier reasoning model - ~1,700 tokens/sec (~26x faster than GPU) - Targeted for coding, agentic workflows, and multilingual use cases - 1.7s time-to-full-answer - 131K context length - Input: $0.60 | Output: $1.2 per M tokens - Open weights for full control 👉 Get your free API key to power your apps: https://lnkd.in/gQR-TaPG 🤗 Head over the Hugging Face and start building with blazing fast thinking: https://lnkd.in/gZP5amNA
-
-
Cerebras Systems is very proud to be offering Qwen3-235B 2507 with our strategic partner , Hugging Face ! https://lnkd.in/dDBcqHpc Thank you Clem Delangue 🤗 Vaibhav Srivastav and Julien Chaumond for the partnership!
-
Qwen 3 235B Instruct on Cerebras runs at an unprecedented 1,400 tokens/s – making it the world’s only frontier model that can generate code, run agents, and carry conversations at instant speed. Don't just take our word for it, get your API key and put it to work: https://lnkd.in/gQR-TaPG
Any questions about who is the fastest on Qwen3 235B Instruct? Just ask Artificial Analysis....or see below. Frontier model accuracy. Blisteringly fast performance on Cerebras Systems API.
-
🚀 Qwen-3‑235B‑A22B‑Instruct‑2507 is now live! Smarter, 11x faster and 70% cheaper than GPT 4.1 ✅ Key Facts: • World’s fastest frontier non-reasoning model at ~1,400+ tokens/sec (~11x faster than GPU) • Ideal for blazing fast instruction following, coding, tool use, STEM, long-context RAG, and multilingual knowledge • 131K context length • Input: $0.60/M tokens Output: $1.2/M tokens • Open weights for full control 👉 Get your free API key and start building the future of AI, today -https://lnkd.in/gQR-TaPG
-
-
🎥 Andrew Feldman and Eric Schmidt gave a masterclass on the evolution of AI at RAISE Summit. A few big ideas answered in this insightful session: - How must hardware and software co-evolve to truly accelerate the building of AI systems? - What should you not pursue when designing breakthrough AI infrastructure? - With token costs dropping 10x annually, are we witnessing the industrialization of AI? Thank you Henri Delahaye and the Raise team for hosting us! Link to full fireside chat in comments 🔽
-
Thank you for the partnership ZS - we are excited to build a future of enterprise AI that's fast, scalable, and agentic. Build with us: https://lnkd.in/g4hCRCbE
Thank you ZS and Gopi Vikranth for the Partnership. This collaboration integrates Cerebras Systems’ CS-3 systems directly into ZS’s MAX.AI platform—creating an ultra-high-performance environment for fine-tuning, training and deploying large language models (LLMs) and agentic AI solutions. https://bwnews.pr/43ZpmpA
-