November 20, 2024
Chatting Your Way Through 4500 NeurIPS papers with Cerebras
A helpful directory of all the papers from NEURIPS 2024,…
August 28, 2024
Integrating LLMs and Software Engineering for Self-Refining Copy Creation
Discover how to build an AI agent that generates marketing…
August 28, 2024
ReadAgent: Bringing Gist Memory to AI
Learn how gist memory improves long context handling for…
August 28, 2024
Llama3.1 Model Quality Evaluation: Cerebras, Groq, SambaNova, Together, and Fireworks
Cerebras' new inference solution redefines AI performance,…
April 12, 2024
Cerebras CS-3 vs. Nvidia B200: 2024 AI Accelerators Compared
In the fast-paced world of AI hardware, the Cerebras CS-3…
March 12, 2024
Cerebras CS-3: the world’s fastest and most scalable AI accelerator
Today Cerebras is introducing the CS-3, our…
February 5, 2024
Key Insights from the 1st Multilingual Workshop
At NeurIPS 2023, we kicked off our inaugural workshop on…
February 5, 2024
Sparsity Made Easy – Introducing the Cerebras PyTorch Sparsity Library
We release our PyTorch-based sparsity library allowing ML…
December 29, 2023
Introducing gigaGPT: GPT-3 sized models in 565 lines of code
GigaGPT is Cerebras’ implementation of Andrei Karpathy’s…
December 5, 2023
Cerebras Pioneers Ethical AI Development through Collaborative AI Initiatives
Today, Cerebras proudly revealed our pivotal role as a…
November 10, 2023
Cerebras Software Release 2.0: 50% Faster Training, PyTorch 2.0 Support, Diffusion Transformers, and More
Today we are excited to announce Cerebras software release…
October 12, 2023
How we fine-tuned Llama2-70B to pass the US Medical License Exam in a week
New open-access model by M42 outperforms GPT-3.5 in…
September 5, 2023
Jais: a New Pinnacle in Open Arabic NLP
Introducing a new state-of-the-art bi-lingual…
July 24, 2023
BTLM-3B-8K: 7B Performance in a 3 Billion Parameter Model
Cerebras and Opentensor introduce a new standard for…
July 22, 2023
Accelerating Large Language Model Training with Variable Sparse Pre-training and Dense Fine-tuning
We reduced pre-training FLOPs by 64% using sparsity. To the…
July 22, 2023
Variable Sequence Length Training for Long-Context Large Language Models
We show it is possible to accelerate the training for large…
July 20, 2023
Introducing Condor Galaxy 1: a 4 exaFLOPS Supercomputer for Generative AI
Cerebras, in partnership with G42 unveils CG-1, a 4…
June 9, 2023
SlimPajama: A 627B token, cleaned and deduplicated version of RedPajama
Today we are releasing SlimPajama – the largest…
May 23, 2023
Efficient Large-Scale GPT Training Using a Cerebras Wafer-Scale Cluster
Cerebras has built a platform for push-button training of…
April 17, 2023
Fine-Tuning with Cerebras AI Model Studio Launchpad
Cerebras shares research showing smaller foundation models…