August 28, 2024
Integrating LLMs and Software Engineering for Self-Refining Copy Creation
Discover how to build an AI agent that generates marketing…
August 28, 2024
ReadAgent: Bringing Gist Memory to AI
Learn how gist memory improves long context handling for…
August 28, 2024
Llama3.1 Model Quality Evaluation: Cerebras, Groq, SambaNova, Together, and Fireworks
Cerebras' new inference solution redefines AI performance,…
June 25, 2024
US DOE Achieves 88x Performance Speedup with Cerebras CS-2 Over H100 in Materials Modeling
Using the Cerebras CS-2, NETL implements the venerable…
May 17, 2024
Cerebras Breaks Exascale Record for Molecular Dynamics Simulations
Cerebras has set a new record for molecular dynamics…
May 1, 2024
Supercharge your HPC Research with the Cerebras SDK
Cerebras SDK 1.1.0, our second publicly available release,…
April 12, 2024
Cerebras CS-3 vs. Nvidia B200: 2024 AI Accelerators Compared
In the fast-paced world of AI hardware, the Cerebras CS-3…
March 12, 2024
Cerebras CS-3: the world’s fastest and most scalable AI accelerator
Today Cerebras is introducing the CS-3, our…
February 5, 2024
Key Insights from the 1st Multilingual Workshop
At NeurIPS 2023, we kicked off our inaugural workshop on…
February 5, 2024
Sparsity Made Easy – Introducing the Cerebras PyTorch Sparsity Library
We release our PyTorch-based sparsity library allowing ML…
December 29, 2023
Introducing gigaGPT: GPT-3 sized models in 565 lines of code
GigaGPT is Cerebras’ implementation of Andrei Karpathy’s…
December 5, 2023
Cerebras Pioneers Ethical AI Development through Collaborative AI Initiatives
Today, Cerebras proudly revealed our pivotal role as a…
November 10, 2023
Cerebras Software Release 2.0: 50% Faster Training, PyTorch 2.0 Support, Diffusion Transformers, and More
Today we are excited to announce Cerebras software release…
October 12, 2023
How we fine-tuned Llama2-70B to pass the US Medical License Exam in a week
New open-access model by M42 outperforms GPT-3.5 in…
September 5, 2023
Jais: a New Pinnacle in Open Arabic NLP
Introducing a new state-of-the-art bi-lingual…
July 24, 2023
BTLM-3B-8K: 7B Performance in a 3 Billion Parameter Model
Cerebras and Opentensor introduce a new standard for…
July 22, 2023
Accelerating Large Language Model Training with Variable Sparse Pre-training and Dense Fine-tuning
We reduced pre-training FLOPs by 64% using sparsity. To the…
July 22, 2023
Variable Sequence Length Training for Long-Context Large Language Models
We show it is possible to accelerate the training for large…
July 20, 2023
Introducing Condor Galaxy 1: a 4 exaFLOPS Supercomputer for Generative AI
Cerebras, in partnership with G42 unveils CG-1, a 4…
June 9, 2023
SlimPajama: A 627B token, cleaned and deduplicated version of RedPajama
Today we are releasing SlimPajama – the largest…
May 23, 2023
Efficient Large-Scale GPT Training Using a Cerebras Wafer-Scale Cluster
Cerebras has built a platform for push-button training of…
May 22, 2023
Cerebras Architecture Deep Dive: First Look Inside the HW/SW Co-Design for Deep Learning [Updated]
Our ML-optimized architecture enables the largest models to…
April 17, 2023
Fine-Tuning with Cerebras AI Model Studio Launchpad
Cerebras shares research showing smaller foundation models…
March 28, 2023
Cerebras-GPT: A Family of Open, Compute-efficient, Large Language Models
Cerebras open sources seven GPT-3 models from 111 million…
March 22, 2023
Can Sparsity Make AI Models More Accurate?
Cerebras introduces Sparse-IFT, a technique that, through…
March 21, 2023
Accelerating Large GPT Training with Sparse Pre-Training and Dense Fine-Tuning [Updated]
We have shown it is possible to reduce the training compute…
February 16, 2023
Cerebras Announces Fine-Tuning on the Cerebras AI Model Studio
Announcing the addition of fine-tuning capabilities for…
February 15, 2023
What’s New in R0.6 of the Cerebras SDK
The latest release of our SDK includes a host of new…
February 13, 2023
Unlocking High-Resolution Computer Vision with Wafer-Scale Technology
We have built a platform for accelerating CV workloads that…
February 7, 2023
Real-Time Computational Physics with Wafer-Scale Processing [updated]
Cerebras and NETL achieve two orders of magnitude…
January 30, 2023
To Bfloat or not to Bfloat? That is the Question!
The bfloat16 data format for deep learning shortens…
January 27, 2023
More Pixels, More Context, More Insight!
The Cerebras architecture is designed to overcome the…
January 17, 2023
Cerebras Software Platform R1.7 is Out!
Our new release expands PyTorch support, releases code…
November 29, 2022
The Cerebras AI Model Studio brings Wafer-Scale Cluster Acceleration to the Cloud
Announcing our dedicated cloud service which makes large…
November 28, 2022
Harnessing the Power of Sparsity for Large GPT AI Models
Enabling innovation of novel sparse ML techniques to…
November 28, 2022
Creating Sparse GPT-3 Models with Iterative Pruning
We have trained extremely sparse GPT-3 1.3B parameter…
November 14, 2022
Genomics in Unparalleled Resolution: Cerebras Wafer-Scale Cluster Trains Large Language Models on the Full COVID Genome Sequence
Our joint work with Argonne National Laboratory (ANL) and…
November 2, 2022
Sandia Awards Major Contract to Cerebras Systems
Learn more about what impressed Sandia Labs about Cerebras…
October 19, 2022
Cerebras Goes Massive at Supercomputing 2022
Visit our booth for unique hardware, expert AI and HPC…
September 14, 2022
What is Appliance Mode?
We created Appliance Mode for simplicity, ease of use, and…
September 14, 2022
The Complete Guide to Scale-Out on Cerebras Wafer-Scale Clusters
Cerebras’ appliance mode offers users easy linear scaling…
September 14, 2022
Linear Scaling Made Possible with Weight Streaming
In a single keystroke, Cerebras can scale large language…
August 15, 2022
Context is Everything: Why Maximum Sequence Length Matters
GPU-Impossible™ sequence lengths on Cerebras systems may…
August 3, 2022
Cerebras Wafer-Scale Engine Inducted into the Computer History Museum
Today was a proud day for the entire Cerebras family.
June 22, 2022
How to Harness the Predictive Power of GPT-J
A look at why this open-source language model is so…
June 22, 2022
Cerebras Sets Record for Largest AI Models Ever Trained on Single Device
Our customers can easily train and reconfigure GPT-3 and…
June 22, 2022
Training Multi-Billion-Parameter Models on a Single Cerebras System is Easy
Changing model size is trivial on Cerebras, rather than a…
June 22, 2022
Cerebras Makes It Easy to Harness the Predictive Power of GPT-J
A look at why this open-source language model is so…
June 1, 2022
Multi-Billion-Parameter Model Training Made Easy with CSoft R1.3
CSoft R1.3 delivers GPT-J continuous pre-training, more…
June 1, 2022
Tensor Shape: Increasing Model Throughput
We write machine learning algorithms to fit the data, not…
April 14, 2022
Getting Started with PyTorch BERT Models
This walk through shows how easy it is to adapt and run…
April 14, 2022
TotalEnergies and Cerebras Create Massively Scalable Stencil Algorithm
TotalEnergies used the Cerebras CS-2 system to turn a…
April 13, 2022
Cerebras Software Platform R1.2 is Out!
Our new release expands PyTorch support and unlocks…
March 21, 2022
When Time is Money: Accelerating NLP Model Training at a Leading Financial Institution
Our CS-2 system delivered the compute performance of more…
March 8, 2022
Celebrating International Women’s Day
At Cerebras Systems, we’re on a mission to revolutionize…
March 2, 2022
TotalEnergies and Cerebras: Accelerating into a Multi-Energy Future (Updated)
Cerebras is helping TotalEnergies to advance their clean…
January 14, 2022
If You’re Doing Pharma and Life Sciences AI Research Without a Cerebras System, You’re Doing it Wrong
Cerebras AI accelerator systems make it possible to train…
December 14, 2021
A Big Chip for Big Science: Watching the COVID-19 Virus in Action
It’s hard to imagine a better example of “AI for good” than…
September 14, 2021
Scaling Up and Out: Training Massive Models on Cerebras Systems using Weight Streaming
A look at how the Wafer-Scale Engine trains models much…
August 17, 2021
An AI Chip With Unprecedented Performance To Do the Unimaginable
AI accelerator chips have made machine learning a reality…
June 16, 2021
Innovations in Virtual Drug Screening
Drug discovery has seen substantial innovation in recent…
April 20, 2020
Error Compensation Mechanism in Online Normalization
Online Normalization is a new technique for normalizing the…