ReadAgent: Bringing Gist Memory to AI

Learn how gist memory improves long context handling for…


Cerebras Breaks Exascale Record for Molecular Dynamics Simulations

Cerebras has set a new record for molecular dynamics…


Supercharge your HPC Research with the Cerebras SDK

Cerebras SDK 1.1.0, our second publicly available release,…


Key Insights from the 1st Multilingual Workshop

At NeurIPS 2023, we kicked off our inaugural workshop on…


Can Sparsity Make AI Models More Accurate?

Cerebras introduces Sparse-IFT, a technique that, through…


Cerebras Announces Fine-Tuning on the Cerebras AI Model Studio

Announcing the addition of fine-tuning capabilities for…


What’s New in R0.6 of the Cerebras SDK

The latest release of our SDK includes a host of new…


Real-Time Computational Physics with Wafer-Scale Processing [updated]

Cerebras and NETL achieve two orders of magnitude…


To Bfloat or not to Bfloat? That is the Question!

The bfloat16 data format for deep learning shortens…


More Pixels, More Context, More Insight!

The Cerebras architecture is designed to overcome the…


Cerebras Software Platform R1.7 is Out!

Our new release expands PyTorch support, releases code…


SC22 is a Wrap!

At this year’s supercomputing show in Dallas, we announced…


Harnessing the Power of Sparsity for Large GPT AI Models

Enabling innovation of novel sparse ML techniques to…


Creating Sparse GPT-3 Models with Iterative Pruning

We have trained extremely sparse GPT-3 1.3B parameter…


Sandia Awards Major Contract to Cerebras Systems

Learn more about what impressed Sandia Labs about Cerebras…


Cerebras Goes Massive at Supercomputing 2022

Visit our booth for unique hardware, expert AI and HPC…


What is Appliance Mode?

We created Appliance Mode for simplicity, ease of use, and…


The Complete Guide to Scale-Out on Cerebras Wafer-Scale Clusters

Cerebras’ appliance mode offers users easy linear scaling…


Linear Scaling Made Possible with Weight Streaming

In a single keystroke, Cerebras can scale large language…


Context is Everything: Why Maximum Sequence Length Matters

GPU-Impossible™ sequence lengths on Cerebras systems may…


How to Harness the Predictive Power of GPT-J

A look at why this open-source language model is so…


Tensor Shape: Increasing Model Throughput

We write machine learning algorithms to fit the data, not…


Getting Started with PyTorch BERT Models

This walk through shows how easy it is to adapt and run…


Cerebras Software Platform R1.2 is Out!

Our new release expands PyTorch support and unlocks…


Celebrating International Women’s Day

At Cerebras Systems, we’re on a mission to revolutionize…


A Big Chip for Big Science: Watching the COVID-19 Virus in Action

It’s hard to imagine a better example of “AI for good” than…


An AI Chip With Unprecedented Performance To Do the Unimaginable

AI accelerator chips have made machine learning a reality…


Innovations in Virtual Drug Screening

Drug discovery has seen substantial innovation in recent…


Error Compensation Mechanism in Online Normalization

Online Normalization is a new technique for normalizing the…