Transformers

4 Posts

Efficient Foundations animation showing layered AI model components.
Transformers

More-Efficient Training for Transformers: Researchers reduce transformer training costs by 20% with minimal performance loss

Researchers cut the processing required to train transformers by around 20 percent with only a slight degradation in performance.
Gemma Scope 2
Transformers

Making LLMs Explainable: Google’s Gemma Scope probes how large language models think

Researchers have probed the inner workings of individual layers of large language models. A new tool applies this approach to all layers.
Throughput and latency at different context lengths
Transformers

Long Context Gets Up to Speed: AI21 Labs’ Jamba 1.5 outpaces transformers in long-text processing

A new model generates tokens faster than current transformers, especially when processing long inputs.
A man with electrodes connected through his skull is connected to a machine.
Transformers

A Lost Voice Regained: Brain implants paired with neural network reconstruct speech for ALS patient

A man who lost the ability to speak four years ago is sounding like his earlier self, thanks to a collection of brain implants and machine learning models.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox