Large Language Models

8 Posts

Reducing Memorization in LLMs: A technique that masks tokens in large language models, protecting data privacy
Large Language Models

Reducing Memorization in LLMs: A technique that masks tokens in large language models, protecting data privacy

Studies have established that large language models can memorize the text passages they’ve been trained on repeatedly and regurgitate them when prompted in adversarial and, though rarely, in benign ways.
Covariant robotic arm clutching an Amazon box.
Large Language Models

Amazon Boosted by Covariant: Amazon strengthens logistics and robotics with new AI partnership

Amazon took on talent and technology from robotics startup Covariant to enhance its warehouse automation, an area critical to its core ecommerce business.
High Gear for Llama 3.1 405B: SambaNova boosts Llama 3.1 performance with fast, free access to largest model
Large Language Models

High Gear for Llama 3.1 405B: SambaNova boosts Llama 3.1 performance with fast, free access to largest model

SambaNova raised the speed limit for access to the largest model in the Llama 3.1 family — and it’s free.
OpenAI's model scores on the GPQA Diamond tests in biology, chemistry, and physics, along with their overall score.
Large Language Models

OpenAI Forges Chains of Thought: OpenAI’s o1 models excel in reasoning, outperform GPT-4o in math and coding

Preliminary versions of OpenAI’s new model family were trained explicitly to think step-by-step, yielding outstanding marks in math, science, and coding — but users can’t see their reasoning steps.
Gemma Scope 2
Large Language Models

Making LLMs Explainable: Google’s Gemma Scope probes how large language models think

Researchers have probed the inner workings of individual layers of large language models. A new tool applies this approach to all layers.
Short, Medium and Long Context RAG
Large Language Models

Models Ranked for Hallucinations: Measuring language model hallucinations during information retrieval

How often do large language models make up information when they generate text based on a retrieved document? A study evaluated the tendency of popular models to hallucinate while performing retrieval-augmented generation (RAG). 
Throughput and latency at different context lengths
Large Language Models

Long Context Gets Up to Speed: AI21 Labs’ Jamba 1.5 outpaces transformers in long-text processing

A new model generates tokens faster than current transformers, especially when processing long inputs.
The SWE-bench full leaderboard shows Cosine Genie outperforming its competitors.
Large Language Models

Agentic Coding Strides Forward: Genie coding assistant outperforms competitors on SWE-bench by over 30 percent

An agentic coding assistant boosted the state of the art in an important benchmark by more than 30 percent.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox