Responsible AI

23 Posts

COMPL-AI workflow diagram showing compliance steps for AI models under the EU AI Act.
Responsible AI

Does Your Model Comply With the AI Act?: COMPL-AI study measures LLMs’ compliance with EU’s AI act

A new study suggests that leading AI models may meet the requirements of the European Union’s AI Act in some areas, but probably not in others.
Nuclear power plant cooling towers emitting steam into the sky.
Responsible AI

AI Giants Go Nuclear: Amazon, Google, and Microsoft bet on nuclear power to meet AI energy demands

Major AI companies plan to meet the growing demand with nuclear energy.
Art Attack: ArtPrompt, a technique that exploits ASCII art to bypass LLM safety measures
Responsible AI

Art Attack: ArtPrompt, a technique that exploits ASCII art to bypass LLM safety measures

Seemingly an innocuous form of expression, ASCII art opens a new vector for jailbreak attacks on large language models (LLMs), enabling them to generate outputs that their developers tuned them to avoid producing.
Hallucination Detector: Oxford scientists propose effective method to detect AI hallucinations
Responsible AI

Hallucination Detector: Oxford scientists propose effective method to detect AI hallucinations

Large language models can produce output that’s convincing but false. Researchers proposed a way to identify such hallucinations. 
Seoul AI Summit Spurs Safety Agreements: AI summit in Seoul achieves safety commitments from companies and governments
Responsible AI

Seoul AI Summit Spurs Safety Agreements: AI summit in Seoul achieves safety commitments from companies and governments

At meetings in Seoul, government and corporate officials from dozens of countries agreed to take action on AI safety.
U.S. and China Seek AI Agreement: U.S. and China open dialogue to prevent AI catastrophes
Responsible AI

U.S. and China Seek AI Agreement: U.S. and China open dialogue to prevent AI catastrophes

The United States and China opened a dialogue to avert hypothetical AI catastrophes.
Disinformation Documented: OpenAI takes action against misuse of its models in propaganda
Responsible AI

Disinformation Documented: OpenAI takes action against misuse of its models in propaganda

OpenAI models were used in five disinformation campaigns, the company said.
Why ChatGPT Acts That Way: OpenAI introduces guidelines for model behavior, seeks public feedback
Responsible AI

Why ChatGPT Acts That Way: OpenAI introduces guidelines for model behavior, seeks public feedback

OpenAI pulled back the curtain on revised rules that will guide its models. 
Standard for Media Watermarks: C2PA introduces watermark tech to combat media misinformation.
Responsible AI

Standard for Media Watermarks: C2PA introduces watermark tech to combat media misinformation.

An alliance of major tech and media companies introduced a watermark designed to distinguish real from fake media starting with images. The Coalition for Content Provenance and Authenticity (C2PA) offers an open standard that marks media files with information about their creation and editing.
OpenAI Revamps Safety Protocol: Inside OpenAI's framework to evaluate and mitigate model risks
Responsible AI

OpenAI Revamps Safety Protocol: Inside OpenAI's framework to evaluate and mitigate model risks

Retrenching after its November leadership shakeup, OpenAI unveiled a new framework for evaluating risks posed by its models and deciding whether to limit their use. 
GPT-4 Wouldn’t Lie to Me . . . Would It?: Researchers showed how GPT-4 can deceive users without being prompted to do so explicitly.
Responsible AI

GPT-4 Wouldn’t Lie to Me . . . Would It?: Researchers showed how GPT-4 can deceive users without being prompted to do so explicitly.

It’s well known that large language models can make assertions that are blatantly false. But can they concoct outright lies? In a proof-of-concept demonstration, Jérémy Scheurer, Mikita Balesni, and Marius Hobbhahn at Apollo Research...
High Anx-AI-ety: A recap of 2023's battle between AI doomsday warnings and regulatory measures
Responsible AI

High Anx-AI-ety: A recap of 2023's battle between AI doomsday warnings and regulatory measures

Angst at the prospect of intelligent machines boiled over in moves to block or limit the technology. Fear of AI-related doomsday scenarios prompted proposals to delay research and soul searching by prominent researchers. Amid the doomsaying, lawmakers took dramatic regulatory steps. 
Champion for Openness: Top companies launch the AI Alliance to ensure safe and open source AI.
Responsible AI

Champion for Openness: Top companies launch the AI Alliance to ensure safe and open source AI.

A new consortium aims to support open source AI. Led by Meta and IBM, dozens of organizations from the software, hardware, nonprofit, public, and academic sectors formed the AI Alliance, which plans to develop tools and programs that aid open development.
Europe Clamps Down: The AI Act, Europe's biggest AI law, moves closer to approval.
Responsible AI

Europe Clamps Down: The AI Act, Europe's biggest AI law, moves closer to approval.

Europe’s sweeping AI law moved decisively toward approval. After years of debate, representatives of the European Union’s legislative and executive branches agreed on a draft of the AI Act, a comprehensive approach to regulating AI.
Colorado flag with a neural network over it
Responsible AI

Limits on AI in Life Insurance: All about the first law that regulates use of AI in life insurance in the U.S.

The U.S. state of Colorado started regulating the insurance industry’s use of AI. Colorado implemented the first law that regulates the use of AI in life insurance and proposed extending the limits to auto insurers.
Load More

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox