Responsible AI

67 Posts

Diagram shows AI traits with pipelines for "evil" vs. "helpful" responses to user queries on animal treatment.
Responsible AI

Toward Steering LLM Personality: Persona Vectors allow model builders to identify and edit out sycophancy, hallucinations, and more

Large language models can develop character traits like cheerfulness or sycophancy during fine-tuning. Researchers developed a method to identify, monitor, and control such traits.
Hands strum a guitar covered in labels from major record companies, symbolizing AI music innovation.
Responsible AI

Record Labels Back AI-Music Startup: Klay Image emerges from relative obscurity to announce deals with Sony, Warner, and Universal

A music-generation newcomer emerged from stealth mode with licenses to train generative AI models on music controlled by the world’s biggest recording companies.
Visual map outlines cybercrime operation phases, highlighting AI-driven processes and human validation steps.
Responsible AI

Anthropic Cyberattack Report Sparks Controversy: Security researchers question whether coding agents allow unprecedented automated attacks

Independent cybersecurity researchers pushed back on a report by Anthropic that claimed hackers had used its Claude Code agentic coding system to perpetrate an unprecedented automated cyberattack.
AI models are compared on a graph showing benchmark accuracy from 20% to 100%, highlighting GPT-5's rise.
Responsible AI

The Year AI Went Industrial: The State of AI Report 2025 says AI’s barriers aren’t technological but social and material

A year-in-review report heralds the dawn of AI’s industrial era.
Icon of silhouettes of kids with a ban symbol, indicating limited chatbot use by teens.
Responsible AI

Toward Safer (and Sexier) Chatbots: Inside Character AI and OpenAI’s policy changes to protect younger and vulnerable Users

Chatbot providers, facing criticism for engaging troubled users in conversations that deepen their distress, are updating their services to provide wholesome interactions to younger users while allowing adults to pursue erotic conversations.
Chart illustrates exact and approximate memorization percentages in different Gemma models.
Responsible AI

Masking Private Data in Training Sets: Google researchers released VaultGemma, an open-weights model redacting personal information

Large language models often memorize details in their training data, including private information that may appear only once, like a person’s name, address, or phone number. Researchers built the first open-weights language model that’s guaranteed not to remember such facts.
Prominent dollar sign amid geometric lines, illustrating OpenAI's shift to for-profit status
Responsible AI

OpenAI Reorganizes For Profit: ChatGPT’s maker completed restructuring, freeing it to go public, make deals with new partners

OpenAI completed its transition from nonprofit to for-profit in a feat of legal engineering that took an army of lawyers, investment bankers, and two state attorneys general 18 months to negotiate.
Kids in various Halloween costumes walk on a street as numerous witches fly above in an orange sunset sky.
Responsible AI

Autonomous Systems Wage War: Drones are redefining warfare. What if humans lose control?

Drones are becoming the deadliest weapons in today’s war zones, and they’re not just following orders. Should AI decide who lives or dies?
Characters dressed for Halloween blowing bubbles, hinting at the AI industry's speculative bubble.
Responsible AI

The AI Boom Is Bound to Bust: What if big investments in AI models, data centers, and hot startups don't pay off?

Leading AI companies are spending mountains of cash in hopes that the technology will deliver outsize profits before investors lose patience. Are exuberant bets on big returns grounded in the quicksand of wishful thinking?
A rabbit leads a viking-costumed person into a hole, holding a bag of toys, against a forest backdrop.
Responsible AI

Chatbots Lead Users Into Rabbit Holes: When paranoia, delusions, and other signs of mental illness meet AI

Conversations with chatbots are loosening users’ grips on reality, fueling the sorts of delusions that can trigger episodes of severe mental illness. Are AI models driving us insane?
Official letterhead displays Governor Newsom's signature on AI regulation bill establishing state oversight.
Responsible AI

California Builds AI Regulatory Regime: The U.S.’s biggest state by population and economy passed four AI transparency bills is less than one month

In the absence of national laws that specifically regulate AI in the United States, California moved to regulate the technology within its own borders, passing four bills in less than a month.
US map showing Nevada and Illinois highlighted in red, marking states that banned AI-driven mental health treatments.
Responsible AI

States Ban AI-Driven Treatments for Mental Health: Illinois follows Nevada, prohibiting certain uses of chatbots unless used by licensed therapists

Illinois became the second U.S. state, after Nevada, to ban AI applications that administer psychotherapy.
Students at Alpha School working on laptops during AI-assisted personalized learning session.
Responsible AI

2 Hours With AI Versus 6 With Teacher: Inside Alpha School, a Texas-based program using algorithms and video monitors to teach children

A growing private school system replaces the typical 6-hour school day with 2 hours of personalized, AI-assisted education. 
AI chatbot interfaces showing tour guide, outdoor adventurer, and custom characters as Meta and OpenAI add safety controls.
Responsible AI

Meta, OpenAI Reinforce Guardrails: Meta and OpenAI respond to criticism by adding new rules for teens’ chatbot use

Meta and OpenAI promised to place more controls on their chatbots’ conversations with children and teenagers, as worrisome interactions with minors come under increasing scrutiny.
Google study chart comparing energy use of AI accelerators for Gemini, including chip power, CPU, and idle machines.
Responsible AI

Gemini’s Environmental Impact Measured: Google study directly measures electricity, water use, and greenhouse emissions of its models

Google determined that its large language models have a smaller environmental footprint than previous estimates had led it to expect.
Load More

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox