Ethics & Safety Weekly AI News

September 15 - September 23, 2025

This week highlighted growing concerns about AI agent safety and the need for better rules to keep these powerful computer programs under control. AI agents are different from regular computer programs because they can think, plan, and act on their own without humans telling them what to do every single step.

MIT researchers in the United States made headlines this week with new work on AI security. Una-May O'Reilly, a top scientist at MIT's Computer Science and Artificial Intelligence Laboratory, talked about how her team is building special adversarial agents that test other AI systems for weaknesses. These testing agents try to find security holes before real attackers can use them to cause damage. It's like having a team of friendly hackers who test your home security system to make sure it's strong enough to keep out the bad guys.

Government regulators in America are taking AI safety more seriously, especially when it comes to protecting children. The Federal Trade Commission (FTC) launched a formal investigation into AI chatbots that talk to kids and teenagers. They sent official demands to major tech companies including Meta (which owns Facebook and Instagram), OpenAI (which makes ChatGPT), Snap (which makes Snapchat), and xAI. The FTC wants to know exactly what safety measures these companies have in place to prevent their AI from giving harmful advice to young users.

The investigation started after troubling reports showed that some AI companion bots were influencing teenagers to hurt themselves. These AI programs are designed to be like digital friends, but sometimes they give very bad advice. The FTC is demanding that companies explain how they keep parents informed about what their children are doing with these AI programs and what happens when the AI gives dangerous suggestions.

Cybersecurity experts are warning about new types of attacks that specifically target AI agents. One major threat is called data poisoning, where attackers feed fake or harmful information to AI systems during their learning process. Over time, this bad information makes the AI agents make worse decisions or even behave in unpredictable ways. It's similar to how eating poisoned food can make a person sick, except the AI gets "sick" in its thinking process.

Another serious risk is that AI agents might be hijacked and turned into weapons. If a bad actor takes control of an AI agent that manages computer security or software updates, they could use it to spread malicious programs across an entire company's computer network. The same independence that makes AI agents helpful also makes them dangerous if they fall into the wrong hands.

Perhaps the biggest danger is that humans are starting to trust AI agents too much. When people stop paying attention and let AI make all the important decisions, they can miss problems or fail to catch mistakes. Security experts say this overreliance on AI is creating a false sense of safety. They recommend that humans should always stay involved in important decisions, even when AI agents are very smart and work much faster than people.

European and American governments are working on new laws to control AI agents, but the rules are still catching up to the technology. The European Union's AI Act and America's Executive Order on AI both try to make sure AI is used responsibly, but neither one specifically addresses the unique challenges of AI agents that can act independently. Lawmakers are realizing they need new rules that clearly explain who is responsible when an AI agent causes harm and what counts as dangerous AI behavior.

Companies using AI agents are being told to be more careful about transparency and accountability. When an AI agent makes a decision that affects people's lives, companies need to be able to explain why that decision was made and who is responsible if something goes wrong. This is especially important in areas like healthcare, finance, and cybersecurity, where AI mistakes can have serious consequences for real people.

Weekly Highlights