Ethics & Safety Weekly AI News
September 22 - September 30, 2025This week brought major warnings about the safety of AI agents as experts and companies raised serious concerns about these powerful new tools.
JumpCloud, a technology company, released a report called "Who Let the Bot In? Control Agentic AI Before It Goes Too Far". The report shows a scary gap between how much companies use AI agents and how well they control them. While 82% of organizations already use AI agents, only 44% have formal policies to manage them. This means most companies are using dangerous tools without proper safety rules.
Agentic AI is different from regular computer programs. Normal software follows exact instructions, like a recipe. But agentic AI acts more like a digital worker who can think and make decisions on their own. This independence makes them incredibly useful for complex tasks, but it also creates new risks that most companies don't understand.
Joel Rennich from JumpCloud explained the main problem: "Agentic AI's ability to act independently is its biggest risk". Without proper controls, these AI agents can cause serious problems before anyone notices. They might delete important data, make costly financial mistakes, or break legal rules.
The solution, according to experts, is identity-first governance. This means treating AI agents like digital employees. Each agent should have a unique identity and be watched constantly, just like human workers. Companies need to know exactly what each AI agent is doing at all times.
Transparency emerged as another major concern this week. Technology leaders report that 64% cite governance, trust, and safety as their top worries when using AI agents. The problem is that these AI systems often work like "black boxes" - nobody can see how they make decisions.
This lack of transparency creates several dangerous situations. Cascading hallucination happens when an AI agent creates false information and then believes its own lies. For example, if an AI agent books vacations for customers, it might make many wrong bookings before anyone catches the mistake. Agent poisoning is another risk where bad actors feed false information to AI agents to make them work incorrectly.
In the United States, especially California, lawmakers are working on new rules for AI agents. Governor Gavin Newsom is considering signing laws that would require companies to tell workers when AI systems make decisions about their jobs. These rules would apply to hiring, firing, and performance reviews. Companies would need to give workers 30 days' notice before using AI for these important decisions.
These regulatory changes show that the "era of unchecked AI development is drawing to a close". Financial markets are already feeling the effects, with increased compliance costs and demands for more transparent AI systems. Companies that can't explain how their AI agents make decisions may face legal problems and lose customer trust.
Scientists in China developed a promising solution called SciGuard. This system works like a security guard for AI agents used in chemistry research. When someone asks an AI agent a question about chemicals, SciGuard checks if the request is safe before allowing an answer. If someone tries to ask how to make dangerous substances, SciGuard refuses to help. But if the question is legitimate, like asking about safe lab procedures, it provides detailed scientific information.
SciGuard represents a new approach called agent-based safeguards. Instead of changing the AI models themselves, which can reduce their usefulness, these systems add a smart security layer on top. This allows AI agents to remain powerful for good purposes while preventing misuse for harmful activities.
Experts recommend several key strategies for managing AI agents safely. Human-in-the-loop systems keep people involved in important decisions rather than letting AI agents work completely alone. Multi-agent architecture involves designing systems where multiple AI agents work together with clear rules about what each can do. AgentOps refers to new tools for monitoring AI agents throughout their entire lifecycle.
The week's developments show that AI agent safety requires urgent attention from companies, governments, and researchers. As one expert noted, "AI has transformative potential for science, yet with that power comes serious risks when it is misused". The goal is not to stop AI agent development but to ensure these powerful tools remain aligned with human values and safety requirements.
Companies that ignore these safety concerns may face significant consequences, including hefty fines, legal liability, and loss of customer trust. The European Union's AI Act, which became effective in 2024, already imposes penalties up to €35 million for companies that misuse AI systems. Similar regulations are likely coming to other countries as governments work to balance innovation with public safety.