Ethics & Safety Weekly AI News
October 27 - November 4, 2025## Weekly AI Agent Safety News Update
This week brought important conversations about keeping AI agents safe, honest, and under control. AI agents are special computer programs that can think, make decisions, and take action all by themselves. Unlike regular computer programs that just follow exact instructions, AI agents can figure out what to do based on what they learn. This makes them powerful and useful, but it also creates new problems that companies have never seen before.
One of the scariest problems experts are talking about is shadow agents. Imagine if your school hired a teacher to help for one year, but after the year ended, nobody told the teacher to leave. The teacher would just keep showing up, going into classrooms, and keeping their access card. Shadow agents work the same way. Someone creates an AI agent to do a specific job, but then nobody turns it off when the job is done. The agent keeps running and keeps having access to important computer systems and files. This is dangerous because hackers could find these forgotten agents and use them like secret doors into the company's computers.
Another big safety problem is called privilege escalation. "Privilege" is just a fancy word for permission. Escalation means it gets bigger or more powerful. When an AI agent has too much permission, it can do damage. For example, if a janitor had a key to every room in a bank, that would be a problem. The janitor should only be able to get into the hallways and closets, not the place where money is stored. AI agents need the same thing—they should only have permission to do the exact job they are supposed to do. If hackers take control of an agent with too much power, they could take over the whole company's computer system.
Data leakage is another scary problem. Data is information, like customer names, addresses, or secret plans. AI agents often move data from one computer system to another. If something goes wrong, or if someone tricks the agent, all that private information could leak out to bad people or to the internet. Some companies are worried that AI agents could even leak information to other AI agents, or could be tricked through small changes in instructions that nobody would notice.
The good news is that companies are taking these problems seriously. This week, a company called Zenity won an award for creating tools to help companies keep AI agents safe. Palo Alto Networks introduced a new product called Cortex AgentiX that helps companies build and control AI agents in a safer way. Trend Micro teamed up with NVIDIA to create protection systems for AI agents in big computer centers. These tools are like security guards for AI agents—they watch what the agents do and stop them if they try to do something dangerous.
Companies like Salesforce are also thinking hard about data privacy and trust. Salesforce has people whose job is to make sure AI agents are ethical and safe. They test their agents over and over to make sure the agents don't accidentally break the rules or hurt anybody. They want customers to trust that when they use an AI agent, their secret information will stay secret. But experts say companies need to be even more clear and honest about exactly what happens to data when AI agents are using it.
Experts agree that the best solution is identity-first security. This is a fancy way of saying: first, you need to know exactly who each AI agent is and what it is supposed to do. You need to know which person is responsible for each agent. You need to make sure each agent only gets permission to do its specific job. You need to watch what each agent does and keep records of it. And you need to be able to turn off an agent really fast if something goes wrong.
Right now, old security tools don't work very well with AI agents because AI agents behave in ways that are different from people. A person usually does the same job the same way every day. An AI agent might do things differently each day depending on what it learns. This confuses the old security tools, kind of like how a dog might be confused if someone acted like a cat. Security experts are building completely new tools designed just for AI agents.
In the coming weeks and months, companies will need to make big changes in how they think about security. They need to discover and inventory all the AI agents they have—meaning find them all and write down where they are. They need to assign clear ownership, meaning make sure one person is responsible for each agent. They need to enforce least privilege, which means giving each agent the smallest amount of permission it needs. They need to propagate identity context through multiple agents, meaning if Agent A asks Agent B to do something, Agent B should know that Agent A only has permission to ask for certain things. They need to monitor agent behavior like security guards watching cameras. They need an emergency kill switch to turn off a bad agent really fast. And they need to bring AI agents into their overall security system.
The bottom line is this: AI agents are going to be everywhere in companies soon, and they can do incredible things to help people work faster. But nobody wants AI agents running wild and causing problems. This week's news shows that smart people are working hard to make sure AI agents are safe, honest, and stay under control. Companies that start thinking about these problems now will be much safer than companies that wait.