AI Is Fixing Accents in Call Centers... and That Should Make Us Pause


I was recently invited to speak at Congress about new AI laws. I did not expect that one short comment I made would upset so many people. I mentioned that in the future, AI could change how our voices sound. One moment you sound Bisaya, the next you sound like someone from BGC. I said it casually. But online, many people took it seriously. Some were angry. Some were hurt.

I understand why.

In the Philippines, accents are personal. They tell people where you grew up. They tell your story. We joke about them. We tease each other. But we also carry pride in how we speak. So when people hear “accent neutralization,” it can feel like erasing who we are.

Still, this is a good time to talk about what this technology really is.

Voice accent neutralization is an AI tool that changes how a person sounds. It does not change what you say. It changes how you say it. The goal is simple: to make speech easier to understand for the listener. In call centers, that listener is usually in the U.S. or Canada. About 70% of calls go there. So “neutral” means closer to how Americans or Canadians speak.

This technology is already being used today.

Why do call centers use it? The answer is not mysterious. It helps business. When customers understand agents more easily, calls are faster. There are fewer repeats. Less frustration. That means lower costs and better customer scores.

It also changes training. If you have worked in a call center, you know how much time is spent on voice training. Agents practice vowels. They practice tone. They practice sounding “right.” With AI, much of that work is done by software. Agents can focus on the problem, not the accent.

Some people say this is good for workers. And in some ways, it is. People who were once rejected because of their accent may now get hired. The AI does the adjusting, not the person. Some companies even say this helps people with speech disabilities. It removes barriers.

But there is another side.

Many critics ask a hard question. Why is the “neutral” accent always Western? Why does neutral sound American or Canadian, and not Filipino? If the standard is always Western, then what are we really saying about our own voices?

Some people feel this is a kind of cultural erasure. Others call it dehumanizing. The idea is that if AI must “fix” your accent, then your natural voice is treated like a defect. Something to hide. Something to correct.

As a Filipino, I feel this tension deeply.

Call center work is a huge part of our economy. Millions of people depend on it. And the truth is, this industry has always required adaptation. Long before AI, workers were already asked to change how they speak. AI did not start this trend. It just made it faster and cheaper.

So the real question is not whether this technology will exist. It already does. The real question is how it will be used.

Will workers have a choice? Will they be told when their voice is being changed? Will this technology protect them from bias, or quietly reinforce it? These questions matter.

When I listen to call center workers outside of work, they sound different. When they are on the phone, they sound “neutral.” That split already exists. AI just makes it more obvious.

Voice accent neutralization is not purely good or bad. It is a tool. Like many tools, it reflects the values of the system using it.

The challenge for us is to make sure that while we improve communication, we do not lose respect for identity. We should ask not just what sounds clear, but what sounds fair.

And maybe the most important question is this: in a world where AI can change how we sound, who gets to decide what a “normal” voice really is?

 

 

Links:

Krisp Demo
https://www.facebook.com/reel/1024438729839105

Sanas Demo
https://www.youtube.com/watch?v=kNjOEqhwBRw

Krisp VA
https://www.youtube.com/shorts/Ia6ZPwgkMD4

Real Time Talking Avatar
https://www.researchgate.net/publication/289118430_Real-Time_Talking_Avatar_on_the_Internet_Using_Kinect_and_Voice_Conversion

Speech Formats of Vowels
https://corpus.eduhk.hk/english_pronunciation/index.php/2-2-formants-of-vowels/

 

About Me:

Dominic “Doc” Ligot is one of the leading voices in AI in the Philippines. Doc has been extensively cited in local and global media outlets including The Economist, South China Morning Post, Washington Post, and Agence France Presse. His award-winning work has been recognized and published by prestigious organizations such as NASA, Data.org, Digital Public Goods Alliance, the Group on Earth Observations (GEO), the United Nations Development Programme (UNDP), the World Health Organization (WHO), and UNICEF.

If you need guidance or training in maximizing AI for your career or business, reach out to Doc via https://docligot.com.

Follow Doc Ligot on Facebook: https://facebook.com/docligotAI