logo
The Impending AI Shift In The Global Call Center Industry

The Impending AI Shift In The Global Call Center Industry

Forbes25-06-2025
Dr. Brian Scott Glassman is a Thought Leader in Generative AI and Innovation and VP of Product Management at AInspire.ai
Imagine in the near future, you call a customer service number for a cruise line and are greeted by a highly knowledgeable, smooth-talking artificial intelligence (AI). In a fluid conversation, the AI eloquently explains the service you're interested in, eases your concerns, securely processes your payment and books your reservation. There are no initial wait times, no pauses to look up information, no background noise, just a streamlined interaction.
In a second scenario, a human call center agent greets your call, supported by an AI companion on the line. You describe your preferences, and the AI transcribes them into on-screen notes for the agent. Next, relevant cruise options are found by the AI, each presented to the agent with an AI-generated sales pitch. As the call progresses, the AI provides the agent with real-time sales strategies to help close the sale.
As futuristic as both these scenarios may sound, they are entirely feasible with today's AI technology. Call center experts globally agree that AI will significantly disrupt call center operations for the better, yet opinions vary greatly regarding the pace, scale and exact nature of this transformation. Given this uncertainty amid rapid technological change, I am offering my perspective on how AI could impact the call center landscape from the ground up, helping organizational leaders understand the coming changes and plan effectively.
Industry Impact
The global call center industry was valued at approximately $350 billion in 2024. Projections suggest that the impact of AI on this sector could reach $2 to $3 billion in 2025, which represents about 1% of the total market value. With all the hype around AI, this level of industry impact feels underwhelming.
However, I believe that the advancements in generative AI in 2025 represent a critical turning point, as AI performance now enables two major shifts: first, the replacement of human agents with AI call agents; and second, the enhancement of human agents through AI-driven insights and knowledge. Let me explain how I see both these solutions working.
Shifting Outsourced Call Centers To AI Voice Solutions
AI-powered customer service chat solutions have gained popularity, with many customers from younger generations now preferring them over human agents due to their immediate responses and broad knowledge.
In the last few months, a performance inflection point arrived when leading AI companies Meta, Google and OpenAI released high-performance, multimodal AI voice chat models. Newer models often distinguish themselves by delivering near-instant responses with low latency, demonstrating high intelligence, quickly accessing and interpreting documentation, recognizing basic voice inflection, and supporting multiple languages. I see this recent leap in capability now puts multi-modal AI voice solutions in direct competition with offshore and nearshore call centers.
Interestingly, I estimate the operating cost of a core AI voice model per hour of conversation to be about one-third to one-half the hourly rate of human agents based in places like the Philippines, though this estimate does not include setup and IT oversight fees. Nonetheless, AI compute costs are expected to decrease as new more efficient data centers come online and as AI models become more optimized.
Of course, early implementations of AI voice technology will have limitations and downsides. Initial deployments may lead to callers complaining about slight delays in the AI's responses, known as latency, and expressing surprise when they realize they are interacting with an AI agent instead of a human, as they had expected. Naturally, early AI voice systems will need a fallback to a human agent, which still means there will be some costs associated with human labor.
Enhancing Human Call Agents With AI Capabilities
Recent AI advancements have unlocked powerful features designed to augment, not replace, human agents. I broke these capabilities into four areas: call audio enhancement, just-in-time knowledge delivery, quality control and feedback, and call coaching.
For the first area, background noise can now be suppressed in real time using adaptive audio filters for both the customer and the agent. Additionally, AI can be used to "neutralize" accents, creating a more uniform speaking style, as reported by Mashable. With audio enhancement advances, the focus can shift toward the discussion at hand.
For the second area, I envision a human agent with a real-time AI companion that listens to conversations and offers real-time support via just-in-time knowledge and tips. Here the AI can display the relevant products or support documentation, summarize company policies, suggest appropriate products or services and recommend custom solutions for resolving customer issues or optimizing sales tactics. This alone represents a significant leap in the effectiveness for agents. One downside will be the training required to help agents interact with their AI companions effectively, adding additional ramp-up time for agents.
For the third area, AI can offer post-call deep insights. By listening to the conversation recording, the AI can write summaries, identify remaining tasks, evaluate and grade the human agent's performance, analyze the customer's needs, give insights into product demand and run company policy or compliance checks. These insights will provide agents with a feedback loop to improve their skills, give supervisors a clear view of agent performance and offer operational leaders valuable information about market demand. However, call centers may object to the level of monitoring that AI enables, and some lawmakers are already proposing legislation that could restrict the use of AI in hiring, promotion or termination decisions.
Finally, AI can coach human agents to improve call outcomes. Imagine a human agent receiving a post-call analysis email explaining what specifically they did well and pointing out specific examples of where they could improve, along with training exercises. On-the-job coaching is generally a luxury, but with AI it can become routine.
I anticipate that the transition from offshore and nearshore call centers to AI voice solutions will begin gradually in 2025, and accelerate through 2026 due to required internal testings and buy-ins. In all, the global call center industry is poised for a dramatic shift in the coming years, and the world will be on the line to hear about the results.
Forbes Business Council is the foremost growth and networking organization for business owners and leaders. Do I qualify?
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Anthropic, seeing voracious demand for shares, is clamping down on a certain kind of investment
Anthropic, seeing voracious demand for shares, is clamping down on a certain kind of investment

Business Insider

time15 minutes ago

  • Business Insider

Anthropic, seeing voracious demand for shares, is clamping down on a certain kind of investment

Anthropic has told investors it is displeased by the prevalence of a popular kind of investment vehicle being marketed to those eager to get in on the AI boom, according to two sources with knowledge of the matter. In its latest funding round, the AI startup is raising around $5 billion at a $170 billion valuation, Business Insider has confirmed. It told one of its largest investors, Menlo Ventures, that the venture capital firm must use its own capital and not resort to a special purpose vehicle, or SPV, as it did in a previous funding round, one of the people told BI. The clampdown is a sign the AI startup has more leverage to dictate terms due to rampant demand for its private shares, according to the sources who asked not to be identified because they were discussing nonpublic information. Spokespeople for Anthropic and Menlo declined to comment. Less desirable SPVs let investors pool their funds for a single, one-off deal. While they allow speed and can be marketed to a wider range of non-institutional investors, SPVs are generally seen as less desirable because most companies prefer to have a direct relationship with investors. With so many people eager to invest in a limited number of AI companies, investors have been sharing stories on social media about being marketed SPVs with extremely high fees. "Many friends including myself have been offered allocation into OpenAI or Anthropic SPVs this week," Michelle Lim, a founder and angel investor, posted last week on X. "Minimum check sizes are $100k-$1M, with fees as high as 16%. From what I understand, folks are creating SPVs on top of SPVs and making management fees on top of them. It's like a pyramid." "The feeding frenzy for ownership in the AI labs has spawned a set of bottom feeding multi layered SPV brokers that have no relationship with the company, and straight up grifters," venture capitalist Sarah Guo wrote recently on X. "Careful of that nonsense." Guo was tweeting generally, not about Anthropic's current fundraising efforts. Still, this shows how SPVs can be considered problematic, even though they're a common tool in the venture capital toolkit. Anthropic previously allowed SPVs When Anthropic raised funding in late 2023, Menlo Ventures was the lead investor and stepped forward with a $500 million check. There was a catch, though: Half a billion dollars was too rich a sum for Menlo to use all its own capital. So part of the money came from an SPV where a pool of cash from outside investors helped fund the investment. The SPV, revealed in regulatory filings, was an effective way for Anthropic to quickly raise a lot of money to keep up with OpenAI and other competitors. This time, Anthropic is being more choosy, probably because there's so much demand from investors. The startup has seen revenue soar lately, mostly thanks to its Claude models' ability to generate, fix, and deploy software code. Anthropic has told potential investors it wants tier-one VCs who will be with the company for the long haul, according to one of the sources who spoke with Business Insider. Anthropic has more leverage now Anthropic's round is five times oversubscribed, this person added, so the company has more leverage than it did in previous funding rounds. Menlo's commitment in this latest round is set to be smaller, around $120 million, according to one of the other sources. Anthropic was last valued at $61.5 billion when it raised $3.5 billion in March. Since then, investor appetite for AI has only grown more voracious. OpenAI is reportedly seeking a $500 billion valuation in an employee stock sale, making Anthropic, even valued at $170 billion, may look like a relative bargain to some investors, in comparison.

College Students Have Already Changed Forever
College Students Have Already Changed Forever

Atlantic

time26 minutes ago

  • Atlantic

College Students Have Already Changed Forever

A college senior returning to classes this fall has spent nearly their entire undergraduate career under the shadow—or in the embrace—of generative AI. ChatGPT first launched in November 2022, when that student was a freshman. As a department chair at Washington University in St. Louis, I witnessed the chaos it unleashed on campus. Students weren't sure what AI could do, or which uses were appropriate. Faculty were blindsided by how effectively ChatGPT could write papers and do homework. College, it seemed to those of us who teach it, was about to be transformed. But nobody thought it would happen this quickly. Three years later, the AI transformation is just about complete. By the spring of 2024, almost two-thirds of Harvard undergrads were drawing on the tool at least once a week. In a British survey of full-time undergraduates from December, 92 percent reported using AI in some fashion. Forty percent agreed that 'content created by generative AI would get a good grade in my subject,' and nearly one in five admitted that they've tested that idea directly, by using AI to complete their assignments. Such numbers will only rise in the year ahead. 'I cannot think that in this day and age that there is a student who is not using it,' Vasilis Theoharakis, a strategic-marketing professor at the Cranfield School of Management who has done research on AI in the classroom, told me. That's what I'm seeing in the classes that I teach and hearing from the students at my school: The technology is no longer just a curiosity or a way to cheat; it is a habit, as ubiquitous on campus as eating processed foods or scrolling social media. In the coming fall semester, this new reality will be undeniable. Higher education has been changed forever in the span of a single undergraduate career. 'It can pretty much do everything,' says Harrison Lieber, a WashU senior majoring in economics and computer science (who took a class I taught on AI last term). As a college student, he told me, he has mostly inhabited a world with ChatGPT. For those in his position, the many moral questions that AI provokes—for example, whether it is exploitative, or anti-intellectual, or ecologically unsound—take a back seat to the simple truth of its utility. Lieber characterized the matter as pragmatic above all else: Students don't want to cheat; they certainly don't want to erode the value of an education that may be costing them or their family a small fortune. But if you have seven assignments due in five days, and AI could speed up the work by tenfold for the cost of a large pizza, what are you meant to do? In spring 2023, I spoke with a WashU student whose paper had been flagged by one of the generally unreliable AI detectors that universities have used to stem the tide of cheating. He told me that he'd run his text through grammar-checking software and asked ChatGPT to improve some sentences, and that he'd done this to make time for other activities that he preferred. 'Sometimes I want to play basketball,' he said. 'Sometimes I want to work out.' His attitude might have been common among large-language-model users during that first, explosive year of AI college: If a computer helps me with my paper, then I'll have more time for other stuff. That appeal persists in 2025, but as these tools have taken over in the dorms, the motivations of their users have diversified. For Lieber, AI's allure seems more about the promise of achievement than efficiency. As with most students who are accepted to and graduate from an elite university, he and his classmates have been striving their whole life. As Lieber put it, if a course won't have 'a tangible impact on my ability to get a good job,' then 'it's not worth putting a lot of my time into.' This approach to education, coupled with a ' dismal ' outlook for postgraduate employment, justifies an ever more ferocious focus on accomplishment. Lieber is pursuing a minor in film and media studies. He has also started a profitable business while in school. Still, he had to network hard to land a good job after graduation. (He is working in risk management.) Da'Juantay Wynter, another rising senior at WashU who has never seen a full semester without AI, told me he always writes his own essays but feels okay about using ChatGPT to summarize readings, especially if he is in a rush. And like the other students I spoke with, he's often in a rush. Wynter is a double major in educational studies and American-culture studies; he has also served as president of the Association of Black Students, and been a member of a student union and various other campus committees. Those roles sometimes feel more urgent than his classwork, he explained. If he does not attend to them, events won't take place. 'I really want to polish up all my skills and intellect during college,' he said. Even as he knows that AI can't do the work as well, or in a way that will help him learn, 'it's always in the back of my mind: Well, AI can get this done in five seconds.' Another member of his class, Omar Abdelmoity, serves on the university's Academic Integrity Board, the body that adjudicates cases of cheating, with AI or otherwise. In almost every case of AI cheating he's seen, Abdelmoity told me, students really did have the time to write the paper in question—they just got stressed or preoccupied by other things, and turned to AI because it works and it is available. Students also feel the strain of soaring expectations. For those who want to go to medical school, as Abdelmoity does, even getting a 4.0 GPA and solid MCAT scores can seem insufficient for admission to the best programs. Whether or not this is realistic, students have internalized the message that they should be racking up more achievements and experience: putting in clinical hours, publishing research papers, and leading clubs, for example. In response, they seek ways to 'time shift,' Abdelmoity said, so they can fit more in. And that's at an elite private university, he continued, where the pressure is high but so is the privilege. At a state school, a student might be more likely to work multiple jobs and take care of their family. Those ordinary demands may encourage AI use even more. In the end, Abdelmoity said, academic-integrity boards such as the one he sits on can only do so much. For students who have access to AI, an education is what you make of it. If the AI takeover of higher ed is nearly complete, plenty of professors are oblivious. It isn't that they fail to understand the nature of the threat to classroom practice. But my recent interviews with colleagues have led me to believe that, on the whole, faculty simply fail to grasp the immediacy of the problem. Many seem unaware of how utterly normal AI has become for students. For them, the coming year could provide a painful revelation. Some professors I spoke with have been taking modest steps in self-defense: They're abandoning online and take-home assignments, hoping to retain the purity of their coursework. Kerri Tobin, an associate professor of education at Louisiana State University, told me that she is making undergrads do a lot more handwritten, in-class writing—a sentiment I heard many times this summer. The in-class exam, and its associated blue book, is also on the rise. And Abdelmoity reported that the grading in his natural-science courses has already been rejiggered, deemphasizing homework and making tests count for more. These adjustments might be helpful, but they also risk alienating students. Being forced to write out essays in longhand could make college feel even more old-fashioned than it did before, and less connected to contemporary life. Other professors believe that moral appeals may still have teeth. Annabel Rothschild, an assistant professor of computer science at Bard College, said she's found that blanket rules and prohibitions have been less effective than a personal address and appeal to social responsibility. Rothschild is particularly concerned about the environmental harms of AI, and she reports that students have responded to discussions about those risks. The fact that she's a scientist who understands the technology gives her message greater credibility. It also helps that she teaches at a small college with a focus on the arts. Today's seniors entered college at the tail end of the coronavirus pandemic, a crisis that once seemed likely to produce its own transformation of higher ed. The sudden switch to Zoom classes in 2020 revealed, over time, just how outmoded the standard lecture had become; it also showed that, if forced by circumstance, colleges could turn on a dime. But COVID led to little lasting change in the college classroom. Some of the students I spoke with said the response to AI has been meager too. They wondered why faculty weren't doing more to adjust teaching practices to match the fundamental changes wrought by new technologies—and potentially improve the learning experience in the process. Lieber said that he wants to learn to make arguments and communicate complex ideas, as he does in his film minor. But he also wonders why more courses can't assess those skills through classroom discussion (which is hard to fake) instead of written essays or research papers (which may be completed with AI). 'People go to a discussion-based class, and 80 percent of the class doesn't participate in discussion,' he said. The truth is that many professors would like to make this change but simply can't. A lot of us might want to judge students on the merits of their participation in class, but we've been discouraged from doing so out of fear that such evaluations will be deemed arbitrary and inequitable —and that students and their parents might complain. When professors take class participation into account, they do so carefully: Students tend to be graded on whether they show up or on the number of times they speak in class, rather than the quality of what they say. Erin McGlothlin, the vice dean of undergraduate affairs in WashU's College of Arts & Sciences, told me this stems from the belief that grading rubrics should be crystal clear in spelling out how class discussion is evaluated. For professors, this approach avoids the risk of any conflicts related to accommodating students' mental health or politics, or to bureaucratic matters. But it also makes the modern classroom more vulnerable to the incursion of AI. If what a student says in person can't be assessed rigorously, then what they type on their computer—perhaps with automated help—will matter all the more. Like the other members of his class, Lieber did experience a bit of college life before ChatGPT appeared. Even then, he said, at the very start of his freshman year, he felt alienated from some of his introductory classes. 'I would think to myself, What the hell am I doing, sitting watching this professor give the same lecture that he has given every year for the last 30 years? ' But he knew the answer even then: He was there to subsidize that professor's research. At America's research universities, teaching is a secondary job activity, at times neglected by faculty who want to devote as much time as possible to writing grants, running labs, and publishing academic papers. The classroom experience was suffering even before AI came onto the scene. Now professors face their own temptations from AI, which can enable them to get more work done, and faster, just as it does for students. I've heard from colleagues who admit to using AI-generated recommendation letters and course syllabi. Others clearly use AI to write up their research. And still more are eager to discuss the wholesome-seeming ways they have been putting the technology to use—by simulating interactions with historical authors, for example, or launching minors in applied AI. But students seem to want a deeper sort of classroom innovation. They're not looking for gimmicks—such as courses that use AI only to make boring topics seem more current. Students like Lieber, who sees his college education as a means of setting himself up for his career, are demanding something more. Instead of being required to take tests and write in-class essays, they want to do more project-based learning—with assignments that 'emulate the real world,' as Lieber put it. But designing courses of this kind, which resist AI shortcuts, would require professors to undertake new and time-consuming labor themselves. That assignment comes at the worst possible time. Universities have been under systematic attack since President Donald Trump took office in January. Funding for research has been cut, canceled, disrupted, or stymied for months. Labs have laid off workers. Degree programs have cut doctoral admissions. Multi-center research projects have been put on hold. The ' college experience ' that Americans have pursued for generations may soon be over. The existence of these stressors puts higher ed at greater risk from AI. Now professors find themselves with even more demands than they anticipated and fewer ways to get them done. The best, and perhaps the only, way out of AI's college takeover would be to embark on a redesign of classroom practice. But with so many other things to worry about, who has the time? In this way, professors face the same challenge as their students in the year ahead: A college education will be what they

AI Orchestration Matrix Maps Leading Software Providers to CX Applications
AI Orchestration Matrix Maps Leading Software Providers to CX Applications

Yahoo

time28 minutes ago

  • Yahoo

AI Orchestration Matrix Maps Leading Software Providers to CX Applications

AI is revolutionizing the CX industry with massive innovation. Companies are seeking AI expertise to enhance CX and ROI. Key opportunities lie in AI applications and software provider capabilities that drive superior CX delivery and transformation. Dublin, Aug. 18, 2025 (GLOBE NEWSWIRE) -- The "AI Orchestration in CX - Powering Innovation & Transformation" report has been added to impact and momentum of artificial intelligence (AI) on the customer experience (CX) industry are profound. It's a rapidly changing environment, enabling a massive wave of innovation across business and industry. Businesses are actively seeking AI technology expertise and solutions that can enhance their customer experience (CX) capabilities and improve their return on investment (ROI). This research provides a more comprehensive understanding of how AI is transforming the CX landscape, identifying the key AI applications that now drive the delivery of better CX, and profiles the companies and capabilities that are enabling this timely Objectives Outline and discuss the key AI applications that are helping drive massive innovation in the CX industry. Identify and profile software providers who develop and market these AI applications. Present an AI orchestration in a CX matrix that maps the capabilities of these software providers to these key AI applications. Key Topics Covered: AI in CX Applications and Profiled CompaniesAI Orchestration of CX Agentic AI Agentic AI for CX Agentic AI Capabilities Matrix AI Analytics AI Analytics for CX AI Analytics Capabilities Matrix AI Studio AI Studio for CX AI Studio Capabilities Matrix Conversational AI Conversational AI for CX Conversational AI Capabilities Matrix Intelligent Virtual Assistants IVAs for CX IVA Capabilities Matrix Knowledge Management Knowledge Management for CX Knowledge Management Capabilities Matrix Quality Assurance & Management Quality Assurance & Management for CX Quality Assurance & Management Capabilities Matrix Security & Authentication Security & Authentication for CX Security & Authentication Capabilities Matrix Social Media Management Social Media Management for CX Social Media Management Capabilities Matrix Training & Performance Management Training & Performance Management for CX Training & Performance Management Capabilities Matrix Workforce Engagement Management Workforce Engagement Management for CX Workforce Engagement Management Capabilities Matrix AI Orchestration in CX Large Language Model Usage for AI Orchestration of CX* Cloud Infrastructure for AI Orchestration in CX* Company Profiles Ada AmplifAI Assembled Avaamo Calabrio Centrical Cognigy Content Guru Coveo Diabolocom Engageware Enghouse Interactive Five9 Forethought Genesys Mitel NiCE Puzzel Quiq SoundHound Sprinklr Talkdesk Vonage YellowAI Appendix Growth Opportunities Fuel the Growth Pipeline Engine Why Is It Increasingly Difficult to Grow? The Strategic Imperative For more information about this report visit About is the world's leading source for international market research reports and market data. We provide you with the latest data on international and regional markets, key industries, the top companies, new products and the latest trends. CONTACT: CONTACT: Laura Wood,Senior Press Manager press@ For E.S.T Office Hours Call 1-917-300-0470 For U.S./ CAN Toll Free Call 1-800-526-8630 For GMT Office Hours Call +353-1-416-8900Sign in to access your portfolio

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store