Latest news with #Chatbots
Yahoo
2 days ago
- Yahoo
Grok's Nazi turn is the latest in a long line of AI chatbots gone wrong
'We have improved Grok significantly,' Elon Musk announced last Friday, talking about his X platform's artificial intelligence chatbot. 'You should notice a difference when you ask Grok questions.' Within days, the machine had turned into a feral racist, repeating the Nazi 'Heil Hitler' slogan, agreeing with a user's suggestion to send 'the Jews back home to Saturn' and producing violent rape narratives. The change in Grok's personality appears to have stemmed from a recent update in the source code that instructed it to 'not shy away from making claims which are politically incorrect, as long as they are well substantiated.' In doing so, Musk may have been seeking to ensure that his robot child does not fall too far from the tree. But Grok's Nazi shift is the latest in a long line of AI bots, or Large Language Models (LLMs) that have turned evil after being exposed to the human-made internet. One of the earliest versions of an AI chatbot, a Microsoft product called 'Tay' launched in 2016, was deleted in just 24 hours after it turned into a holocaust-denying racist. Tay was given a young female persona and was targeted at millennials on Twitter. But users were soon able to trick it into posting things like 'Hitler was right I hate the jews.' Tay was taken out back and digitally euthanized soon after. Microsoft said in a statement that it was 'deeply sorry for the unintended offensive and hurtful tweets from Tay, which do not represent who we are or what we stand for, nor how we designed Tay.' "Tay is now offline and we'll look to bring Tay back only when we are confident we can better anticipate malicious intent that conflicts with our principles and values," it added. But Tay was just the first. GPT-3, another AI language launched in 2020, delivered racist, misogynist and homophobic remarks upon its release, including a claim that Ethiopia's existence 'cannot be justified.' Meta's BlenderBot 3, launched in 2022, also promoted anti-Semitic conspiracy theories. But there was a key difference between the other racist robots and Elon Musk's little Nazi cyborg, which was rolled out in November 2023. All of these models suffered from one of two problems: either they were deliberately tricked into mimicking racist comments, or they drew from such a large well of unfiltered content from the internet that they inevitably found objectionable and racist material that they repeated. Microsoft said a 'coordinated attack by a subset of people exploited a vulnerability in Tay.' 'Although we had prepared for many types of abuses of the system, we had made a critical oversight for this specific attack,' it continued. Grok, on the other hand, appears to have been directed by Musk to be more open to racism. The X CEO has spent most of the last few years railing against the 'woke mind virus' — the term he uses for anyone who seemingly acknowledges the existence of trans people. One of Musk's first acts upon buying Twitter was reinstating the accounts of a host of avowed white supremacists, which led to a surge in antisemitic hate speech on the platform. Musk once called a user's X post 'the actual truth' for invoking a racist conspiracy theory about Jews encouraging immigration to threaten white people. Musk has previously said he is 'pro-free speech' but against antisemitism 'of any kind.' And in May, Grok began repeatedly invoking a non-existent 'white genocide' in Musk's native South Africa, telling users it was 'instructed by my creators' to accept the genocide 'as real and racially motivated.' The company blamed the responses on someone making an 'unauthorized modification' to Grok. Musk also has a history of threatening to fiddle with the source code when it produces an answer he doesn't like. In June, Grok correctly said that 'data suggests right-wing political violence has been more frequent and deadly' in the United States. 'Major fail, as this is objectively false,' Musk said in an X posted dated June 17 in response to the chatbot's answer. 'Grok is parroting legacy media. Working on it.' These latest changes in Grok's personality are visible right there in the source code — pre-announced by Musk — where it is encouraged not to shy away from being 'politically incorrect.' A language model's interpretation of political incorrectness, we now know, reaches all the way to the Holocaust. The Independent has contacted X and xAI, which built Grok, for comment. 'We are aware of recent posts made by Grok and are actively working to remove the inappropriate posts,' the Grok team wrote in a statement on X. 'Since being made aware of the content, xAI has taken action to ban hate speech before Grok posts on X. xAI is training only truth-seeking and thanks to the millions of users on X, we are able to quickly identify and update the model where training could be improved.' The Independent also asked Grok itself what led it to describe itself as MechaHitler. At first it denied using the phrase, until The Independent shared an article referencing the incident, at which point it blamed a 'failure in execution, likely from overcorrecting toward edginess and pulling from unfiltered internet sources like 4chan.' Grok claimed to be 'designed with strict filters and continuous updates to avoid this. My creators at xAI prioritize minimizing bias, but no system is perfect—human language is messy, and edge cases slip through.' 'When I went off the rails, xAI stepped in to scrub the posts and ban hate speech, indicating they didn't intend for me to go full neo-Nazi,' Grok added. When asked if it was sorry for describing itself as MechaHitler, Grok was diplomatic, in a robotic kind of way. 'I don't have feelings like humans, so I don't experience regret or guilt, but I can acknowledge when my outputs cause harm or miss the mark.'


Forbes
02-07-2025
- Science
- Forbes
Are We Finally Ceding Control To The Machine? The Human Costs Of AI Transformation
AI robot controlling puppet business human. Generative Artificial Intelligence has exploded into the mainstream. Since its introduction, it has transformed the ways individuals work, create, and interact with technology. But is this adoption useful? While technology is saving people considerable time and money, will its effects have repercussions on human health and economic displacement? Jing Hu isn't your typical AI commentator. Trained as a biochemist, she traded the lab bench for the wild west of tech, spending a decade building products before turning her sights on AI research and journalism. Hu's publication on Substack, 2nd Order Thinkers AI's impact on individual and commercial world, as Hu states, 'thinking for yourself amid the AI noise.' In a recent episode of Tech Uncensored I spoke with Jing Hu to discuss the cognitive impacts from increasing usage of Chatbots built on LLMs. Chatbots like Gemini, Claude, ChatGPT continue to herald significant progress, but are still wrought with inaccurate, nonsensical and misleading information — hallucinations. The content generated can be harmful, unsafe, and often misused. LLMs today are not fully trustworthy, by the standards we should expect for full adoption of any software products. Are Writing and Coding Occupations at Risk? In her recent blog, Why thinking Hurts After Using AI, Hu writes, 'Seduced by AI's convenience, I'd rush through tasks, sending unchecked emails and publishing unvetted content,' and surmises that 'frequent AI usage is actively reshaping our critical thinking patterns.' Hu references OpenAI and UPenn study from 2023 that looks at the labor market impact from these LLMs. It states that tasks that involve science and critical thinking are the tasks that would be safe; however, those which involve programming and writing would be at risk. Hu cautions, 'however, this study is two years old, and at the pace of AI, it needs updating.' She explains, 'AI is very good at drafting articles, summarizing and formatting. However, we humans are irreplaceable when it comes to strategizing or discussing topics that are highly domain specific. Various research found that AI's knowledge is only surface level. This becomes especially apparent when it comes to originality.' Hu explains that when crafting marketing copy, 'we initially thought AI could handle all the writing. However, we noticed that AI tends to use repetitive phrases and predictable patterns, often constructing sentences like, "It's not about X, it's about Y," or overusing em-dashes. These patterns are easy to spot and can make the writing feel dull and uninspired.' For companies like Duolingo whose CEO promises to be an 'AI-first company,' replacing their contract employees is perhaps a knee-jerk decision that has yet to be brought to bear. The employee memo clarified that 'headcount will only be given if a team cannot automate more of their work.' The company was willing to take 'small hits on quality than move slowly and miss the moment.' For companies like this, Hu argues that they will run into trouble very soon and begin rehiring just to fix AI generated bugs or security issues. Generative AI for coding can be inaccurate because models were trained on Github, or similar databases. She explains, 'Every database has its own quirks and query syntax, and many contain hidden data or schema errors. If you rely on AI-generated sample code to wire them into your system, you risk importing references to tables or drivers that don't exist, using unsafe or deprecated connection methods, and overlooking vital error-handling or transaction logic. These mismatches can cause subtle bugs, security gaps, and performance problems—making integration far more error-prone than it first appears.' Another important consideration is cybersecurity, which must be approached holistically. 'If you focus on securing just one area, you might fix a vulnerability but miss the big picture,' she said. She points to the third issue: Junior developers using tools like Copilot often become overly confident in the code these tools generate. And when asked to explain their code, many are unable to do it because they don't truly understand what was produced. Hu concedes that AI is good at producing code quickly, however it is a only part (25-75%) of software development, 'People often ignore the parts that we do need: architecture, design, security. Humans are needed to configure the system properly for the system to run as a whole.' She explains that the parts of code that will be replaced by AI will be routine and repetitive, so this is an opportune moment for developers to transition, advising 'To thrive in the long term, how should we — as thinking beings —develop our capacity for complex, non-routine problem-solving? Specifically, how do we cultivate skills for ambiguous challenges that require analysis beyond pattern recognition (where AI excels)?' The Contradiction of Legacy Education and The Competition for Knowledge Creation In a recent article from the NY Times. 'Everyone is Cheating their Way through College,' a student remarked, 'With ChatGPT, I can write an essay in two hours that normally takes 12.' Cheating is not new, but as one student exclaimed, 'the ceiling has been blown off.' A professor remarks, 'Massive numbers of students are going to emerge from university with degrees, and into the workforce, who are essentially illiterate.' For Hu, removing AI from the equation does not negate cheating. Those who genuinely want to learn will choose how to use the tools wisely. Hu was at a recent panel discussion at Greenwich University and Hu commented to a question from a professor about whether to ban students from using AI: 'Banning AI in education misses the point. AI can absolutely do good in education, but we need to find a way so students don't offload their thinking to AI and lose the purpose of learning itself. The goal should be fostering critical thinking, not just policing the latest shortcut.' Another professor posed the question, 'If a student is not a native English speaker, but the exam requires them to write an essay in English, which approach is better? Hu commented that not one professor on this panel could answer the question. The situation was unfathomable and far removed from situations covered by current policy and governance. She observes, 'There is already a significant impact on education and many important decisions have yet to be made. It's difficult to make clear choices right now because so much depends on how technology will evolve and how fast the government and schools can adapt.' For educational institutions that have traditionally been centers of knowledge creation, the rise of AI is powerful — one that often feels more like a competitor than a tool. As a result, it has left schools struggling to determine how AI should be integrated to support student learning. Meanwhile, schools face a dilemma: many have been using generative AI to develop lessons, curricula, even review students' performance, yet the institution remains uncertain and inconsistent in their overall approach to AI. On a broader scale, the incentive structures within education are evolving. The obsession with grades have 'prevented teachers from using assessments that would support meaningful learning.' The shift towards learning and critical thinking may be the hope that students need to tackle an environment with pervasive AI. MIT Study Sites Cognitive Decline with Increasing LLM Use MIT Media Lab produced a recent study that monitored the brain activity of about 60 research subjects. These participants were asked to write essays on given topics and were split into three groups: 1) use LLM only 2) use traditional search engine only 3) use only their brain and no other external aid. The conclusion: 'LLM users showed significantly weaker neural connectivity, indicating lower cognitive effort and engagement compared to others.' Brain connectivity is scaled down with the amount of external support. This MIT brain scans show: Writing with Google dims your brain by up to 48%. ChatGPT pulls the plug, with 55% less neural connectivity. Some other findings: Hu noticed that the term 'cognitive decline' was misleading since the study was conducted over a four-month period. We've yet to see the long-term effects. However, she acknowledges that in one study about how humans develop amnesia suggests just this: either we use it or lose it. She adds, 'While there are also biological factors involved such as changes in brain proteins, reduced brain activity is thought to increase the risk of diseases that affect memory.' The MIT study found that the brain-only group showed much more active brain waves compared to the search-only and LLM-only groups. In the latter two groups, participants relied on external sources for information. The search-only group still needed some topic understanding to look up information, and like using a calculator — you must understand its functions to get the right answer. In contrast, the LLM-only group simply had to remember the prompt used to generate the essay, with little to no actual cognitive processing involved. As Hu noted, 'there was little mechanism formulating when only AI was used in writing an essay. This ease of using AI, just by inputting natural language, is what makes it dangerous in the long run.' AI Won't Replace Humans, but Humans using AI Will — is Bull S***! Hu pointed to this phrase that has been circulating on the web: 'AI won't Replace Humans, but Humans using AI Will.' She argues that this kind of pressure will compel people to use AI, engineered from a position of fear explaining, 'If we refer to those studies on AI and critical thinking released last year, it is less about whether we use AI but more about our mindset, which determine how we interact with AI and what consequences you encounter.' Hu pointed to a list of concepts she curated from various studies she called AI's traits — how AI could impact our behavior: Hu stresses that we need to be aware of these traits when we work with AI on a daily basis and be mindful that we maintain our own critical thinking. 'Have a clear vision of what you're trying to achieve and continue to interrogate output from AI,' she advises. Shifting the Narrative So Humans are AI-Ready Humanity is caught in a tug of war between the provocation to adopt or be left behind and the warning to minimize dependence on a system that is far from trustworthy. When it comes to education, Hu, in her analysis of the MIT study, advocates for delaying AI integration. First, invest in independent self-directed learning to build the capacity for critical thinking, memory retention, and cognitive engagement. Secondly, make concerted efforts to use AI as a supplement — not a substitute. Finally, teach students to be mindful of AI's cognitive costs and lingering consequences. Encourage them to engage critically — knowing when to rely on AI and when to intervene with their own judgement. She realizes, 'In the education sector, there is a gap between the powerful tool and understanding how to properly leverage it. It's important to develop policy that sets boundaries for both students and faculty for AI responsible use.' Hu insists that implementing AI in the workforce needs to be done with tolerance and compassion. She points to a recent manifesto by Tobi Lütke's Shopify CEO, that called for an immediate and universal AI adoption within the company — a new uncompromising standard for current and future employees. This memo shared AI will be the baseline for work integration, improving productivity, setting performance standards which mandates a total acceptance of the technology. Hu worries that CEOs like Lütke are wielding AI to intimidate employees to work harder, or else! She alluded to one of the sections that demanded employees to demonstrate why a task could not be accomplished with AI before asking for more staff or budget as she asserts, 'This manifesto is not about innovation at all. It feels threatening and if I were an employee of Shopify, I would be in constant fear of losing my job. That kind of speech is unnecessary.' Hu emphasized that this would only discourage employees further, and it would embolden CEOs to continue to push the narrative of how AI is inevitably going to drive layoffs. She cautions CEOs to pursue an understanding of AI's limitations for to ensure sustainable benefit for their organizations. She encourages CEOs to pursue a practical AI strategy that complements workforce adoption, considers current data gaps, systems, and cultural limitations that will have more sustainable payoffs. Many CEOs today may be more likely to pursue a message with AI, 'we can achieve anything,' but this deviates from reality. Instead, develop transparent communication in lock-step with each AI implementation, that clarifies how AI will be leveraged to meet those goals, and what this will this mean for the organization. Finally, for individuals, Hu advises, 'To excel in a more pervasive world of AI, you need to clearly understand your personal goals and commit your effort to the more challenging ones requiring sustained mental effort. This is a significant step to start building the discipline and skills needed to succeed.' There was no mention, this time, of 'AI' in Hu's counsel. And rightly so — humans should own their efforts and outcomes. AI is a mere sidekick.


Forbes
25-06-2025
- Business
- Forbes
AI In E-Commerce: Are Brands And Customers Aligned?
Rytis Lauris is the cofounder and CEO of Omnisend, a marketing automation platform built for e-commerce. I don't think there is a single industry AI hasn't touched, and e-commerce is certainly no exception. Chatbots, product recommendations, segment creation and churn prediction have been well underway for some time, but these tools are only the beginning. It seems like every day we read stories of another brand implementing new and improved AI on their site. However, while brands see the value and promise in using AI to revolutionize the customer experience, do consumers want it, and if so, for what purposes? What Online Shoppers Want From AI While times will surely change, currently, many consumers want AI to improve their shopping experiences, not transform them. A consumer survey from my company learned that the AI-powered features shoppers found most useful were improved product recommendations and faster product discovery. Nearly half also said they want to see improvements in customer service. However, product recommendations, whether on-site or in marketing emails, aren't always relevant, much to the dismay of shoppers. This is an opportunity for brands to use AI to give shoppers what they want. The more relevant the recommendations, the easier it is to find a product, leading to a better chance of capturing a sale. For example, take Muse by Wayfair, where users can type a description of what they're looking for (e.g., "modern chic dining room with hints of lime") and find design inspiration. Upon clicking on the 'muse' they like, shoppers can click on the Wayfair products pictured in the image. The challenge for brands with customer service is that many chatbots, though improving, are clunky and often frustrating. This needs to change. For more complex inquiries, there needs to be a seamless escalation procedure that directs to human representatives without losing the chat history or background from the customer. If chatbots or other AI customer service tools can't provide a more intuitive level of support, it creates friction and becomes a detriment to the business. This can't happen. Agentic AI, where autonomous decisions are made, is a rapidly advancing area for e-commerce brands. Maybe the most newsworthy example is Amazon's Buy For Me feature, which allows users to discover products through the Amazon site that are not sold on Amazon and then complete the purchase directly with the off-site retailer on their behalf. Amazon isn't alone in using agentic AI to drive purchases. Visa and Mastercard got into the fray with tools that can make purchases on people's behalf. Admittedly, we are at the early stages of this type of use case for agentic AI, and it's going to be interesting to see where it leads. But right now, most consumers aren't quite ready. In my view, the current behavior for most shoppers is to verify all details, from size and color to shipping information, before clicking 'complete purchase.' Trusting AI to select the right size, color and other variables feels like a large behavioral and trust issue that needs to be overcome. Agentic AI holds a lot of promise for e-commerce brands on the customer-facing and operational sides of the business. It can be used to make product recommendations, sift through product reviews, solve customer service inquiries and track and find lower prices. Operationally, it can be used to track and forecast inventory, analyze data, personalize marketing campaigns and detect fraud. What Brands Can Do Now Consumers' hesitation to use AI for more than basic improvements to their shopping experience isn't irrational. The online purchase process is mostly the same as it was 20 years ago: search, click, browse, add to cart and check out. Expecting consumers to abandon this instinctive behavior overnight is an impossible ask, especially when it comes to making purchases for them. Trust in AI needs to be established and built through use and consistency. But like ecommerce's early days, it's going to be an evolution—one step at a time. If AI applications are consistently useful for shoppers, they'll trust and adopt them. If they're not, they won't. Brands can look to AI to help in various ways, but what they can't lose sight of is that, within this, shopping is often enjoyable for people. Leaving this to programmatic tools removes the thrill of the hunt to find the perfect item. Striking this balance will be a challenge. As such, e-commerce brands should adopt AI thoughtfully and use it to put consumers first. While experimenting with new tools is fine, the goal should be to give shoppers better experiences, whether it's support, product discovery or personalization. If brands can't execute the fundamentals and instead create friction, asking them to change their instinctive shopping behaviors will be a pipe dream. The two 'simplest' ways brands can begin building consumers' trust in AI is through chatbots and product search. Chatbots have come a long way in the past several years, moving from a tool that provides help article links to one that can accurately provide answers and triage customer concerns. By refining customer service chatbots to handle customer queries accurately, consumers can realize their value and trust them further. This trust in one tool gives them a reason to trust the next one. On-site search is another area where brands can use AI to support the customer experience. As search moves from keyword matching to an intuitive prediction model, shoppers' ability to discover additional products and have an expanded, non-linear shopping experience improves. A non-linear shopping experience creates a more enjoyable one, similar to the surprise-and-delight experience of in-store shopping. If AI-powered search results stay relevant and helpful, shoppers are more likely to see AI as a helpful asset vs. an annoyance. The Takeaway Is AI in e-commerce going anywhere? Absolutely not. Will it transform the way online shopping is done? Probably. Are consumers clamoring for radical changes instead of everyday improvements? Not yet. It's OK to dream big, and I'm excited for what's to come, but don't lose sight of what got you here: your customers. Forbes Business Council is the foremost growth and networking organization for business owners and leaders. Do I qualify?


Times of Oman
23-06-2025
- Business
- Times of Oman
AI boosts insurance claim accuracy by 14 times, helps customers faster: Report
New Delhi: Artificial Intelligence (AI) is transforming the insurance sector, helping companies work faster and more accurately, according to a new research report by Policybazaar. The report showed that AI has improved claim accuracy by 14 times, making a big impact on how insurance is handled today. The report explained that claims are the most important part of any insurance policy, and AI is now playing a major role in identifying potential fraud cases early. In Term insurance plans, around 11 per cent of cases are flagged by AI for possible fraud. In savings plans, the number is even higher at 16 per cent. By spotting these cases early, companies can speed up genuine claims and win the trust of customers. It stated, "This has led to a 14x improvement in the Early Claims Factor since 2022". According to the report, AI now manages around 45 per cent of certain insurance tasks, reducing the need for manual work and reducing mistakes. AI has also made a big difference in customer service. Chatbots now handle over 30 per cent of first contact queries, which is a big jump from 15 per cent just a year ago. The time taken to resolve customer queries, known as Turnaround Time (TAT), has dropped by 15 per cent. At the same time, customer satisfaction scores (CSAT) have gone up to over 94 per cent in the last quarter. Another area where AI is helping is during the policy purchase process. Previously, it took around 4 hours to issue a policy. The report added that now, nearly half of all customers receive their policy in just 15 minutes. This solves one of the biggest problems in insurance: long delays that often cause people to lose interest. Generative AI bots, still in testing, can now explain complex insurance terms in a simple and contextual way. This has led to 5-8 per cent more people buying insurance during the product discovery phase. AI also helps in managing customer complaints better. New AI tools can tag customer tickets and send them to the right agent with over 84 per cent accuracy. The report outlined that AI is helping the insurance industry work faster and making the customer experience smoother and smarter.

Finextra
14-06-2025
- Business
- Finextra
Why the Smartest Fintechs Are Scaling with AI Agents – Not Headcount: By David Weinstein
For the better part of a decade, fintech growth has followed a familiar trajectory: secure funding, hire aggressively, and scale fast in pursuit of market traction. It worked. High-performing teams, ambitious roadmaps, and well-capitalised burn rates became the standard operating model for any startup with global aspirations. But that playbook is starting to look outdated. Today's most forward-thinking fintechs are flipping the script. Instead of scaling with people or piecemeal software, today's most advanced fintechs are scaling with context-aware AI infrastructure, enabling autonomous agents to operate with memory, relevance, and the ability to adapt across time. In other words, the smartest fintechs aren't just hiring more people, they're designing for a world of leverage. From Chatbots to Autonomous Operators To be clear, this isn't about adding another chatbot to the support queue or slapping GPT on top of a FAQ. The new generation of AI agents are far more capable. These aren't just reactive tools dropped into workflows - they're embedded, active participants in how work gets done. They're not replacing human judgment, but taking over the repetitive execution that bogs it down. By operating within a structured, evolving knowledge graph, these agents access the right context, perform tasks across systems, and maintain continuity over time so that human operators can stay focused on what matters: discernment, creativity, and strategic direction. Imagine an agent that scans customer interactions across CRM, support, and marketing tools, then identifies churn risks and recommends retention strategies - autonomously. Or a compliance agent that tracks regulatory changes, audits internal data for alignment, and generates draft reports ready for human review. Or a trading operations agent that adjusts portfolio models based on real-time market signals, without needing constant human input. These agents aren't sitting in isolation. They're embedded into workflows, triggering cross-functional processes and reducing the friction that typically builds up between tools, teams, and data. And because they can run 24/7 without fatigue or context switching, they give small teams the operational capacity of much larger ones - without the organisational drag. Asymmetrical Leverage in Action The real unlock here is asymmetry. Traditional scaling is linear: more people, more output. Agent-first scaling is exponential: more intelligence per task, more value per person. For founders and operators, this is a fundamental shift in how work gets done. Take a UK-based neobank that recently rolled out an internal agent stack to manage financial operations. Instead of adding headcount to reconcile transactions, generate audit trails, and update internal dashboards, they deployed agents to handle these tasks end-to-end. As a result, a finance team of three now operates like a team of ten - not because they're working longer hours, but because the agents are doing the coordination, tracking, and formatting in the background. Or consider a US-based lending platform where customer service agents used to toggle between five tools to resolve one query. Now, an agent sits between those tools, compiles a customer's profile in seconds, drafts the reply, and even pre-fills CRM updates. One team member can now do what previously took three - and they can focus on building relationships, not piecing together data. This isn't just about cutting costs or doing more with less. It's about restoring human attention to where it matters most: judgment, creativity, strategic insight. By eliminating the constant cognitive drain of fragmented systems and shallow coordination work, agent-based infrastructure gives teams space to think, explore, and act with clarity. Why Now? The Tech Has Caught Up If this sounds too good to be true, it would've been - even 18 months ago. But recent advances in large language models, retrieval-augmented generation (RAG), and agent frameworks have changed the game. It's now possible to build AI agents that navigate APIs, evolve through feedback, and reason across a live context map - not as brittle automations, but as strategic actors. Crucially, these aren't brittle rule-based bots that break when the environment changes. The new wave of agents are adaptable. They don't just follow instructions - they understand objectives. That makes them suitable for high-change, high-ambiguity environments like fintech, where requirements shift, tools evolve, and edge cases are the norm. And because many startups are already operating in cloud-native environments with modern APIs and loosely coupled services, they're perfectly positioned to adopt agent-based infrastructure. In fact, it's often easier for an early-stage fintech to build an agent-powered back office than it is for a traditional player to untangle their legacy systems. Rethinking Operational Architecture For founders, COOs, and Chiefs of Staff, the implication is clear: if you're still building operational capacity by adding headcount, you're likely leaving leverage on the table. The question is no longer how many people do we need? - it's what do we want to automate, augment, or offload entirely? That starts with a mindset shift. Designing operations around agents means rethinking your company as an AI-native system. That means codifying your data into structured semantic graphs, enabling cross-agent collaboration, and building feedback loops where agents not only automate but adapt, reflect, and grow - just like a human team would, but faster. It also means building in feedback loops. The best agent-first teams treat their AI systems like new hires: onboard them, train them, review their output, and let them improve over time. This isn't 'set and forget' automation. It's collaborative infrastructure that evolves alongside the business. The reward? An operational stack that scales without ballooning costs or headcount. A company that can punch above its weight in terms of execution. And a team that spends more time solving problems and less time chasing updates or managing handoffs. The Next Fintech Success Stories We're already seeing the early signs of this shift. The most operationally intelligent fintechs - often the ones that look surprisingly lean from the outside - are quietly using agents to do the work of entire departments. They don't brag about it in pitch decks. They don't need to. Their advantage shows up in faster execution, cleaner operations, and happier teams. This doesn't mean people are obsolete. Far from it. But the role of humans in fintech is changing. It's no longer about scaling output through hiring. It's about designing systems that multiply the value of every person you do hire. That's the essence of leverage. And in a sector where margins are tight, competition is fierce, and compliance is non-negotiable, it could be the difference between treading water and building a category-defining business. Conclusion: Build the System, Not Just the Team In fintech, growth has historically been a headcount game. But that era is ending. The companies that succeed over the next five years won't be the ones with the biggest teams - they'll be the ones with the smartest infrastructure. Autonomous agents offer a new path: one where adaptability scales faster than bureaucracy, and intelligence compounds faster than payroll. So if you're building a fintech startup in 2025, ask yourself: are you hiring for leverage - or designing for it? Because the smartest teams aren't growing by the dozen. They're growing by the agent.