
The GenAI Revolution: Balancing Innovation With Responsibility
Sandy Anderson, Experian's executive vice president of the data office and operations.
Generative AI (GenAI) is a disruptive and transformative force for companies across industries, reshaping how businesses operate and innovate. It's no different for those in the financial services community, where GenAI is more than a breakthrough technology—it's a game-changer.
Recently, the rise of assistants powered by agentic AI has revolutionized how the broader financial services community explores vast datasets, drives productivity across teams and ultimately, delivers more value to current and prospective customers. It has completely transformed the landscape, turning what once took months into a matter of days or even hours, allowing organizations to develop complex analytical models at lightning speed.
Despite these advancements and the meteoric rise of GenAI, the truth is we've only begun scratching the surface of possibility. Future applications of GenAI could lead to improved regulatory compliance and an enhanced customer experience, alongside continued innovation in product development. But as the financial services community continues down the path of unlocking the potential of GenAI, the one constant must remain: The technology has to be used responsibly and ethically.
Operational efficiency and product innovation are paramount to the success of any company, but for the financial services community, the obligation to the consumer comes first. While the promise of gen AI can take the financial services community to the next level, its adoption comes with inherent risks—data privacy concerns, data security and appropriate data usage. The responsible and ethical use of GenAI, and AI in general, is not just a regulatory requirement; it's a business imperative.
People rely on the financial services community for banking services, access to loans and credit and wealth management needs, among a wide range of services, and as a result, organizations have access to sensitive consumer information. Customers need to trust that their data is being protected, used responsibly and that any data-driven decisions are made without bias. A financial organization's failure to live up to that trust could jeopardize an institution's relationships with its customers.
That's why we view the responsible and ethical use of AI as a cornerstone of our approach. In fact, our governance frameworks are founded on intentional and deliberate choices, empowering us to simultaneously innovate and build trust with the broader community.
The use of AI and GenAI must be built on a foundation of transparency, fairness and accountability. Although the implementation of the technology can lead to a more inclusive financial ecosystem, potential bias can lead us in the opposite direction, resulting in the unfair treatment of certain consumer groups.
To address potential challenges, every member of the financial services community, including banks, credit reporting agencies, credit unions, fintechs and nonprofits, among others, must be open and honest about how they use AI, as well as the outcomes. That means implementing rigorous testing and validation processes to ensure their systems are free from bias and providing clear explanations of AI-driven decisions and allowing customers to challenge those decisions if necessary. Investing in explainable AI helps financial institutions harness the power of AI while building and maintaining consumer trust.
Beyond the models themselves, industry collaboration can create an opportunity to ensure transparency on the front end of development. Engaging with various external stakeholders, including regulators and other industry experts, organizations can pressure test their AI practices against societal values and expectations, and identify potential issues and foster a culture of accountability.
We've witnessed first-hand how collaboration with clients and industry leaders can shape the future of AI and allow us to contribute to the standards that help promote transparency, fairness and accountability.
As we look to the future, the promise of GenAI in the financial services industry is immense. However, balancing the innovation and potential with responsible and ethical use will be the key to unlocking its full potential. Embedding responsible and ethical considerations within their AI strategies will empower the financial services community to not only exceed regulatory compliance but also meet the expectations of their customers.
Forbes Business Development Council is an invitation-only community for sales and biz dev executives. Do I qualify?

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
an hour ago
- Yahoo
China's AI lab unveils RoboBrain 2.0 model to accelerate humanoid robot development
In a move that will further assert China's bid to scale robotics industry, the Beijing Academy of Artificial Intelligence (BAAI)—a not for profit research laboratory—unveiled last week a series of new open-source artificial intelligence (AI) models, dubbed RoboBrain 2.0, that will function as the 'brain' of robots. According to BAAI head Wang Zhongyuan, the use of powerful AI models in China's booming robotics market could accelerate the development and adoption of humanoids, as the sector works to overcome key challenges such as limited model capabilities and a lack of high-quality training data. Wang further explained that BAAI is actively seeking collaboration across the embodied intelligence industry, emphasizing the importance of joint efforts to accelerate progress. He noted that the institute is working with more than 20 leading companies in the sector and is looking to expand its network of partners to drive continued growth. Unveiled as part of China's broader push to advance intelligent machines, RoboBrain 2.0 was described by Wang as the world's most powerful open-source AI model designed to enhance a wide range of robots, including humanoids. Its debut positions BAAI as a potential key player in the evolving sector, the South China Morning Post reported. Furthermore, RoboBrain 2.0 introduces major improvements in spatial intelligence and task planning, delivering a 17% boost in speed and a 74% increase in accuracy compared to the previous version launched just three months earlier. With enhanced spatial intelligence, robots can now perceive distances from surrounding objects more precisely, while advanced task planning enables them to autonomously deconstruct complex activities into manageable steps, significantly improving overall performance. The RoboBrain model is part of the Wujie series, which also includes RoboOS 2.0—a cloud platform for deploying robotics AI models—and Emu3, a multimodal system capable of interpreting and generating text, images, and video. BAAI is one of China's early developers of open-source large language models, the technology behind generative AI chatbots. Several former employees have used their experience at BAAI to start their own AI companies, helping to grow the AI startup community in China. China's push to lead in robotics AI involves multiple players, with BAAI joined by the Beijing Humanoid Robot Innovation Centre, which earlier this year launched Hui Si Kai Wu—a general-purpose embodied AI platform. The center is also known for developing the Tien Kung humanoid robot, which made headlines after completing a half-marathon in Beijing in April. The Chinese institution aims to have its platform become the "Android of humanoid robots," serving as a standard operating system much like Google's Android does in the smartphone industry. Moreover, this year's edition of the BAAI Conference attracted over 100 AI researchers from around the world and more than 200 industry experts, including leaders from major Chinese tech companies such as Baidu, Huawei Technologies, and Tencent Holdings. Additionally, the Chinese academy also announced a strategic partnership with the Hong Kong Investment Corporation to collaborate on talent development, technology advancement, and capital investment aimed at fostering innovation and entrepreneurship in the country's AI sector.
Yahoo
2 hours ago
- Yahoo
Klarna CEO warns AI may cause a recession as the technology comes for white-collar jobs
The CEO of payments company Klarna has warned that AI could lead to job cuts and a recession. Sebastian Siemiatkowski said he believed AI would increasingly replace white-collar jobs. Klarna previously said its AI assistant was doing the work of 700 full-time customer service agents. The CEO of the Swedish payments company Klarna says that the rise of artificial intelligence could lead to a recession as the technology replaces white-collar jobs. Speaking on The Times Tech podcast, Sebastian Siemiatkowski said there would be "an implication for white-collar jobs," which he said "usually leads to at least a recession in the short term." "Unfortunately, I don't see how we could avoid that, with what's happening from a technology perspective," he continued. Siemiatkowski, who has long been candid about his belief that AI will come for human jobs, added that AI had played a key role in "efficiency gains" at Klarna and that the firm's workforce had shrunk from about 5,500 to 3,000 people in the last two years as a result. It's not the first time the exec and Klarna have made headlines along these lines. In February 2024, Klarna boasted that its OpenAI-powered AI assistant was doing the work of 700 full-time customer service agents. The company, most famous for its "buy now, pay later" service, was one of the first firms to partner with Sam Altman's company. Later that year, Siemiatkowski told Bloomberg TV that he believed AI was already capable of doing "all of the jobs" that humans do and that Klarna had enacted a hiring freeze since 2023 as it looked to slim down and focus on adopting the technology. However, Siemiatkowski has since dialed back his all-in stance on AI, telling an audience at the firm's Stockholm headquarters in May that his AI-driven customer service cost-cutting efforts had gone too far and that Klarna was planning to now recruit, according to Bloomberg. "From a brand perspective, a company perspective, I just think it's so critical that you are clear to your customer that there will be always a human if you want," he said. In the interview with The Times, Siemiatkowski said he felt that many people in the tech industry, particularly CEOs, tended to "downplay the consequences of AI on jobs, white-collar jobs in particular." "I don't want to be one of them," he said. "I want to be honest, I want to be fair, and I want to tell what I see so that society can start taking preparations." Some of the top leaders in AI, however, have been ringing the alarm lately, too. Anthropic's leadership has been particularly outspoken about the threat AI poses to the human labor market. The company's CEO, Dario Amodei, recently said that AI may eliminate 50% of entry-level white-collar jobs within the next five years. "We, as the producers of this technology, have a duty and an obligation to be honest about what is coming," Amodei said. "I don't think this is on people's radar." Similarly, his colleague, Mike Krieger, Anthropic's chief product officer, said he is hesitant to hire entry-level software engineers over more experienced ones who can also leverage AI tools. The silver lining is that AI also brings the promise of better and more fulfilling work, Krieger said. Humans, he said, should focus on "coming up with the right ideas, doing the right user interaction design, figuring out how to delegate work correctly, and then figuring out how to review things at scale — and that's probably some combination of maybe a comeback of some static analysis or maybe AI-driven analysis tools of what was actually produced." Read the original article on Business Insider Sign in to access your portfolio
Yahoo
3 hours ago
- Yahoo
‘Significant challenges' in use of AI within UK screen sector
The use of artificial intelligence (AI) within the UK screen sector raises 'significant legal, ethical, and practical challenges' such as the use of copyrighted material being used without the permission of the rights holders, a report has warned. Other issues highlighted by the British Film Institute (BFI) report include the safeguarding of human creative control, the fear of jobs being lost as positions are replaced through the use of AI, and investment in training in new skills. High energy consumption and carbon emissions, and the risks to creative content around biased data, are also described as being of concern. The report, which has been carried out in partnership with CoSTAR universities Goldsmiths, Loughborough and Edinburgh, analyses how the screen sector is using and experimenting with rapidly evolving generative AI technologies. It warned that the 'primary issue' was the use of copyrighted material – such as hundreds of thousands of film and TV scripts – in the training of generative AI models, without payment or the permission of rights-holders. 'This practice threatens the fundamental economics of the screen sector if it devalues intellectual property creation and squeezes out original creators,' the report said. But it added that the UK's strong foundation in creative technology – as it is home to more than 13,000 creative technology companies – means that the UK screen sector is well positioned to adapt to the technological shift. The report – titled AI in the Screen Sector: Perspectives and Paths Forward – said generative AI promises to democratise and revolutionise the industry, with the BBC, for example, piloting AI initiatives. Meanwhile, projects such as the Charismatic consortium, which is backed by Channel 4 and Aardman Animations, aim to make AI tools accessible to creators regardless of their budget or experience. It said this could empower a new wave of British creators to produce high-quality content with modest resources, though concerns about copyright and ethical use remain significant barriers to full adoption. The report sets out nine key recommendations it suggests should be addressed within the next three years to enable the UK screen sector to thrive in using AI. These include establishing the UK as a world-leading market of IP licensing for AI training, and embedding sustainability standards to reduce AI's carbon footprint. It also calls for structures and interventions to pool knowledge, develop workforce skills and target investments in the UK's creative technology sector, while it urges support for independent creators through accessible tools, funding and ethical AI products. The BFI's director of research and innovation, Rishi Coupland, said: 'AI has long been an established part of the screen sector's creative toolkit, most recently seen in the post-production of the Oscar-winning The Brutalist, and its rapid advancement is attracting multimillion investments in technology innovator applications. 'However, our report comes at a critical time and shows how generative AI presents an inflection point for the sector and, as a sector, we need to act quickly on a number of key strategic fronts. 'Whilst it offers significant opportunities for the screen sector such as speeding up production workflows, democratising content creation and empowering new voices, it could also erode traditional business models, displace skilled workers, and undermine public trust in screen content. 'The report's recommendations provide a roadmap to how we can ensure that the UK's world-leading film, TV, video games and VFX industries continue to thrive by making best use of AI technologies to bring their creativity, innovations and storytelling to screens around the globe.' Professor Jonny Freeman, director of CoSTAR Foresight Lab, said: 'This latest CoSTAR Foresight Lab report, prepared by the BFI, navigates the complex landscape of AI in the screen sector by carefully weighing both its transformative opportunities and the significant challenges it presents. 'The report acknowledges that while AI offers powerful tools to enhance creativity, efficiency, and competitiveness across every stage of the production workflow – from script development and pre-production planning, through on-set production, to post-production and distribution – it also raises urgent questions around skills, workforce adaptation, ethics, and sector sustainability.' CoSTAR is a £75.6 million national network of laboratories that are developing new technology to maintain the UK's world-leading position in gaming, TV, film, performance, and digital entertainment. Last month stars including Sir Elton John, Sir Paul McCartney and Sir Ian McKellen, wrote a joint letter to Sir Keir Starmer, urging the Prime Minister to introduce safeguards against work being plundered for free. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data