logo
College grads are lab rats in the Great AI Experiment

College grads are lab rats in the Great AI Experiment

COMPANIES are eliminating the grunt work that used to train young professionals – and they don't seem to have a clear plan for what comes next.
Artificial intelligence (AI) is analysing documents, writing briefing notes, creating PowerPoint presentations or handling customer service queries, and – surprise! – now the younger humans who normally do that work are struggling to find jobs. Recently, the chief executive officer of AI firm Anthropic predicted that AI would wipe out half of all entry-level white-collar jobs. The reason is simple. Companies are often advised to treat ChatGPT 'like an intern', and some are doing so at the expense of human interns.
This has thrust college graduates into a painful experiment across multiple industries, but it doesn't have to be all bad. Employers must take the role of scientists, observing how AI helps and hinders their new recruits, while figuring out new ways to train them. And the young lab rats in this trial must adapt faster than the technology trying to displace them, while jumping into more advanced work.
Consulting giant KPMG, for instance, is giving graduates tax work that would previously go to staff with three years of experience. Junior staff at PwC have started pitching to clients. Hedge fund Man Group tells me its junior analysts who use AI to scour research papers now have more time to formulate and test trading ideas, what the firm calls 'higher-level work'.
I recently interviewed two young professionals about using AI in this way, and, perhaps not surprisingly, neither of them complained about it. One accountant who had just left university said he was using ChatGPT to pore over filings and Moody's Ratings reports, saving him hours on due diligence.
Another young executive at a public relations (PR) firm, who'd graduated last year from the London School of Economics, said tools such as ChatGPT had cut down her time spent tracking press coverage from two-and-a-half hours to 15 minutes, and while her predecessors would have spent four or five hours reading forums on Reddit, that now takes her only 45 minutes.
A NEWSLETTER FOR YOU
Friday, 3 pm Thrive
Money, career and life hacks to help young adults stay ahead of the curve.
Sign Up
Sign Up
I'm not convinced, however, that either of these approaches is actually helping recruits learn what they need to know. The young accountant, for instance, might be saving time, but he's also missing out on the practice of spotting something fishy in raw data. How do you learn to notice red flags if you don't dig through numbers yourself? A clean summary from AI doesn't build that neural pathway in your brain.
The PR worker also didn't seem to be doing 'higher-level work', but simply doing analysis more quickly. The output provided by AI is clearly useful to a junior worker's bosses, but I'm sceptical that it's giving them a deeper understanding of how a business or industry works.
What's worse is that their opportunities for work are declining overall. 'We've seen a huge drop in the demand for 'entry-level' talent across a number of our client sets,' says James Callander, CEO of a Freshminds, a London recruitment firm that specialises in finding staff for consultancies. An increasing number of clients want more 'work-ready' professionals who already have a first job under their belt, he adds.
That corroborates a trend flagged by venture capital firm SignalFire, whose State of Talent 2025 report pointed to what it called an 'experience paradox', where more companies post for junior roles but fill them with senior workers. The data crunchers at LinkedIn have noticed a similar trend, prompting one of its executives to claim that the bottom rung of the career ladder was breaking.
Yet some young professionals seem unfazed. Last week, a University of Oxford professor asked a group of 70 executive Master of Business Administration students from the National University of Singapore if Gen Z jobs were being disproportionately eroded by AI. Some said 'no', adding that they, younger workers, were best placed to become the most valuable people in a workplace because of their strength in manipulating AI tools, recounts Dr Alex Connock, a senior fellow at Oxford's Said Business School, who specialises in the media industry and AI.
The students weren't just using ChatGPT, but a range of tools such as Gemini, Claude, Firefly, HeyGen, Gamma, Higgsfield, Suno, Udio, NotebookLM and Midjourney, says Dr Connock.
The lesson here for businesses is that sure, in the short term you can outsource entry-level work to AI and cut costs; but that means missing out on capturing AI-native talent.
It's also dangerous to assume that giving junior staff AI tools will automatically make them more strategic. They could instead become dependent, even addicted to AI tools, and not learn business fundamentals. There are lessons here from social media. Studies show that young people who use it actively tend not to get the mental-health harms of those who use it passively. Posting and chatting on Instagram, for instance, is better than curling up on the couch and doom-scrolling for an hour.
Perhaps businesses should similarly look for healthy engagement by their newer staff with AI, checking that they're using it to sense-check their own ideas and interrogating a chatbot's answers, rather than going to it for all analysis and accepting whatever the tools spit out.
That could spell the difference between raising a workforce that can think strategically, and one that can't think beyond the output from an AI tool. BLOOMBERG

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Exclusive-OpenAI taps Google in unprecedented cloud deal despite AI rivalry, sources say
Exclusive-OpenAI taps Google in unprecedented cloud deal despite AI rivalry, sources say

CNA

time3 hours ago

  • CNA

Exclusive-OpenAI taps Google in unprecedented cloud deal despite AI rivalry, sources say

SAN FRANCISCO :OpenAI plans to add Alphabet's Google cloud service to meet its growing needs for computing capacity, three sources told Reuters, marking a surprising collaboration between two prominent competitors in the artificial intelligence sector. The deal, which has been under discussion for a few months, was finalized in May, one of the sources added. It underscores how massive computing demands to train and deploy AI models are reshaping the competitive dynamics in AI, and marks OpenAI's latest move to diversify its compute sources beyond its major supporter Microsoft, including its high-profile Stargate data center project. It is a win for Google's cloud unit, which will supply additional computing capacity to OpenAI's existing infrastructure for training and running its AI models, sources said, who requested anonymity to discuss private matters. The move also comes as OpenAI's ChatGPT poses the biggest threat to Google's dominant search business in years, with Google executives recently saying that the AI race may not be winner-take-all. OpenAI, Google and Microsoft declined to comment. Since ChatGPT burst onto the scene in late 2022, OpenAI has dealt with increasing demand for computing capacity - known in the industry as compute - for training large language models, as well as for running inference, which involves processing information so people can use these models. OpenAI said on Monday that its annualized revenue run rate surged to $10 billion as of June, positioning the company to hit its full-year target amid booming adoption of AI. Earlier this year, OpenAI partnered with SoftBank and Oracle on the $500 billion Stargate infrastructure program, and signed deals worth billions with CoreWeave for more compute. It is on track this year to finalize the design of its first in-house chip that could reduce its dependency on external hardware providers, Reuters reported in February. The partnership with Google is the latest of several maneuvers made by OpenAI to reduce its dependency on Microsoft, whose Azure cloud service had served as the ChatGPT maker's exclusive data center infrastructure provider until January. Google and OpenAI discussed an arrangement for months but were previously blocked from signing a deal due to OpenAI's lock-in with Microsoft, a source told Reuters. Microsoft and OpenAI are also in negotiations to revise the terms of their multibillion-dollar investment, including the future equity stake Microsoft will hold in OpenAI. For Google, the deal comes as the tech giant is expanding external availability of its in-house chip known as tensor processing units, or TPUs, which were historically reserved for internal use. That helped Google win customers including Big Tech player Apple as well as startups like Anthropic and Safe Superintelligence, two OpenAI competitors launched by former OpenAI leaders. Google's addition of OpenAI to its customer list shows how the tech giant has capitalized on its in-house AI technology from hardware to software to accelerate the growth of its cloud business. Google Cloud, whose $43 billion of sales comprised 12 per cent of Alphabet's 2024 revenue, has positioned itself as a neutral arbiter of computing resources in an effort to outflank Amazon and Microsoft as the cloud provider of choice for a rising legion of AI startups whose heavy infrastructure demands generate costly bills. Alphabet faces market pressure to demonstrate financial returns on its AI-related capital expenditures, which are expected to hit $75 billion this year, while maintaining its bottom line against the threat of competing AI offerings, as well as antitrust enforcement. Google's DeepMind AI unit also competes directly with OpenAI and Anthropic in a race to develop the best models and integrate those advances into consumer applications. Selling computing power reduces Google's own supply of chips while bolstering capacity-constrained rivals. The OpenAI deal will further complicate how Alphabet CEO Sundar Pichai allocates the capacity between the competing interests of Google's enterprise and consumer business segments. Google already lacked sufficient capacity to meet its cloud customers' demands as of the last quarter, Chief Financial Officer Anat Ashkenazi told analysts in April. Although ChatGPT holds a large lead over Google's competing chatbot in terms of monthly users and analysts have predicted it could reduce Google's dominant search market share, Pichai has brushed aside concerns that OpenAI will usurp Google's business dominance.

Atlassian competitor Linear raises funding at $1.25-billion valuation
Atlassian competitor Linear raises funding at $1.25-billion valuation

CNA

time4 hours ago

  • CNA

Atlassian competitor Linear raises funding at $1.25-billion valuation

SAN FRANCISCO :Enterprise software maker Linear has raised an $82 million Series C funding round valuing the startup at $1.25 billion, the company said on Tuesday. Venture capital fund Accel led the round, with participation from existing investors 01A and Sequoia, and new investors Seven Seven Six and Designer Fund. Linear, a maker of software development and project planning tools, competes with Atlassian's project management tool Jira. Linear said its profits grew 280 per cent last year, and it now has over 15,000 customers, including buzzy AI companies OpenAI, Scale AI and Perplexity. The 80-person, remote-first company will use its funding to build more products and attract larger enterprises to its customer base, said CEO Karri Saarinen. Linear focuses on specific product development use cases, a contrast to other tools that offer extensive customization but often overwhelm users, Saarinen said. For example, the company has specific functionalities around common software development workflows, such as a built-in "triage inbox" for software bugs and feature requests, and management for software development cycles, called sprints. It also has functionality for managing an AI like a team member, enabling humans and AIs to build software together effectively, something that is becoming common, Saarinen said. The company's commitment to customer-first product development over a technology-first approach is a focus that has been overlooked in the AI era, Linear investor Miles Clements, a partner at Accel, said. "There are a lot of vendors that are pushing a lot of unwanted AI slop into the market, and the Linear team instead is clued into what users are looking for and then providing them something they want," he said.

Nvidia, HPE to build new supercomputer in Germany
Nvidia, HPE to build new supercomputer in Germany

CNA

time8 hours ago

  • CNA

Nvidia, HPE to build new supercomputer in Germany

SAN FRANCISCO :Nvidia and Hewlett Packard Enterprise said on Tuesday they are partnering with the Leibniz Supercomputing Centre to build a new supercomputer using Nvidia's next-generation chips. The Blue Lion supercomputer, as the project is called, will become available to scientists in early 2027, using Nvidia's "Vera Rubin" chips. The announcement, made at a supercomputing conference in Hamburg, Germany, follows Nvidia's announcement that the Lawrence Berkeley National Lab in the United States also plans to build a system using the chips next year. Separately, Nvidia also said that Jupiter, another supercomputer using its chips at German national research institute Forschungszentrum Julich, has officially become Europe's fastest system. The deals represent European institutions aiming to stay competitive against the U.S. in supercomputers used for scientific fields from biotechnology to climate research. Long before it became an artificial intelligence powerhouse, Nvidia set out to persuade scientists to use its chips to speed up complex computer problems, such as modeling climate change. Those problems required many precise calculations that could take months at a time. Nvidia is now working to persuade scientists to use artificial intelligence. Those AI systems can take the results of a few precise calculations and use them to make predictions that, while not as accurate as the fully calculated results, can still be useful while taking far less time. Nvidia on Tuesday unveiled what it calls its "Climate in a Bottle" AI model. In a press briefing, Dion Harris, head of data center product marketing at Nvidia, said scientists will be able to input a few initial conditions such as sea surface temperatures and generate a forecast for 10 to 30 years in the future and see what the weather may be like at any kilometer or so of the earth's surface. "Researchers will use combined approach of classic physics and AI to resolve turbulent atmospheric flows," Harris said. "This technique will allow them to analyze thousands and thousands more scenarios in greater detail than ever before."

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store