logo
#

Latest news with #AI-transformed

CodeSignal Report Ranks Universities by Measurable Technical Skills, Highlighting Top Engineering Talent Nationwide
CodeSignal Report Ranks Universities by Measurable Technical Skills, Highlighting Top Engineering Talent Nationwide

Yahoo

time14-05-2025

  • Science
  • Yahoo

CodeSignal Report Ranks Universities by Measurable Technical Skills, Highlighting Top Engineering Talent Nationwide

Nearly 1 in 3 top-performing students come from universities overlooked by traditional rankings SAN FRANCISCO, May 14, 2025 /PRNewswire/ -- CodeSignal, a leading skills assessment and experiential learning platform, today unveils its fourth annual University Ranking Report, an university ranking methodology based purely on students' verified coding skills. Unlike traditional rankings that rely on legacy signals, CodeSignal's report offers an objective, data-driven alternative: one that evaluates universities based on how well their students perform on an assessment of real-world coding skills. In an AI-transformed workforce, the ability to think computationally, solve problems, and write strong foundational code remains critical, regardless of where a student went to school. By analyzing thousands of General Coding Assessments (GCA) completed by students worldwide, CodeSignal's Talent Science Team reveals a powerful conclusion: top engineering talent is everywhere. Here are the top 15 universities for 2025: Carnegie Mellon University Massachusetts Institute of Technology Stony Brook University University of California, Los Angeles University of Pennsylvania California Institute of Technology University of California, San Diego Duke University San José State University University of Southern California Rice University Yale University Georgia Institute of Technology Johns Hopkins University Indiana University High-level results: 28.4% of high-scorers come from schools not included in the US News & World Report's top 50 undergraduate engineering programs. 12 of the top 50 schools in our skill-based ranking did not make the US News & World top 50. Two of the top 10 US schools in our rankings, Stony Brook University (#3) and San José State University (#9), didn't make the US News & World top 50. Korea Advanced Institute of Science & Technology is the top non-US school for software engineering talent this year, ranking just below Rice University (#12 on the US list). "This report is a celebration of the universities equipping students with the skills that matter most," said Tigran Sloyan, CEO and Co-Founder of CodeSignal. "When we focus on what students can actually do, not just where they studied, we uncover incredible talent from institutions of all types. It's a reminder that great engineers are everywhere, and we need to broaden how we recognize and recruit them." While traditional rankings reward legacy signals, CodeSignal's 2025 University Ranking Report focuses on outcomes – what students can actually do when faced with real-world engineering challenges. CodeSignal's data makes the case that technical talent isn't confined to a short list of name-brand schools. It's everywhere. For employers competing in an AI-driven economy, this report is a call to rethink where, and how, they discover their next generation of engineers. To view the full report, please visit: About CodeSignalCodeSignal is how the world discovers and develops the skills that will shape the future. Our AI-native skills assessment and experiential learning platform helps organizations hire, train, and grow talent at scale while empowering individuals to advance their careers. Whether you're growing your team's potential or unlocking your own, CodeSignal meets you where you are and gets you where you need to go. With millions of skills assessments completed, CodeSignal is trusted by companies like Netflix, Capital One, Meta, and Dropbox and used by learners worldwide. For more information, visit or connect with CodeSignal on LinkedIn. View original content to download multimedia: SOURCE CodeSignal

What International AI Safety report says on jobs, climate, cyberwar and more
What International AI Safety report says on jobs, climate, cyberwar and more

The Guardian

time29-01-2025

  • Business
  • The Guardian

What International AI Safety report says on jobs, climate, cyberwar and more

The International AI Safety report is a wide-ranging document that acknowledges an array of challenges posed by a technology that is advancing at dizzying speed. The document, commissioned after the 2023 global AI safety summit, covers numerous threats from deepfakes to aiding cyberattacks and the use of biological weapons, as well as the impact on jobs and the environment. Here are some of the key points from the report chaired by Yoshua Bengio, a world-leading computer scientist. In a section on 'labour market risks', the report warns that the impact on jobs will 'likely be profound', particularly if AI agents – tools that can carry out tasks without human intervention – become highly capable. 'General-purpose AI, especially if it continues to advance rapidly, has the potential to automate a very wide range of tasks, which could have a significant effect on the labour market. This means that many people could lose their current jobs,' said the report. The report added that many economists believe job losses could be offset by the creation of new jobs or demand from sectors not touched by automation. According to the International Monetary Fund, about 60% of jobs in advanced economies such as the US and UK are exposed to AI and half of these jobs may be negatively affected. The Tony Blair Institute has said AI could displace up to 3m private-sector jobs in the UK, though the ultimate rise in unemployment will be in the low hundreds of thousands because growth in the technology will create new roles in an AI-transformed economy. 'These disruptions could be particularly severe if autonomous AI agents become capable of completing longer sequences of tasks without human supervision,' the report says. It adds that some experts have pointed to scenarios where work is 'largely' eliminated. In 2o23 Elon Musk, the world's richest person, told the former UK prime minister, Rishi Sunak, that AI could ultimately replace all human jobs. However, the report said such views were controversial and there was 'considerable uncertainty' over how AI might impact labour markets. The report described AI's impact on the environment as a 'moderate but rapidly growing contributor' as datacentres – the central nervous systems of AI models – consume electricity to train and operate the technology. Datacentres and data transmission account for about 1% of energy-related greenhouse gas emissions, said the report, with AI constituting up to 28% of datacentre energy consumption. It added that models are using more energy as they become more advanced and warned that a 'significant portion' of global model training relies on high-carbon energy sources such as coal or natural gas. Use of renewable energy by AI firms and improvements in efficient have not kept pace with rising demand for energy, said the report, which also points to tech firms admitting that AI development is harming their ability to meet environmental targets. The report also warns that water consumption by AI, used for cooling equipment in datacentres, could pose a 'substantial threat to the environment and the human right to water'. However, the report adds that there is a shortage of data about the environmental impact of AI. An all-powerful AI system evading human control is the central concern of experts who fear the technology could extinguish humanity. The report acknowledges those fears but says opinion varies 'greatly'. 'Some consider it implausible, some consider it likely to occur, and some see it as a modest-likelihood risk that warrants attention due to its high severity,' it says. Bengio told the Guardian that AI agents, which carry out tasks autonomously, are still being developed and so far are unable to carry out the long-term planning necessary for those systems to eradicate jobs wholesale – or evade safety guidelines. 'If an AI cannot plan over a long horizon, it's hardly going to be able to escape our control,' he said The report states that new models can create step-by-step guides to creating pathogens and toxins that surpass PhD-level expertise. However, it cautions that there is uncertainty over whether they can be used by novices. There is evidence of advancement since an interim safety report last year, the experts say, with OpenAI producing a model that could 'meaningfully assist experts in the operational planning of reproducing known biological threats'. A fast-growing threat from AI in terms of cyber-espionage is autonomous bots being able to find vulnerabilities in open-source software, the term for code that is free to download and adapt. However, relative shortcomings in AI agents mean the technology is not able to plan and carry out attacks autonomously. The report lists an array of known examples of AI deepfakes being used maliciously, including tricking companies into handing over money and creating pornographic images of people. However, the report says there is not enough data to fully measure the amount of deepfake incidents. 'Reluctance to report may be contributing to these challenges in understanding the full impact of AI-generated content intended to harm individuals,' the report says. For example, institutions often hesitate to disclose their struggles with AI-powered fraud. Similarly, individuals attacked with AI-generated compromising material about themselves may stay silent out of embarrassment and to avoid further harm.' The report also warns that there are 'fundamental challenges' to tackling deepfake content, such as the ability to remove digital watermarks that flag AI-generated content.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store