
Human wins coding contest … but AI finishes a close second
Dębiak, who happens to be a former employee of OpenAI, was showing off his programming skills at the AtCoder World Tour Finals held in Tokyo last week.
Recommended Videos
The 10-hour session left the Pole, who competed under the name Psyho, feeling 'completely exhausted' and 'barely alive,' according to a post on X.
As for the AI model, well, presumably it felt just fine and could've carried on without complaint.
It's believed to be the first coding contest to have allowed the participation of an AI model, with the technology gaining prominence for its rapidly improving programming smarts.
OpenAI's AI model participated as part of a special 'Humans vs AI' exhibition match conducted within the Heuristic division of the contest. Called OpenAIAHC, the AI model took on 12 top-ranking human programmers — with Dębiak among them — on a 10-hour optimization challenge.
The Polish programmer managed to finish in first place by a margin of just over 9%, with the AI placing second — ahead of all of the other human contestants.
Interestingly, Dębiak said in comments after the contest that it was the AI model's presence that motivated him to keep going, as he could see during the match-up that he was just ahead of the model, and he very much wanted to keep it that way. Without the AI's participation, his score would have been 'much, much lower,' Dębiak told Business Insider.
The AI's impressive performance is a clear demonstration that advanced AI is making real progress when it comes to edging out the world's best human programmers in open-ended, creative problem-solving tasks. Indeed, it's hard to think that it won't come out on top at next year's contest … if it's invited back, that is.

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
14 minutes ago
- Yahoo
Introducing Latent-X, a Frontier Generative AI Model for Protein Binder Design Accessible via No-Code Platform for Push-Button Protein Design
Latent-X generates lab-ready macrocycles and protein mini-binders at all-atom resolution to accelerate drug design The model can be accessed through Latent's web-based platform for push-button protein design. Sign ups are open now for early access: Extensive lab validation shows picomolar binding affinities outperforming prior models, with 91-100% hit rates for macrocycles and 10-64% for mini-binders LONDON & SAN FRANCISCO, July 22, 2025--(BUSINESS WIRE)--Today, Latent Labs is launching Latent-X, a frontier AI model for push button protein design, outperforming competing models under identical laboratory conditions. The model is available for early access on Latent's no-code AI protein design platform, where users can upload protein targets and generate cyclic peptides and mini-binders directly in the browser. Through the platform, users can generate, explore, and score binder designs, selecting top-ranked structures for further lab testing. The platform includes a free tier for both commercial and non-commercial users. Sign up is available at Latent Labs is a frontier AI lab working to transform the expensive, labor intensive, and high failure rate processes of drug discovery into automated drug design. Traditional drug discovery requires screening millions of random molecules—a process where hit rates are typically well below 1% and each experiment takes months and costs thousands of dollars. With Latent-X, drug designers can generate high-confidence binders with the push of a button, achieving what would typically require testing millions of candidates by testing as little as 30 candidates per target. AI models have recently enabled solutions to previously insurmountable technical challenges in biology. With generative models, frontier AI can go beyond predicting structures to creating new sequences and structures of candidate drugs. "We envision a future where effective therapeutics can be designed entirely in a computer, much like how space missions or semiconductors are designed today," said Simon Kohl, CEO and founder of Latent Labs. "Our platform empowers scientists with lab-validated protein binder design at their fingertips, whether they're experts or new to AI-powered drug design, and without needing AI infrastructure. This is the first step on our mission toward making biology programmable in order to make drug design instantaneous." Latent-X generates functional, high affinity de novo binders with breakthrough laboratory performance. In extensive wet lab experiments across 7 therapeutic targets, Latent-X achieved 91-100% hit rates for macrocycles and 10-64% hit rates for mini-binders. The model delivered picomolar binding affinities for mini-binders and single-digit micromolar affinities for macrocycles, with generated binders showing strong target specificity. In head-to-head experimental comparisons, Latent-X exceeded the prior state-of-the-art, outperforming existing generative tools in both in silico evaluations and laboratory validation. Macrocycles are a sought after drug modality for their potential oral deliverability, with their compactness promising tissue permeability while retaining specificity. Mini-binders are a versatile new binder modality that offers high specificity in a flexible format. Full results are available in our technical report: The Latent Labs Platform allows users to access the state of the art in protein binder design in an intuitive platform for target upload, hotspot selection, binder design, and computational ranking. The platform features structure visualization, predicted structure overlays, and computational metric rankings allowing to replicate the AI workflows used to generate our successfully lab-validated binders. Latent-X is a general purpose frontier model that creates binders from scratch for unseen or previously untargeted proteins, solving the geometric puzzle of binding at the all-atom level. The model generates designs over 10x faster than previous methods and co-samples sequence and structure simultaneously, allowing for computational experimentation within seconds. Latent-X generalizes beyond nature's repertoire by generating all-atom binder structures that obey atomic-level biochemical rules, opening doors to other therapeutic modalities that depend on target-specific binding—nanobodies and antibodies being prime examples. The company is now open to partnerships to bring these expanded capabilities to new drug applications. Only five months ago Latent Labs announced its $50M funding round co-lead by Radical Ventures and Sofinnova Partners, with participation by Google's Chief Scientist Jeff Dean, Anthropic's CEO Dario Amodei and Eleven Labs' CEO Mati Staniszewski. The team consists of former AlphaFold 2 co-developers, ex-DeepMind team leads, and brings rich experience from Microsoft, Apple, Stability AI, Exscientia, Mammoth Bio, Altos Labs and Zymergen. View source version on Contacts contact@ Error while retrieving data Sign in to access your portfolio Error while retrieving data Error while retrieving data Error while retrieving data Error while retrieving data


Forbes
17 minutes ago
- Forbes
Audacious Idea That America Is Going To Have An Unnerving Sputnik Moment When It Comes To Attaining AGI And AI Superintelligence
Will the United States attain AGI and ASI first, before any other country, and does it really matter ... More which country is first? In today's column, I examine the provocative chatter that the United States might experience a said-to-be Sputnik moment when it comes to attaining artificial general intelligence (AGI) and artificial superintelligence (ASI). How so? The audacious idea postulates that rather than America being the first to achieve AGI and ASI, some other country manages to beat us to the punch. It is a seemingly unimaginable proposition. You see, the United States is indisputably a world leader in AI and known for the development of leading-edge advances in AI. It is nearly inconceivable that the U.S. won't arrive at AGI and ASI first. But is that wishful thinking rather than real-world thinking? Let's talk about it. This analysis of an innovative AI breakthrough is part of my ongoing Forbes column coverage on the latest in AI, including identifying and explaining various impactful AI complexities (see the link here). Heading Toward AGI And ASI First, some fundamentals are required to set the stage for this weighty discussion. There is a great deal of research going on to further advance AI. The general goal is to either reach artificial general intelligence (AGI) or maybe even the outstretched possibility of achieving artificial superintelligence (ASI). AGI is AI that is considered on par with human intellect and can seemingly match our intelligence. ASI is AI that has gone beyond human intellect and would be superior in many, if not all, feasible ways. The idea is that ASI would be able to run circles around humans by outthinking us at every turn. For more details on the nature of conventional AI versus AGI and ASI, see my analysis at the link here. We have not yet attained AGI. In fact, it is unknown whether we will reach AGI, or that maybe AGI will be achievable in decades or perhaps centuries from now. The AGI attainment dates that are floating around are wildly varying and wildly unsubstantiated by any credible evidence or ironclad logic. ASI is even more beyond the pale when it comes to where we are currently with conventional AI. The Saga Of Sputnik 1 There is immense speculation going on about AGI and ASI regarding which country will be the first to achieve the vaunted pinnacle of AI. One fiery comment that's floating around is that this could turn out to be another semblance of the infamous Sputnik crisis. You might be somewhat familiar with the unnerving exploits of Sputnik that occurred in the 1950s and 1960s, and beyond. I'll provide a quick recap for ease of recollection and then tie the historical reference to our modern times. In October 1957, the Soviet Union launched a small spacecraft known as Sputnik 1 that traveled in a low Earth orbit. A radio signal was then beamed from this orbiting spacecraft. People across the globe could hear the beeping sounds on their radios as retransmitted by amateur radio operators. This made enormous history as it was the first-ever artificial Earth satellite. At the time, this frightening action triggered the American Sputnik crisis. Worries were that the Soviet Union could end up controlling outer space. The Russians could potentially launch military weapons that orbited the planet and would readily threaten the United States and other countries of the world. The action also suggested that the scientific prowess showcased by the Soviet Union was superior to that of America. How much farther behind might the U.S. really be? The sky was the limit, or maybe not, and extended to the heavens far above. This served as a mighty impetus to spur the Space Race. Indeed, a few years later, President John F. Kennedy made his famous speech in 1962 that called for the United States to land on the Moon before the end of the decade. The oft quoted line was this: 'We choose to go to the Moon in this decade and do the other things, not because they are easy, but because they are hard; because that goal will serve to organize and measure the best of our energies and skills, because that challenge is one that we are willing to accept, one we are unwilling to postpone, and one we intend to win, and the others, too.' Will AI Be The Sputnik 2 Let's tie the Sputnik saga with the ongoing efforts to attain pinnacle AI. There is heated debate in dark backrooms that maybe the United States won't be the first to arrive at AGI and ASI. Some other country might get there first. Lots of big-name countries are vying for that prized position. Smaller countries are doing so too. An eclectic race is avidly underway. Suppose the U.S. isn't first? It would be reminiscent of the Sputnik 1 circumstance. Perhaps such an instance would cheekily be labeled as a kind of Sputnik 2 phenomenon (as an aside, there really was a Sputnik 2 in terms of a second spacecraft launched in November 1957 by the Soviet Union and was the first to put an animal in space, the dog named Laika). America could end up as a second fiddle in the AI race. The idea seems absurd at face value. The United States undeniably has many of the top AI makers, along with amazing academic institutions that are globally recognized as AI leaders, and gobs of first-class AI researchers. Billions upon billions of dollars are flowing into the AI race by American companies and via U.S. federal, state, and local governmental agencies. Any notion of the United States not landing on AGI and ASI before any other country would seem utterly ludicrous and summarily rejected. The Logical Suppositions Whoa, comes the retort, you can't blindly assume that the United States will necessarily be the first to attain AGI and ASI. That is a haughty assumption. It belies the intense efforts taking place beyond the United States. This begs the question as to why America would not be the first to reach that desired goal. I will go ahead and give you a rundown on some of the most compelling reasons that have been expressed on this dicey matter. They consist of these five primary contentions: Let's briefly unpack each of those. Unknown Path To AGI And ASI First, no one anywhere can say for sure how AGI and ASI can be achieved. The whole endeavor is pretty much a shot in the dark. There isn't a pristine map that lays out the steps involved. Furthermore, it is conceivable that AGI and ASI will not be attained at all, i.e., no one will achieve the pinnacle AI. The United States is in the same boat as everyone else, namely, trying all sorts of clever ways to move toward AGI and ASI. No guarantees are to be had. All countries might come up blank on the AGI and ASI pursuit. Thus, no matter how much money or brainpower is employed, the end result might consist nicely of more advanced AI, but not the total package of true pinnacle AI. Marching To The Same Tune A second point is that perhaps a birds of a feather mindset could undermine the United States. Here's what that entails. Some have criticized that, by-and-large, we are using the same methods and similar AI internal structures across the board to reach AGI and ASI, see my analysis at the link here and the link here. If that's the case, our all-alike AI approach could be akin to putting all our eggs into one basket. The true path to pinnacle AI might be something outside of that presumed avenue. Unwilling To Take Risky Chances Another somewhat related consideration is that with the vast investments going into AI efforts, this might be making us more risk-averse. The logic is this. You would find it difficult to take in bucko bucks and not be pursuing AGI and ASI like others are. The investors won't be happy that you are trying some oddball angle. If you don't succeed but have followed the same approach as others, you can hold your head high and proclaim that everyone was caught off guard. On the other hand, if you opt for a risky path that no one else chooses to pursue, you'll have little headspace cover when it comes to explaining why your zany method didn't arrive at AGI and ASI. You will be fully exposed and readily vulnerable to reputational attack. Stealing AI To Reach The Pinnacle Here's a twist for you. Theft might come into play. It is suggested that maybe another country will steal our budding AI and manage to undertake the final steps to AGI and ASI before we do. In other words, suppose we have gotten down to the 90% mark and are struggling to get the final 10% done. Some other country that isn't anywhere near AGI and ASI decides to take a shortcut by stealing the AI that we have. Next, they manage to get the remaining 10% undertaken under their own auspices. Of course, they tout to the world that they did the pinnacle AI by themselves, entirely from A to Z. For more details on the chances of stealing AI, see my coverage at the link here. Discovery By Luck Or Chance One of the most intriguing reasons for the U.S. not being the first to achieve pinnacle AI is that perhaps there is some out-of-the-blue discovery that needs to be made. The ardent belief is that there is a missing piece that nobody has identified yet. No one knows what that piece is. There isn't any definition of it. It is the classical dilemma of not knowing what we don't know. The kicker is this. Suppose that discovering the missing piece is going to be based mainly on luck rather than skill. Assume that there is no inherent advantage in having the biggest AI labs and the biggest AI budgets. AGI and ASI might hinge on a completely left-field discovery that could happen anywhere and at any time. I've pointed out that this particular theory or conjecture has given rise to the credence that AGI and ASI might be achieved on a solo basis, see my discussion at the link here. Yes, instead of vast teams arriving at pinnacle AI, some enterprising individual in their pajamas and in their basement arrives there first. If you believe in this fanciful missing piece concept, it seems plausible that a solo developer with incredible luck might discover it. The solo developer might be in the tiniest of countries, and ergo, bring AGI and ASI to that country before any other country figures it out. Presumably, unbelievable fame and fortune await that solo developer. Being First Does Matter Those above-described handful of mainstay reasons are on the minds of many. Please know that additional reasons are being bandied around. It's a hot topic and raises the heat when emphatically discussed. A smarmy viewpoint about this dire handwringing conundrum is that being first is perhaps overrated. If the U.S. doesn't get to AGI and ASI before some other country, maybe it's not such a big deal, and we are making an undue fuss. A preoccupation with being first can be a bad thing. Go with the flow. However things perchance go, they go. The counterargument to this offhandness is that we all pretty much acknowledge that AGI and ASI have supremely dual-use consequences, doing grand good for the world but also potentially grand bad for the world. The first to get to pinnacle AI might unleash quite a vicious storm upon the globe and muscle themselves into a geo-economic position of a disconcerting nature (see my analysis at the link here, along with why the United Nations also is trying to have a role in the AGI/ASI arrival, see the link here). The planetary and humankind existential stakes underlying AGI and ASI are huge. Whichever country gets there first is, in fact, an important consideration. Humanity And The Future A final thought to ruminate on. Some have likened the attainment of AGI and ASI to the likes of achieving atomic energy and the atomic bomb. Historically, the case can be made that getting there first did make a difference. We now know that being first was significant, and we also know that what happens after the first attainment is an ongoing struggle and vitally crucial too. Thinking further ahead in terms of pinnacle AI, the question arises whether some or all other countries of the world will eventually possess and/or control AGI and ASI. That's another substantive topic worthy of keen chatter. Per the wise words of Albert Einstein, we earnestly need to keep this pointed remark in mind: 'The solution to this problem lies in the heart of mankind.'


Business Insider
17 minutes ago
- Business Insider
AMD Stock Rallies as Bernstein Analyst Raises Price Target
Advanced Micro Devices (AMD) stock rallied on Monday after the semiconductor company's shares received new analyst coverage. Five-star Bernstein analyst Stacy Rasgon assigned a Hold rating to AMD stock and increased the firm's price target for the shares to $140 from $95, suggesting a 12.17% downside. Elevate Your Investing Strategy: Take advantage of TipRanks Premium at 50% off! Unlock powerful investing tools, advanced data, and expert analyst insights to help you invest with confidence. Make smarter investment decisions with TipRanks' Smart Investor Picks, delivered to your inbox every week. While Rasgon stuck with a Hold rating for AMD stock, he highlighted several positive catalysts for the company. Among these are the lifted ban on AI chip sales to China, better times for the PC gaming market, and conditions similar to its rivals. In light of all this, Rasgon updated his estimates for AMD earnings. These includes Q2 adjusted EPS of 49 cents on revenue of $7.52 billion, compared to a prior estimate of 47 cents and $7.4 billion. He also updated his Q3 estimates to $1.20 per share and $8.43 billion from $1.09 per share and $8.08 billion, as well as his full-year 2025 estimate to $3.89 per share and $32 billion from $3.71 per share and $31.4 billion. AMD will next report earnings on Aug. 5, 2025. AMD Stock Movement & More Analyst Coverage AMD stock was up 1.43% as of Monday morning, extending a 31.83% year-to-date rally and a 0.72% increase over the past 12 months. AMD stock also received additional analyst coverage today. Five-star Citi analyst Christopher Danely reiterated a Hold rating and raised his price target for the shares to $165 from $145. Is AMD Stock a Buy, Sell, or Hold? Turning to Wall Street, the analysts' consensus for AMD is Moderate Buy, based on 25 Buy and 10 Hold ratings over the past three months. With that comes an average AMD stock price target of $145.90, representing a potential 8.33% downside for the shares.