logo
China's DeepSeek releases an update to its R1 reasoning model

China's DeepSeek releases an update to its R1 reasoning model

Yahooa day ago

SHANGHAI -Chinese artificial intelligence startup DeepSeek released an update to its R1 reasoning model in the early hours of Thursday, stepping up competition with U.S. rivals such as OpenAI.
DeepSeek launched R1-0528 on developer platform Hugging Face, but has yet to make an official public announcement. It did not publish a description of the model or comparisons.
But the LiveCodeBench leaderboard, a benchmark developed by researchers from UC Berkeley, MIT, and Cornell, ranked DeepSeek's updated R1 reasoning model just slightly behind OpenAI's o4 mini and o3 reasoning models on code generation and ahead of xAI's Grok 3 mini and Alibaba's Qwen 3.
Bloomberg earlier reported the update on Wedneday. It said that a DeepSeek representative had told a WeChat group that it had completed what it described as a "minor trial upgrade" and that users could start testing it.
DeepSeek earlier this year upended beliefs that U.S. export controls were holding back China's AI advancements after the startup released AI models that were on a par or better than industry-leading models in the United States at a fraction of the cost.
The launch of R1 in January sent tech shares outside China plummetting in January and challenged the view that scaling AI requires vast computing power and investment. Since R1's release, Chinese tech giants like Alibaba and Tencent have released models claiming to surpass DeepSeek's.
Google's Gemini has introduced discounted tiers of access while OpenAI cut prices and released an o3 Mini model that relies on less computing power.
The company is still widely expected to release R2, a successor to R1. Reuters reported in March, citing sources, that R2's release was initially planned for May. DeepSeek also released an upgrade to its V3 large language model in March.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

France says talks with China failed to solve cognac dispute
France says talks with China failed to solve cognac dispute

Yahoo

time29 minutes ago

  • Yahoo

France says talks with China failed to solve cognac dispute

PARIS (Reuters) -Talks between China and France on Thursday failed to end a dispute between the two countries over cognac tariffs, French Finance Minister Eric Lombard said, although he added the door was open for further discussions. After trade tensions flared between Beijing and Brussels, China began an anti-dumping investigation into European Union brandy in January. It was extended in April, giving EU exporters more time to find a way to avoid Chinese penalties. "At this stage, the economic dialogue has not led to a definitive solution to this dispute," Lombard told journalists after wide-ranging talks with visiting Chinese Vice Premier He Lifeng at the finance ministry in Paris. "I know, having discussed it at length with His Excellency, that the door to discussion remains open," he added. In France, the Chinese anti-dumping probe is regarded as retaliation for Paris' support for EU tariffs on imports of Chinese electric vehicles. "The investigating authorities will make a decision on the basis of investigation in accordance with laws and regulations and on the basis of the facts," He said. Chinese official news agency Xinhua's readout of the meeting focused on He speaking about China's readiness to "enrich bilateral economic and financial cooperation" and create a good trade and investment environment with France. Cognac lobby, the National Interprofessional Bureau of Cognac, said in a statement the absence of an agreement could have "catastrophic effects" on its industry and region, saying that every day that passed without a resolution was painful. "We expect full mobilisation of our government to avoid the definitive imposition of these tariffs on July 5," it said. Separately, the French agriculture ministry said the two countries had signed two protocols to formalise an agreement to allow unaffected regions to continue poultry exports in the event of a bird flu outbreak in France, and to provide access for French poultry to the Chinese market. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data

Remy Cointreau names luxury veteran Marilly as new CEO
Remy Cointreau names luxury veteran Marilly as new CEO

Yahoo

time29 minutes ago

  • Yahoo

Remy Cointreau names luxury veteran Marilly as new CEO

(Reuters) - Remy Cointreau on Wednesday named Franck Marilly as its new CEO, tasked with leading the French cognac maker through a period of spiralling sales and tariff threats in its key U.S. and Chinese markets. Marilly, 59, who has worked in luxury and cosmetics groups including Chanel, Unilever and Japan's Shiseido, will take office on June 25, replacing Eric Vallat, who resigned earlier this year after more than five years at the helm. Vallat led Remy through the COVID-19 pandemic and a subsequent boom in luxury spirits that has since gone into reverse, with sales tanking. Remy makes some 70% of its sales from cognac, mostly in the U.S. and China. Marilly now has to turn performance around at a time when drinkers in those nations are not buying the brandy, and when Remy faces tariffs on its exports to both. Vallat will work alongside Marilly to ensure a harmonious transition, Remy's statement said, adding Marilly had successfully navigated "high-stakes" environments in his previous roles. "We are convinced that he will bring a new dynamic and will be able to confidently address the new challenges of the group's growth," Remy Cointreau Chairperson Marie-Amelie de Leusse said in a statement. Remy's shares, which have already fallen substantially since 2022, were down 1.2% at 0721 GMT. VALUE STRATEGY Remy's sales have come under pressure amid a sluggish Chinese economy and tariffs on European brandy. Meanwhile, high interest rates and inflation have led to steep declines in the U.S., where Remy's rivals have also been cutting prices and taking market share. Remy however has maintained its approach of cultivating the value of its high-end brands, a strategy it says would be undermined by price cuts. It said Marilly will continue this value-led approach and build up in new markets. Investors however have warned diversification will take time to show results. The new CEO now also faces the threat of steep tariffs from the U.S. Remy is more exposed to tariffs than peers due to its reliance on Chinese and U.S. cognac sales. Marilly said he was delighted to join the group, and would focus on brand value and sustainable results. "Together we will accelerate the group's development... while meeting the needs of a constantly evolving sector," he said. Previously, Marilly led the Europe, Middle East and Africa and fragrance divisions at Shiseido Group and Chanel, where he spent 17 years. Sign in to access your portfolio

OpenAI Can Stop Pretending
OpenAI Can Stop Pretending

Atlantic

time43 minutes ago

  • Atlantic

OpenAI Can Stop Pretending

OpenAI is a strange company for strange times. Valued at $300 billion—roughly the same as seven Fords or one and a half PepsiCos—the AI start-up has an era-defining product in ChatGPT and is racing to be the first to build superintelligent machines. The company is also, to the apparent frustration of its CEO Sam Altman, beholden to its nonprofit status. When OpenAI was founded in 2015, it was meant to be a research lab that would work toward the goal of AI that is 'safe' and 'benefits all of humanity.' There wasn't supposed to be any pressure—or desire, really—to make money. Later, in 2019, OpenAI created a for-profit subsidiary to better attract investors—the types of people who might otherwise turn to the less scrupulous corporations that dot Silicon Valley. But even then, that part of the organization was under the nonprofit side's control. At the time, it had released no consumer products and capped how much money its investors could make. Then came ChatGPT. OpenAI's leadership had intended for the bot to provide insight into how people would use AI without any particular hope for widespread adoption. But ChatGPT became a hit, kicking 'off a growth curve like nothing we have ever seen,' as Altman wrote in an essay this past January. The product was so alluring that the entire tech industry seemed to pivot overnight into an AI arms race. Now, two and a half years since the chatbot's release, Altman says some half a billion people use the program each week, and he is chasing that success with new features and products—for shopping, coding, health care, finance, and seemingly any other industry imaginable. OpenAI is behaving like a typical business, because its rivals are typical businesses, and massive ones at that: Google and Meta, among others. Now 2015 feels like a very long time ago, and the charitable origins have turned into a ball and chain for OpenAI. Last December, after facing concerns from potential investors that pouring money into the company wouldn't pay off because of the nonprofit mission and complicated governance structure, the organization announced plans to change that: OpenAI was seeking to transition to a for-profit. The company argued that this was necessary to meet the tremendous costs of building advanced AI models. A nonprofit arm would still exist, though it would separately pursue 'charitable initiatives'—and it would not have any say over the actions of the for-profit, which would convert into a public-benefit corporation, or PBC. Corporate backers appeared satisfied: In March, the Japanese firm Softbank conditioned billions of dollars in investments on OpenAI changing its structure. Resistance came as swiftly as the new funding. Elon Musk—a co-founder of OpenAI who has since created his own rival firm, xAI, and seems to take every opportunity to undermine Altman— wrote on X that OpenAI 'was funded as an open source, nonprofit, but has become a closed source, profit-maximizer.' He had already sued the company for abandoning its founding mission in favor of financial gain, and claimed that the December proposal was further proof. Many unlikely allies emerged soon after. Attorneys general in multiple states, nonprofit groups, former OpenAI employees, outside AI experts, economists, lawyers, and three Nobel laureates all have raised concerns about the pivot, even petitioning to submit briefs to Musk's lawsuit. OpenAI backtracked, announcing a new plan earlier this month that would have the nonprofit remain in charge. Steve Sharpe, a spokesperson for OpenAI, told me over email that the new proposed structure 'puts us on the best path to' build a technology 'that could become one of the most powerful and beneficial tools in human history.' (The Atlantic entered into a corporate partnership with OpenAI in 2024.) Yet OpenAI's pursuit of industry-wide dominance shows no real signs of having hit a roadblock. The company has a close relationship with the Trump administration and is leading perhaps the biggest AI infrastructure buildout in history. Just this month, OpenAI announced a partnership with the United Arab Emirates and an expansion into personal gadgets—a forthcoming ' family of devices ' developed with Jony Ive, former chief design officer at Apple. For-profit or not, the future of AI still appears to be very much in Altman's hands. Why all the worry about corporate structure anyway? Governance, boardroom processes, legal arcana—these things are not what sci-fi dreams are made of. Yet those concerned with the societal dangers that generative AI, and thus OpenAI, pose feel these matters are of profound importance. The still more powerful artificial 'general' intelligence, or AGI, that OpenAI and its competitors are chasing could theoretically cause mass unemployment, worsen the spread of misinformation, and violate all sorts of privacy laws. In the highest-flung doomsday scenarios, the technology brings about civilizational collapse. Altman has expressed these concerns himself—and so OpenAI's 2019 structure, which gave the nonprofit final say over the for-profit's actions, was meant to guide the company toward building the technology responsibly instead of rushing to release new AI products, sell subscriptions, and stay ahead of competitors. 'OpenAI's nonprofit mission, together with the legal structures committing it to that mission, were a big part of my decision to join and remain at the company,' Jacob Hilton, a former OpenAI employee who contributed to ChatGPT, among other projects, told me. In April, Hilton and a number of his former colleagues, represented by the Harvard law professor Lawrence Lessig, wrote a letter to the court hearing Musk's lawsuit, arguing that a large part of OpenAI's success depended on its commitment to safety and the benefit of humanity. To renege on, or at least minimize, that mission was a betrayal. The concerns extend well beyond former employees. Geoffrey Hinton, a computer scientist at the University of Toronto who last year received a Nobel Prize for his AI research, told me that OpenAI's original structure would better help 'prevent a super intelligent AI from ever wanting to take over.' Hinton is one of the Nobel laureates who has publicly opposed the tech company's for-profit shift, alongside the economists Joseph Stiglitz and Oliver Hart. The three academics, joining a number of influential lawyers, economists, and AI experts, in addition to several former OpenAI employees, including Hilton, signed an open letter in April urging the attorneys general in Delaware and California—where the company's nonprofit was incorporated and where the company is headquartered, respectively—to closely investigate the December proposal. According to its most recent tax filing, OpenAI is intended to build AGI 'that safely benefits humanity, unconstrained by a need to generate financial return,' so disempowering the nonprofit seemed, to the signatories, self-evidently contradictory. In its initial proposal to transition to a for-profit, OpenAI still would have had some accountability as a public-benefit corporation: A PBC legally has to try to make profits for shareholders alongside pursuing a designated 'public benefit' (in this case, building 'safe' and 'beneficial' AI as outlined in OpenAI's founding mission). In its December announcement, OpenAI described the restructure as 'the next step in our mission.' But Michael Dorff, another signatory to the open letter and a law professor at UCLA who studies public-benefit corporations, explained to me that PBCs aren't necessarily an effective way to bring about public good. 'They are not great enforcement tools,' he said—they can 'nudge' a company toward a given cause but do not give regulators much authority over that commitment. (Anthropic and xAI, two of OpenAI's main competitors, are also public-benefit corporations.) OpenAI's proposed conversion also raised a whole other issue—a precedent for taking resources accrued under charitable intentions and repurposing them for profitable pursuits. And so yet another coalition, composed of nonprofits and advocacy groups, wrote its own petition for OpenAI's plans to be investigated, with the aim of preventing charitable organizations from being leveraged for financial gain in the future. Regulators, it turned out, were already watching. Three days after OpenAI's December announcement of the plans to revoke nonprofit oversight, Kathy Jennings, the attorney general of Delaware, notified the court presiding over Musk's lawsuit that her office was reviewing the proposed restructure to ensure that the corporation was fulfilling its charitable interest to build AI that benefits all of humanity. California's attorney general, Rob Bonta, was reviewing the restructure, as well. This ultimately led OpenAI to change plans. 'We made the decision for the nonprofit to stay in control after hearing from civic leaders and having discussions with the offices of the Attorneys General of California and Delaware,' Altman wrote in a letter to OpenAI employees earlier this month. The for-profit, meanwhile, will still transition to a PBC. The new plan is not yet a done deal: The offices of the attorneys general told me that they are reviewing the new proposal. Microsoft, OpenAI's closest corporate partner, has not yet agreed to the new structure. One could be forgiven for wondering what all the drama is for. Amid tension over OpenAI's corporate structure, the organization's corporate development hasn't so much as flinched. In just the past few weeks, the company has announced a new CEO of applications, someone to directly oversee and expand business operations; OpenAI for Countries, an initiative focused on building AI infrastructure around the world; and Codex, a powerful AI 'agent' that does coding tasks. To OpenAI, these endeavors legitimately contribute to benefiting humanity: building more and more useful AI tools; bringing those tools and the necessary infrastructure to run them to people around the world; drastically increasing the productivity of software engineers. No matter OpenAI's ultimate aims, in a race against Google and Meta, some commercial moves are necessary to stay ahead. And enriching OpenAI's investors and improving people's lives are not necessarily mutually exclusive. The greater issue is this: There is no universal definition for 'safe' or 'beneficial' AI. A chatbot might help doctors process paperwork faster and help a student float through high school without learning a thing; an AI research assistant could help climate scientists arrive at novel insights while also consuming huge amounts of water and fossil fuels. Whatever definition OpenAI applies will be largely determined by its board. Altman, in his May letter to employees, contended that OpenAI is on the best path 'to continue to make rapid, safe progress and to put great AI in the hands of everyone.' But everyone, in this case, has to trust OpenAI's definition of safe progress. The nonprofit has not always been the most effective check on the company. In 2023, the nonprofit board—which then and now had 'control' over the for-profit subsidiary— removed Altman from his position as CEO. But the company's employees revolted, and he was reinstated shortly thereafter with the support of Microsoft. In other words, 'control' on paper does not always amount to much in reality. Sharpe, the OpenAI spokesperson, said the nonprofit will be able to appoint and remove directors to OpenAI's separate for-profit board, but declined to clarify whether its board will be able to remove executives (such as the CEO). The company is 'continuing to work through the specific governance mandate in consultation with relevant stakeholders,' he said. Sharpe also told me that OpenAI will remove the cap on shareholder returns, which he said will satisfy the conditions for SoftBank's billions of dollars in investment. A top SoftBank executive has said 'nothing has really changed' with OpenAI's restructure, despite the nonprofit retaining control. If investors are now satisfied, the underlying legal structure is irrelevant. Marc Toberoff, a lawyer representing Musk in his lawsuit against OpenAI, wrote in a statement that 'SoftBank pulled back the curtain on OpenAI's corporate theater and said the quiet part out loud. OpenAI's recent 'restructuring' proposal is nothing but window dressing.' Lessig, the lawyer who represented the former OpenAI employees, told me that 'it's outrageous that we are allowing the development of this potentially catastrophic technology with nobody at any level doing any effective oversight of it.' Two years ago, Altman, in Senate testimony, seemed to agree with that notion: He told lawmakers that 'regulatory intervention by governments will be critical to mitigate the risks' of powerful AI. But earlier this month, only a few days after writing to his employees and investors that 'as AI accelerates, our commitment to safety grows stronger,' he told the Senate something else: Too much regulation would be 'disastrous' for America's AI industry. Perhaps—but it might also be in the best interests of humanity.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store