logo
Étienne-Émile Baulieu, Father of the Abortion Pill, Is Dead at 98

Étienne-Émile Baulieu, Father of the Abortion Pill, Is Dead at 98

New York Times31-05-2025
Étienne-Émile Baulieu, the French biochemist and physician who was often called the father of the abortion pill — and who was also known for his pioneering studies on the role of steroid hormones in human reproduction and aging — died on Friday at his home in Paris. He was 98.
His wife, Simone Harari Baulieu, confirmed the death on social media.
Dr. Baulieu's early research focused on hormones, notably DHEA, one of the key hormones in the adrenal gland, as well as groundbreaking work on estrogen and progesterone. But it was his development in the early 1980s of the synthetic steroid RU-486, or mifepristone, that thrust him onto the public stage.
Unlike the morning-after pill, which is used after sex to delay ovulation, RU-486 works as a kind of 'anti-hormone,' in Dr. Baulieu's words, by blocking the uterus from receiving progesterone, thereby preventing a fertilized egg from implanting.
Taking the drug with misoprostol, a drug that causes uterine contractions, essentially triggers a miscarriage, enabling women to terminate early pregnancies without surgery.
The two-dose treatment has been proved safe and highly effective — with a success rate of about 95 percent — and is commonly used in many countries; in the United States, medication abortions accounted for more than 50 percent of all abortions in 2020. After the 2022 Supreme Court decision overturning Roe v. Wade, demand for the pills surged, and abortion opponents began seeking ways to ban the drug nationwide.
Controversy over RU-486 began as soon as its release in the 1980s. Dr. Baulieu developed the drug in partnership with the French drug company Roussel-Uclaf, where he was an independent consultant.
Want all of The Times? Subscribe.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Lithuania Asks NATO for Air Defense After Russian Drone Scare
Lithuania Asks NATO for Air Defense After Russian Drone Scare

Bloomberg

timean hour ago

  • Bloomberg

Lithuania Asks NATO for Air Defense After Russian Drone Scare

Lithuania called on its NATO allies to immediately help enhance its air defense capabilities, after two Russian military drones crossed into its airspace in less than a month. The Baltic nation's foreign and defense ministers sent a letter to NATO Secretary General Mark Rutte on Tuesday urging the alliance to quickly implement a rotational model which requires members to take turns sending air defense systems.

The EU AI Act aims to create a level playing field for AI innovation. Here's what it is.
The EU AI Act aims to create a level playing field for AI innovation. Here's what it is.

TechCrunch

timean hour ago

  • TechCrunch

The EU AI Act aims to create a level playing field for AI innovation. Here's what it is.

The European Union's Artificial Intelligence Act, known as the EU AI Act, has been described by the European Commission as 'the world's first comprehensive AI law.' After years in the making, it is progressively becoming a part of reality for the 450 million people living in the 27 countries that comprise the EU. The EU AI Act, however, is more than a European affair. It applies to companies both local and foreign, and it can affect both providers and deployers of AI systems; the European Commission cites examples of how it would apply to a developer of a CV screening tool, and to a bank that buys that tool. Now, all of these parties have a legal framework that sets the stage for their use of AI. Why does the EU AI Act exist? As usual with EU legislation, the EU AI Act exists to make sure there is a uniform legal framework applying to a certain topic across EU countries — the topic this time being AI. Now that the regulation is in place, it should 'ensure the free movement, cross-border, of AI-based goods and services' without diverging local restrictions. With timely regulation, the EU seeks to create a level playing field across the region and foster trust, which could also create opportunities for emerging companies. However, the common framework that it has adopted is not exactly permissive: Despite the relatively early stage of widespread AI adoption in most sectors, the EU AI Act sets a high bar for what AI should and shouldn't do for society more broadly. What is the purpose of the EU AI Act? According to European lawmakers, the framework's main goal is to 'promote the uptake of human centric and trustworthy AI while ensuring a high level of protection of health, safety, fundamental rights as enshrined in the Charter of Fundamental Rights of the European Union, including democracy, the rule of law and environmental protection, to protect against the harmful effects of AI systems in the Union, and to support innovation.' Yes, that's quite a mouthful, but it's worth parsing carefully. First, because a lot will depend on how you define 'human centric' and 'trustworthy' AI. And second, because it gives a good sense of the precarious balance to maintain between diverging goals: innovation vs. harm prevention, as well as uptake of AI vs. environmental protection. As usual with EU legislation, again, the devil will be in the details. How does the EU AI Act balance its different goals? To balance harm prevention against the potential benefits of AI, the EU AI Act adopted a risk-based approach: banning a handful of 'unacceptable risk' use cases; flagging a set of 'high-risk' uses calling for tight regulation; and applying lighter obligations to 'limited risk' scenarios. Techcrunch event Tech and VC heavyweights join the Disrupt 2025 agenda Netflix, ElevenLabs, Wayve, Sequoia Capital — just a few of the heavy hitters joining the Disrupt 2025 agenda. They're here to deliver the insights that fuel startup growth and sharpen your edge. Don't miss the 20th anniversary of TechCrunch Disrupt, and a chance to learn from the top voices in tech — grab your ticket now and save up to $675 before prices rise on August 7. Tech and VC heavyweights join the Disrupt 2025 agenda Netflix, ElevenLabs, Wayve, Sequoia Capital — just a few of the heavy hitters joining the Disrupt 2025 agenda. They're here to deliver the insights that fuel startup growth and sharpen your edge. Don't miss the 20th anniversary of TechCrunch Disrupt, and a chance to learn from the top voices in tech — grab your ticket now and save up to $675 before prices rise. San Francisco | REGISTER NOW Has the EU AI Act come into effect? Yes and no. The EU AI Act rollout started on August 1, 2024, but it will only come into force through a series of staggered compliance deadlines. In most cases, it will also apply sooner to new entrants than to companies that already offer AI products and services in the EU. The first deadline came into effect on February 2, 2025, and focused on enforcing bans on a small number of prohibited uses of AI, such as untargeted scraping of internet or CCTV for facial images to build up or expand databases. Many others will follow, but unless the schedule changes, most provisions will apply by mid-2026. What changed on August 2, 2025? Since August 2, 2025, the EU AI Act applies to 'general-purpose AI models with systemic risk.' GPAI models are AI models trained with a large amount of data, and that can be used for a wide range of tasks. That's where the risk element comes in. According to the EU AI Act, GPAI models can come with systemic risks; 'for example, through the lowering of barriers for chemical or biological weapons development, or unintended issues of control over autonomous [GPAI] models.' Ahead of the deadline, the EU published guidelines for providers of GPAI models, which include both European companies and non-European players such as Anthropic, Google, Meta, and OpenAI. But since these companies already have models on the market, they will also have until August 2, 2027, to comply, unlike new entrants. Does the EU AI Act have teeth? The EU AI Act comes with penalties that lawmakers wanted to be simultaneously 'effective, proportionate and dissuasive' — even for large global players. Details will be laid down by EU countries, but the regulation sets out the overall spirit — that penalties will vary depending on the deemed risk level — as well as thresholds for each level. Infringement on prohibited AI applications leads to the highest penalty of 'up to €35 million or 7% of the total worldwide annual turnover of the preceding financial year (whichever is higher).' The European Commission can also inflict fines of up to €15 million or 3% of annual turnover on providers of GPAI models. How fast do existing players intend to comply? The voluntary GPAI code of practice, including commitments such as not training models on pirated content, is a good indicator of how companies may engage with the framework law until forced to do so. In July 2025, Meta announced it wouldn't sign the voluntary GPAI code of practice meant to help such providers comply with the EU AI Act. However, Google soon after confirmed it would sign, despite reservations. Signatories so far include Aleph Alpha, Amazon, Anthropic, Cohere, Google, IBM, Microsoft, Mistral AI, and OpenAI, among others. But as we have seen with Google's example, signing does not equal a full-on endorsement. Why have (some) tech companies been fighting these rules? While stating in a blog post that Google would sign the voluntary GPAI code of practice, its president of global affairs, Kent Walker, still had reservations. 'We remain concerned that the AI Act and Code risk slowing Europe's development and deployment of AI,' he wrote. Meta was more radical, with its chief global affairs officer Joel Kaplan stating in a post on LinkedIn that 'Europe is heading down the wrong path on AI.' Calling the EU's implementation of the AI Act 'overreach,' he stated that the code of practice 'introduces a number of legal uncertainties for model developers, as well as measures which go far beyond the scope of the AI Act.' European companies have expressed concerns as well. Arthur Mensch, the CEO of French AI champion Mistral AI, was part of a group of European CEOs who signed an open letter in July 2025 urging Brussels to 'stop the clock' for two years before key obligations of the EU AI Act came into force. Will the schedule change? In early July 2025, the European Union responded negatively to lobbying efforts calling for a pause, saying it would still stick to its timeline for implementing the EU AI Act. It went ahead with the August 2, 2025, deadline as planned, and we will update this story if anything changes.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store