
UK's online safety law is putting free speech at risk, X says
But it has attracted criticism from politicians, free-speech campaigners and content creators, who have complained that the rules had been implemented too broadly, resulting in the censorship of legal content.
Users have complained about age checks that require personal data to be uploaded to access sites that show pornography, and more than 468,000 people have signed an online petition calling for the act to be repealed.
The government said on Monday it had no plans to do so and it was working with regulator Ofcom to implement the act as quickly as possible.
Technology Secretary Peter Kyle said on Tuesday that those who wanted to overturn it were 'on the side of predators'.
Elon Musk's X, which has implemented age verification, said the law's laudable intentions were at risk of being overshadowed by the breadth of its regulatory reach.
'When lawmakers approved these measures, they made a conscientious decision to increase censorship in the name of 'online safety',' it said in a statement.
'It is fair to ask if UK citizens were equally aware of the trade-off being made.'
X said the timetable for meeting mandatory measures had been unnecessarily tight, and despite being in compliance, platforms still faced threats of enforcement and fines, encouraging over-censorship.
It said a balanced approach was the only way to protect liberty, encourage innovation and safeguard children.
'It's safe to say that significant changes must take place to achieve these objectives in the UK,' it said. A UK government spokesperson said it is 'demonstrably false' that the Online Safety Act compromises free speech. 'As well as legal duties to keep children safe, the very same law places clear and unequivocal duties on platforms to protect freedom of expression,' the spokesperson said. Ofcom said on Thursday it had launched investigations into the compliance of four companies, which collectively run 34 pornography sites.

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Hindustan Times
26 minutes ago
- Hindustan Times
Decoding UK's OSA, ChatGPT adds emotional support, and pondering Windows' future
The UK's newly updated enforcements for the Online Safety Act (OSA) are worth having a conversation about, this week. Most of the online media coverage would lead you to believe that it's a new act that's been passed into a new law, but it isn't — the Online Safety Act received Royal Assent on 26 October 2023, some provisions came into force on 10 January 2024, while additional elements took effect on 1 April 2024. We're talking about this now, because the critical age verification requirements took effect on 25 July 2025, which means all online platforms accessible in that part of the world are legally required to implement "highly effective" age assurance measures. In fact, this will not have a UK-only fallout, because it could potentially reshape the digital landscape globally, much in the way the GDPR or General Data Protection Regulation of 2016 has had on how online platforms and services collect and handle user data, in subsequent regulations worldwide. OpenAI The mandatory age verification measures that came into place late last month, are meant to provide a substantial legal assurance of a user's age and consent, the idea being to reduce access to content such as pornography, or anything that encourages self-harm, suicide or eating disorders for instance, on the World Wide Web. Apple There are two sides to this coin. Tech companies and content creators are alarmed by the OSA's new sweeping requirements. If any site accessible in the UK—including social media, search engines, music sites, and adult content providers— does not enforce age checks to prevent children from seeing harmful content, they now face potential fines up to 10% of their revenue for non-compliance. This could very well pressure them into implementing invasive verification systems. Depending how a specific platform does it, methods include scanning your face, credit card, or an identity document, if you want to access content. UK's regulators have been at it for a while, a recent case in point being the Investigatory Powers Act, which we decoded in my Tech Tonic column recently which would have forced tech companies to disable active encryption methods, putting user data at significant risk. There are privacy and access implications of this, something digital rights advocates warn about, detailing that these measures have the potential to create an unprecedented surveillance infrastructure, with these massive databases of personal and biometric information inevitably vulnerable to breaches and misuse. Users must now choose between privacy and access, fundamentally altering the internet's traditionally open nature. 'The Act, which is now coming in enforcement stages exemplifies how well-intended laws can cause unintended consequences on other aspects of technologies. The mandatory use of accredited technology is bound to weaken end-to-end encryption which is the hallmark of a free digital society without which commerce or personal communications systems cannot work. Any of the current age verification methods cannot be imposed without addressing biometric surveillance creep, data breaches and misuse, and increased centralization of user data,' explains a spokesperson of Software Freedom Law Centre India ( in a conversation with us. 'The OSA's age assurance rules require platforms to use facial scans, upload IDs, or verify age through banking or telecom data. These measures raise serious privacy concerns and discourage online anonymity. Larger platforms are testing third-party software for this, but the risk does not disappear, it spreads. User data could now sit with multiple external vendors, increasing the chances of leaks or misuse,' points out Vikram Jeet Singh, Partner at BTG Advaya, a law firm. Possible global implications cannot be ignored, considering the OSA's impact extends far beyond British borders, potentially influencing online speech frameworks worldwide. There can be an argument that while it is effective in some form, it breaks the right to privacy and free speech, while also compromising cybersecurity. Countries such as India, already grappling with content regulation challenges, are likely to be closely watching the UK's approach as a potential model, or cautionary tale. The precedent set by Britain's age verification requirements could normalise similar measures globally, which alongside a fragmented internet where access to information depends on geography, also depends on a willingness to submit to digital surveillance. This is something the spokesperson details, 'Such laws generally have global ripple effects like the GDPR. Companies may choose to adopt UK-compliant policies to avoid the costs related to fragmentation. Countries will emulate such provisions to curb dissent and justify surveillance under the guise of child protection or moral regulation by the state.' What's the way forward? The UK's now complete Online Safety Act effectively represents a watershed moment for internet governance, confronted with often opposing scenarios of fundamental questions about digital rights and any government's commitment to protecting children. The intent of the UK government is commendable, in terms of what it is trying to achieve — the internet as a safe space for children. However, the immediate surge in VPN downloads in the UK on the Apple App Store and Google Play Store suggest, citizens aren't likely to play along. Does that potentially undermine the Act's effectiveness? EMOTIONAL SUPPORT OpenAI says that they are updating ChatGPT (irrespective of which model you use specifically) giving it an ability to detect mental or emotional distress. The AI company wants ChatGPT to work better for users when they want guidance and perhaps a pep talk, than pure facts or information. 'I'm feeling stuck—help me untangle my thoughts' is an example OpenAI mentions, among others, to indicate the GPT models will be more capable of listening to the reasoning of a user's thoughts, rather than just tokenise those words into a response. Newly added are also gentle reminders during long sessions to encourage breaks. OpenAI isn't building this out of its own hat, but instead suggests they've worked with over 90 physicians across over 30 countries (including psychiatrists, paediatricians, and general practitioners) to build custom rubrics for evaluating complex, multi-turn conversations, as well as engaging human-computer-interaction (HCI) researchers and clinicians to give feedback on how well they've identified concerning behaviours, as well as convening an advisory group of experts in mental health, youth development, and HCI. The company admits they haven't gotten it right earlier, case in point being an update earlier this year, which made the model respond in a tone that was too agreeable, bordering on saying what sounded nice instead of what was actually helpful. MAKING NOTES Notes In what is usually the most difficult quarter for iPhone sales, with the spectre of the traditional September refresh looming large, Apple has reported Q3 earnings higher than expectations. The company has reported quarterly revenue of $94 billion, up 10 percent year over year, a June quarter revenue record, and beating $89.3 billion expectations. Apple CEO Tim Cook again emphasised the importance of India, for Apple's growth trajectory. 'The world of mousing and keyboarding around will feel as alien as it does to Gen Z [using] MS-DOS,' the words of David Weston, Microsoft's Corporate Vice President of Enterprise & Security in what is apparently the first of Microsoft's 'Windows 2030 Vision' video series. What does this mean? Since he doesn't elaborate any further than this breadcrumb, I'll lay out the possibility for you — another attempt at Windows which is overloaded with AI, perhaps even more within the OS itself and the apps you use on the computing device, with some element of agentic features that'll utilise natural language understanding and context from a user's data as well as what's on the screen. Ready for the future? MEET'S GETTING SMART Google Meet Google Meet is getting a rather interesting new feature, and it may seem like there's some sorcery to it, instead it is more of attention to details. Google says that if you are now joining a Meet call from your own device such as a laptop, the video meeting platform can detect when you may be doing so in a large conference room-esque physical space. In an attempt to reduce or totally eliminate the problem of sound echo on such a call, Meet will suggest joining the call using something called a 'Companion Mode'. Mind you, this presently only works if you are joining a meet call from your laptop on the Google Chrome web browser — and rolls out for all Google Workspace customers with Google Meet hardware devices. Meet uses your laptop's microphone to intelligently know when you are in a room using an ultrasonic signal. 'This wayfinding feature helps ensure a seamless, echo-free start to your meeting. When you join using the highlighted Companion mode button, you will also be automatically checked into the correct room,' says Google, in an official post. Basically, this will require your Google Workspace admins (basically, your organisation's IT folks) to enable 'Proximity Detection' that will allow that hardware to detect nearby devices, as a feature on the Google Meet Hardware installed in a conference room (for this I am sure there will be typical inertia reasoned around 'compatibility' and 'security' to mask ineptitude). At this point, based on my experiences with IT folks, easier said than done. End of the story.


NDTV
2 hours ago
- NDTV
YouTubers Attacked In Dharmasthala, 4 Hospitalised
Mangaluru (Karnataka): Four men, representing three YouTube channels, including a cameraman, were attacked by a large mob near Dharmasthala in Dakshina Kannada district on Wednesday evening, when they were interviewing a person, amid an ongoing probe by SIT into the alleged mass burial case, police said. The incident occurred close to the residence of a college-goer, who was raped and found dead in 2012 in Dharmasthala town. Ajay Anchan, Abhishek, Vijay and a cameraman, were interviewing Rajat of "Bigg Boss Kannada" when they were assaulted by an estimated 50 to 60 individuals near Pangal Cross here. They also damaged the camera equipment. All four were rushed to Benaka Hospital in Ujire. One of them is reported to be in critical condition, police said. Police have launched an investigation and no arrests have been reported yet. Following the incident at Pangal Cross, two groups started abusing and pelting stones on each other. The police tried to pacify the mob, but when it showed signs of escalation, they resorted to a mild lathi charge to disperse them. The police have beefed up the security in sensitive areas. Dakshina Kannada district Superintendent of Police Arun K urged people to remain calm and said that action would be taken once the identities of the suspects are confirmed. Authorities said the motive behind the incident remains unclear and are reviewing nearby CCTV footage. The Special Investigation Team was formed after the Karnataka Women's Commission chairperson Nagalakshmi Choudhary, several members from civil societies and lawyers demanded an investigation into the allegation of mass burial of people in Dharmasthala. There were allegations that "powerful people" were behind the murder of girls and young women, which prompted the state government to form an SIT. The SIT has so far dug out 13 locations and reportedly found bones and skeletons in two of them.


India Today
5 hours ago
- India Today
Federal court upholds SEC ‘gag rule' in 3-0 ruling over free speech objections
A federal appeals court on Wednesday upheld the US Securities and Exchange Commission's so-called "gag rule," rejecting a claim it illegally silences defendants who want to criticise the regulator after settling civil enforcement a 3-0 decision, the 9th US Circuit Court of Appeals said the rule was not unconstitutional on its face, but could violate the First Amendment depending on how it is rule, reflecting SEC policy dating to 1972, often requires settling defendants to say at least that they neither admit nor deny the regulator's allegations. Twelve petitioners had been appealing the SEC's decision in January 2024 not to amend the rule, including eight people whose SEC settlements triggered petitioner, former Xerox chief financial officer Barry Romeril, took a similar case to the US Supreme Court in 2022 in an appeal backed by billionaire and longtime SEC critic Elon Musk, but that court refused to consider Wednesday's decision, Circuit Judge Daniel Bress said that while some defendants find the rule coercive, they remained free not to settle, and instead to speak out against the also said the SEC had an interest in deciding how to try its own cases, including by giving defendants different options, knowing that scrapping the rule could lead to fewer settlements."Provided that any limitation on speech remains within proper bounds, and given the background ability to waive First Amendment rights at least to some extent, the SEC has an interest in giving defendants the option to agree to a speech restriction as part of a broader settlement agreement," Bress said challenges to applying the rule could still be brought before the SEC brings enforcement cases, while judges consider settlements, or when the SEC reopens settled cases because of alleged petitioners included the New Civil Liberties Alliance, which challenges perceived administrative law senior litigation counsel Peggy Little said in a statement the nonprofit was disappointed. "Past practice does not excuse unconstitutional government action," she SEC had no immediate Commissioner Hester Peirce dissented from the regulator's decision not to amend the found "scant factual basis" for the rule, and said prohibiting denials of wrongdoing "prevents the American public from ever hearing criticisms that might otherwise be lodged against the government, let alone assessing their credibility."- Ends