In a significant move aimed at bolstering online safety, Elon Musk’s social media platform, X, has pledged to block UK access to accounts associated with banned terrorist organisations. This commitment, announced in collaboration with the UK’s communications regulator Ofcom, comes as part of ongoing efforts to combat the proliferation of terrorist and hate-related content online. With the UK grappling with rising concerns over such material, this agreement marks a critical step towards ensuring user safety in the digital space.
Stricter Measures for Content Monitoring
Under the terms of the agreement, X will implement robust measures to tackle illegal terrorist and hate content. The company has committed to reviewing at least 85% of flagged content within a tight 48-hour window. This rapid response aims to ensure that harmful material is swiftly addressed, thereby reducing the potential risks to users.
Oliver Griffiths, the director of Ofcom’s online safety group, expressed optimism about the partnership, stating, “Following intensive engagement carried out by Ofcom’s online safety team, X have committed to implementing stronger protections for UK users, which we will now monitor closely.” His remarks underscore the urgency of the situation, particularly in light of a recent uptick in hate crimes targeting the Jewish community in the UK.
Action Against Terrorist Content
The recent agreement will see X proactively blocking access to accounts that share illegal terrorist content linked to organisations banned by the UK government. This initiative is part of a broader strategy outlined by the UK’s Online Safety Act, which aims to shield citizens from illegal online material.

In addition to the ban on terrorist-linked accounts, X will also seek expert advice on how to manage user reports of suspected illegal content. This collaborative approach highlights the platform’s intention to enhance its content moderation practices significantly.
Ongoing Concerns and Criticism
Despite these positive steps, concerns remain about the effectiveness of X’s moderation strategies. Danny Stone, chief executive of the Antisemitism Policy Trust, acknowledged the agreement as a “good start” but noted that X still falls short in its efforts to combat racism on the platform. Meanwhile, Adam Hadley, executive director of Tech Against Terrorism, praised the announcement as a “powerful example of what constructive dialogue between regulators and platforms can deliver.”
X has faced ongoing scrutiny since Musk acquired the platform for $44 billion (£33 billion) in 2022. The platform, previously known as Twitter, has been accused of exacerbating hate speech and extremist content, particularly following the riots triggered by the Southport murders in 2024.
The Path Ahead
As Ofcom continues to investigate the use of manipulated images on X, concerns about content moderation persist. The platform declined to comment on these ongoing issues, but stakeholders remain hopeful that the commitments made will lead to a safer online environment.

Why it Matters
The commitment from X to block accounts linked to extremist groups and enhance content moderation is more than just a regulatory checkbox; it represents a pivotal moment in the ongoing struggle against online hate and terrorism. As digital platforms continue to wield significant influence over public discourse, the measures adopted by X could set a precedent for how social media companies address dangerous content. The collaboration between regulators and tech giants is crucial in shaping a safer online landscape, ensuring that the internet remains a space for constructive dialogue rather than a breeding ground for hate.