Your Instagram feed feels engineered to keep you scrolling—and now Irish regulators want to know if Meta deliberately makes it harder for you to escape the algorithm. On May 5, 2026, Ireland’s media watchdog Coimisiún na Meán launched two formal investigations into Facebook and Instagram, suspecting the platforms violate EU law by steering users away from chronological, non-personalized feeds.
The Dark Pattern Problem
Manipulative interface tricks make healthier feed options harder to find for users.
The probes center on whether Meta uses “dark patterns”—manipulative interface tricks that nudge users toward specific choices. Think buried settings, confusing toggles, or making the healthier option harder to find. EU law requires platforms to offer easily accessible alternatives to algorithm-driven content, especially for Stories and Reels where harmful material can loop endlessly.
“It is unacceptable for platforms to prevent people from using their rights under the law, or to try to manipulate people away from making empowered choices,” said Digital Services Commissioner John Evans. The concern hits particularly hard for young users, where recommender systems can repeatedly push disturbing content into feeds.
Meta Pushes Back
The company disputes breach claims while promising regulatory cooperation.
Meta disagrees with any breach claims, insisting it has “introduced substantial changes” for EU compliance, including options for non-personalized feeds. The company committed to engage with regulators—though this typically signals legal preparation alongside minimal changes.
Real consequences make this probe different from previous investigations. Violations could trigger fines up to 6% of Meta’s global annual turnover. For a company that generated over $130 billion in 2023 revenue, that’s potentially $7.8 billion—enough to fund Ireland’s entire health service.
Your Feed, Your Choice
The Digital Services Act demands straightforward alternatives to algorithmic content curation.
The Digital Services Act, effective since 2023, mandates that massive platforms give users genuine alternatives to algorithmic feeds. No buried settings, no dark patterns, no manipulation. Just straightforward choice between “show me everything chronologically” or “let the algorithm decide.”
This investigation follows a pattern of escalating EU enforcement against Big Tech. The same regulator recently probed TikTok, X, and Shein, while Meta already faced a €251 million GDPR fine in 2024. The message is clear: European regulators have moved beyond strongly worded letters to wielding financial weapons that actually hurt.
Whether you’re doom-scrolling through political outrage or watching your teenager disappear into algorithmic rabbit holes, this case matters. Your right to control what you see shouldn’t require a computer science degree to exercise.





























