7 min read
7 min read

The European Union has reached a landmark political agreement on rules that, for the first time, can make major online platforms legally responsible for certain scams that start on their services when they fail to act after being warned.
Think Meta or TikTok, which carry bogus investment ads and then share the blame when victims lose money.
It represents a significant escalation in Europe’s long-standing campaign to prevent tech giants from disregarding the real-world costs of fraud.
Lawmakers are responding to a surge in investment schemes, impersonation scams, and fake celebrity endorsements circulating on social networks.
For years, platforms earned advertising money while banks and customers ate the losses. Internal documents about huge scam ad revenues made many in Europe feel that enough was enough.
The new rules aim to flip that equation by targeting platforms where revenue is most vulnerable.

Negotiators argued fiercely about who should pay when people are tricked. The compromise is quite simple.
Under the new rules, banks must reimburse customers in defined cases, including when a criminal successfully pretends to be the bank or when a payment is made without the customer’s consent.
Online platforms, including major social media and marketplace services, must compensate banks when they have been alerted to fraudulent ads or listings, fail to remove them, and customers are defrauded as a result. For the first time, everyone in the chain has skin in the game.

Under the new framework, major online platforms and search engines can promote financial services to people in a country only if the company providing those services is properly regulated in that country.
That means fewer shady offshore trading apps slipping into local feeds. This is a quiet yet powerful change, as it forces platforms to verify who is buying financial ad space, rather than letting almost anyone target vulnerable users.

The rules are part of a broader payment and fraud package that also targets banks and other payment providers.
They must implement stronger fraud detection measures, quickly freeze suspicious transfers, and verify that account numbers match the intended recipient’s name.
If those safeguards fail, they pay the price. Combined with platform liability, this shifts fraud prevention into a shared responsibility, not just a banking issue.

Today, many scam victims bounce between banks and platforms, with each side blaming the other. Under the new regime, consumers primarily deal with their bank, which has clearer obligations to refund in specified cases.
Behind the scenes, banks can then recover costs from platforms that ignored warning signs. It will not magically fix every case, but from a user’s perspective, the path to compensation should finally be simplified.

Some European politicians call this a seismic shift for the biggest platforms. They point to reports, based on internal Meta documents, that roughly a tenth of Meta’s 2024 revenue may have come from scam or banned ads and argue that companies have quietly profited from misery.
The message from Europe is blunt. If you host blatant fraud, drag your feet on takedowns, and continue to cash checks, regulators will now have a compelling reason to come after you.

To avoid liability, platforms will need convincing systems to spot and remove scam content. That means stricter checks on new advertisers, more intelligent detection of impersonation accounts, and much faster reaction when users flag fraud.
I expect to see more use of artificial intelligence to catch patterns that humans miss, as well as more human review teams focused on financial abuse, rather than just general content moderation.

From a user perspective, these rules will likely make social feeds look cleaner over time, with fewer miracle investment offers and suspicious trading gurus.
There may be more checks when you try to advertise anything financial, and more hoops to jump through if your post touches money.
Some people will complain about friction, but I prefer a bit of friction to another wave of deepfake get-rich-quick schemes.

The package is not gentle on banks either. They are expected to offer genuine human support when fraud hits, not bury customers in automated menus.
Payment charges should be clearer, and data sharing between banks and other payment providers should be improved.
The idea is that a bank cannot quietly profit from instant payments and then claim helplessness when those same instant systems are used to empty accounts.

The Digital Services Act and Digital Markets Act already showed how Brussels can set global tech standards. These fraud rules build on that pattern.
If they work, regulators in the United Kingdom, United States, and elsewhere will have a ready-made template to follow.
I would not be surprised if social platforms eventually design a single global fraud playbook that meets the strictest European rules by default.

Not everyone is cheering. Some experts fear that platforms will over-remove borderline content to avoid risk, potentially harming legitimate advertisers.
Others say that sophisticated fraudsters will simply move to less-regulated channels, such as encrypted messaging or smaller platforms.
There are also real concerns about privacy as advertisers and users face stronger checks. The key test will be whether scam losses actually decrease, rather than simply remaining the same.
And if you’re curious about how this ties into the bigger picture of online trust, you’ll want to read Altman’s warning that bots are making social media seem fake.
For me, the most crucial shift is psychological. For years, social networks framed scams as unfortunate side effects of an open internet.
This law says something different: if you profit from fraud and fail to act, you share responsibility. That principle could extend to deepfakes, romance scams, and other abuses.
If Europe enforces strict and early regulations, we may finally see platforms treat scam content as a core business risk, rather than background noise.
And if you want to see how these debates are already playing out elsewhere, take a look at Google warns Australia’s teen social media ban may be nearly impossible to police.
What do you think about the new EU rules that social media will be held accountable for any scams/frauds online? Please share your thoughts and drop a comment.
Read More From This Brand:
Don’t forget to follow us for more exclusive content on MSN.
This slideshow was made with AI assistance and human editing.
This content is exclusive for our subscribers.
Get instant FREE access to ALL of our articles.
Dan Mitchell has been in the computer industry for more than 25 years, getting started with computers at age 7 on an Apple II.
We appreciate you taking the time to share your feedback about this page with us.
Whether it's praise for something good, or ideas to improve something that
isn't quite right, we're excited to hear from you.
Stay up to date on all the latest tech, computing and smarter living. 100% FREE
Unsubscribe at any time. We hate spam too, don't worry.

Lucky you! This thread is empty,
which means you've got dibs on the first comment.
Go for it!