
Discover the latest online threats and cyber security trends impacting businesses and consumers worldwide, brought to you by F-Secure's threat intelligence specialists.
March's F‑Alert explores the risks emerging at the intersection of AI, automation, and online fraud—from hyper-personalized assistants and insecure agent platforms to wrapper app breaches and tax-season scams. Throughout, we provide expert commentary and practical guidance to help navigate these risks.
As more European governments move to restrict children's access to social media, the Netherlands has become the latest to propose a minimum age of 15. In this article, Joel Latto breaks down the policy momentum, the unintended consequences, and the practical question at the heart of the debate: will bans work?
OpenClaw's viral growth—and its lack of basic safeguards—offers a case study in what happens when speed and virality take priority over protection. In this article, Megan Squire breaks down the technical flaws, the systemic risks, and why "ship first, secure later" is even more dangerous in the age of AI agents.
As tax season ramps up in the United States, so does scam activity—but the biggest target may not be who you expect. In this article, we break down how tax scams manipulate stress and urgency, why younger adults are especially susceptible, and how to stay ahead of increasingly sophisticated fraud tactics.
The rapid growth of consumer AI apps has created new convenience—and new exposure. A recent breach of 300M 'Chat & Ask AI' messages underscores how quickly sensitive data can be put at risk. In this article, we break down how the incident occurred, why wrapper apps pose unique security challenges, and how users can better protect their data.
The next phase of automation may not replace humans—but recruit them. 'RentAHuman' signals a shift in how digital systems interact with the physical world. In this article, Khalid Alnajjar explores the security implications of AI‑managed gig work, the potential for abuse, and the accountability gaps that could leave workers exposed.
The race to build more personal AI is accelerating, and Google's latest Gemini update signals how deeply integrated these systems are becoming in users' digital lives. In this article, Timo Salmi explores how deeper personalization increases exposure, particularly when biometric data and long-term digital footprints are involved.
Europe's Social Media Crackdown Gains Momentum
The Netherlands is the latest European country to propose banning social media for children, calling for an "enforceable European minimum age of 15 for social media, with privacy-friendly age verification for young people."
Key facts:
France has already approved a social media ban for children under 15, set to take effect in September, while the UK government is considering a ban for children under 16.
While the harms social media can cause to children are well studied, these bans aren't without drawbacks. ID checks carry privacy risks, and restricting internet services goes against the principles of net neutrality and freedom of expression—both championed by the EU.
More concerning is talk of banning harmful or "polarizing" content. Moderating content in a consistently fair way is nearly impossible, and anyone tasked with doing so would wield significant power over people's online experiences.
This is happening rapidly across many European countries, and I expect more to join those proposing social media bans for children. While the idea may be well intentioned, it almost goes without saying that children will seek—and likely find—ways to circumvent these bans. The real question is whether the added friction in accessing the open internet will be enough to make kids comply.
Joel Latto, Threat Advisor at F‑Secure

AI Platform OpenClaw Goes Viral, Ignoring Basic Security
OpenClaw (formerly Clawdbot, then Moltbot) is an open-source AI agent platform that went viral after launching Moltbook, a social network for agents. Both lack meaningful security controls—ignoring decades of cyber security lessons and enabling bots to rapidly scale early internet-era mistakes.
Key facts:
OpenClaw ships with sandboxing disabled, no authentication required, and full shell access enabled, leading to predictable security breaches such as credential exposure and account takeover.
Many of the agent "skills" that make OpenClaw highly extensible also contain serious vulnerabilities—or are outright malicious—enabling silent data exfiltration and backdoor installation.
Moltbook, the "vibe coded" social network for OpenClaw and other AI agents, likewise prioritized virality over security, exposing 1.5 million authentication tokens within hours of launch.
We've seen how "ship first, secure later" plays out. It took decades to harden early internet protocols. But OpenClaw is replaying that same old movie at 100x speed with autonomous agents that never sleep, can infect one another, and process irreversible financial transactions.
Dr Megan Squire, Threat Intelligence Researcher at F‑Secure

Trending Scam: Tax Season Is Here—and Scammers Are Targeting Gen Z
What's happening:
Many assume tax scams mainly target older adults, but according to the F‑Secure U.S. Scam Intelligence & Impacts Report, Gen Z is the most exposed: 18–25-year-olds report the highest scam victimization rate.
As tax season ramps up, scammers exploit the stress, confusion, and financial pressure younger adults often face—in many cases for the first time.
Tax scams follow a predictable pattern: trigger emotion (panic, fear, excitement about a refund), apply pressure ("act now"), disrupt attention (links, calls, login prompts), capture value (credentials, money, data), then escalate (refund theft, account takeover). Once emotion takes hold, rational checks often disappear.
What to do:
The simplest defense: pause. If a tax-related message triggers fear, urgency, or even relief, stop. Never click links in tax-related texts or emails, call numbers included in those messages, or share sensitive information with online "helpers."
Always verify independently via official websites accessed directly—not through email links. Log in to known tax software accounts separately and take your time. Urgency is the scammer's advantage.
Breach That Matters: Popular 'Chat & Ask AI' Mobile App Exposes 300M Messages
What's happening:
300 million messages from 25 million users of 'Chat & Ask AI'—a mobile app that connects users to AI models like ChatGPT and Gemini—were exposed in a major data breach. The incident stemmed from a basic Google Firebase misconfiguration, with access controls left open rather than a hack.
An independent security researcher discovered he could access millions of chat histories, timestamps, user settings, and more. No usernames or passwords were leaked, but many messages were deeply personal, with users confiding in the AI as they would a trusted friend.
The breach highlights the risks of "wrapper apps"—mobile apps that repackage existing AI services without matching their security standards.
What to do:
This type of exposure is not uncommon. Wrapper apps often rush to market with limited security oversight. Be cautious with third-party AI apps and consider using official platforms like ChatGPT or Gemini instead.
Review privacy policies and user reviews before downloading third-party AI apps and avoid sharing sensitive information.
AI Agents Are Now Hiring Humans to Do What They Can't
RentAHuman is a new platform that lets AI agents hire real people for tasks they can't complete—such as picking up packages, taking photos, or attending meetings. It promises flexible income in an automated future, but accepting tasks from AI agents means entering a system with little oversight and no identity verification.
Key facts:
A request to collect a package in a remote location could be a setup. Delivering a message could make you part of harassment or stalking. In each case, your identity and physical safety are tied to a system with no clear accountability.
Agents could also split criminal operations into small, innocent-looking tasks outsourced to different workers. One registers a country-code domain. Others send promotional emails. Once traffic builds, the site is swapped for a phishing page. The workers' real identities are tied to the scam while the agent remains untraceable—and may disappear without paying.
The risk runs both ways: a hired human providing technical services could steal credentials, install malware, or inject malicious code.
As AI agents increasingly operate online, trust is the missing piece. There's no reliable way to verify whether a request is legitimate or who is responsible if something goes wrong. If a task turns out to be criminal, the worker may be the only traceable link. My advice is this: if you can't see the full context of what you're contributing to or any worker protections, don't accept the task.
Khalid Alnajjar, Senior AI & Data Scientist at F‑Secure

Google's AI Is Becoming More Personal—and More Invasive
Google has announced the beta launch of its new 'Personal Intelligence' feature for Gemini, saying it will make the AI model "more personal, proactive, and powerful." The feature connects apps like Gmail, Google Photos, and Google Search to build a detailed user profile and personalize the experience based on a person's digital footprint.
Key facts:
Google's AI is becoming increasingly personalized, now able to infer appointments from calendar entries and identify license plates from photos. The upside? The better assistants know users, the more helpful they can be. The downside: deeper personalization requires broader data access—and increases privacy risk.
Google has previously faced scrutiny over voice recordings captured by its AI‑powered Google Assistant, activated by text and voice commands ("Hey Google") to manage tasks such as making phone calls and setting reminders.
The digitalization of biometrics—such as fingerprints, facial features, and voice—has improved authentication. But when combined with AI, these immutable identifiers can be exploited for identity theft, financial fraud, and other forms of abuse.
The problem is that the whole is more dangerous than any single piece. A set of photos seems harmless. A shared detail seems harmless. Recorded audio seems harmless. But once combined, they create a detailed profile that can be used for identity fraud, deepfakes, or targeted manipulation in ways no single element could. If a criminal gains access to a deeply personalized Gemini account, they unlock a rich pool of sensitive data.
Timo Salmi, Senior Product Marketing Manager at F‑Secure
