Fraudsters used advanced AI to impersonate the prime minister, tricking Ann Jensen into taking out a loan she’ll repay over nearly three decades.
A Salisbury resident has fallen victim to a devastating AI-driven investment scam that used deepfake technology to impersonate Prime Minister Sir Keir Starmer. Ann Jensen, a Wiltshire woman, believed she had discovered a promising financial opportunity after viewing what appeared to be a genuine promotional video featuring the Labour leader. However, the enticing offer soon revealed itself as a sophisticated con that left Jensen with a £23,000 debt she will be paying off for the next 27 years.
The scam began innocently enough. Jensen stumbled upon a video online where Sir Keir Starmer appeared to champion a cryptocurrency investment. “He was talking about the benefits of this wonderful opportunity,” Jensen explained, her voice still tinged with disbelief. “If you put in just £200, you could start making significant returns.” Convinced by what seemed like a credible source, Jensen clicked. That click set in motion a nightmare she never anticipated.
Embed from Getty ImagesAt first, everything seemed to go as promised. Jensen’s initial £200 investment appeared to skyrocket to over £2,500, or so the scammers claimed. It wasn’t long before they convinced her to take the next step: a substantial £20,000 loan to “prove financial stability” and access even greater returns. They assured her she would get every penny back, and for a while, the façade held firm.
But the moment the loan’s cooling-off period expired, the illusion shattered. The fraudsters vanished, leaving Jensen with mounting debt and no trace of the supposed financial windfall. The reality hit her like a tidal wave. “It was a physical sensation,” she said, her voice trembling. “It felt like my whole body had turned to liquid like I wasn’t even solid anymore.”
Jensen’s bank delivered the final blow, confirming she was liable for the loan. A letter from the bank spelt it out bluntly: she would be responsible for the entire sum. The betrayal stung, but Jensen refused to blame herself. “I never allowed myself to feel stupid,” she asserted. “I’m not. I was the victim of a crime.”
Her story is far from unique. Investment fraud, especially scams involving AI and deepfake technology, has surged, with victims in the UK alone losing £612 million to such schemes in the past year. These scams are meticulously designed to mimic legitimate financial opportunities, making them almost indistinguishable from real investment platforms. In Jensen’s case, the use of deepfake technology—where AI-generated videos mimic public figures—added an unsettling layer of believability.
The psychological toll is just as damaging as the financial loss. Jensen spoke of the lingering impact: “It’s tainted me for life.” She described feeling haunted by the experience as if the trust she once held for online interactions had been irrevocably shattered. Yet, she remains defiant. “This isn’t about being naïve; it’s about being targeted by criminals who know exactly how to manipulate.”
While Jensen’s story is deeply personal, it also serves as a chilling warning. Scammers are evolving, leveraging cutting-edge technology to ensnare victims with increasingly sophisticated methods. What once might have been a simple email phishing attempt has now morphed into elaborate schemes involving deepfakes of trusted public figures. The level of detail in these operations, from realistic videos to plausible financial scenarios, makes them exceptionally hard to spot.
The broader implications are sobering. If even well-informed individuals can be duped by such scams, how can others protect themselves? Financial experts recommend remaining sceptical of investment opportunities that seem too good to be true and verifying any claims independently, especially if they involve well-known figures. Most importantly, they urge people to pause before making financial commitments and to seek professional advice when in doubt.
Jensen’s courage in sharing her story highlights the urgent need for greater awareness and more robust protections against such crimes. Despite the emotional scars, she remains determined to move forward. “I refuse to let this define me,” she said. “I was a victim, yes, but I’m also a survivor.”
As technology advances, so too do the risks. Stories like Jensen’s underscore the critical importance of vigilance in the digital age. Scammers may be getting smarter, but with the right awareness, individuals can outsmart them.