Connect with us

Technology

Is the FTC calling you? Probably not. Here's how to avoid a phone scam targeting you

Published

on

Is the FTC calling you? Probably not. Here's how to avoid a phone scam targeting you

Just when you thought phone scams couldn’t get any trickier, there’s another one out there that you must watch out for. 

And this one is even more sophisticated because the government agency you’re meant to report these types of scams to – the FTC – is being impersonated.

Woman on the phone (Kurt “CyberGuy” Knutsson)

One reason this impersonation has been on the rise is the progression of artificial intelligence. 

AI voice-over tools and deep-fake technology make it even easier for impersonators to commit crimes. But you can protect yourself. Here’s what you need to know.

Advertisement

CLICK TO GET KURT’S FREE CYBERGUY NEWSLETTER WITH SECURITY ALERTS, QUICK VIDEO TIPS, TECH REVIEWS AND EASY HOW-TO’S TO MAKE YOU SMARTER

What is the FTC impersonation scam?

Like a phishing scam, the FTC impersonation scam involves scammers pretending to be Federal Trade Commission employees calling your number and trying to convince you to wire money to them. These scammers have become very clever in how they target their victims.

First, they may spoof an FTC phone number so that when your phone rings, and you see your caller ID, it says FTC and/or a title related to a government agency.

Next, the scammer may impersonate an actual FTC employee. So, when you ask for their name to check whether or not it’s a scam (which you should always do!), your quick internet search will reveal an actual FTC employee. While checking whether or not this is an actual scam, some victims might not think twice after checking this.

Finally, they’ll tell you that you owe money to them. While it’s not clear exactly what they are saying – perhaps they are telling you that you owe an invoice, debt or a fine – some scammers even tell you to send money in other peculiar ways, such as going to Bitcoin ATMs or withdrawing cash to pay them in person.

Advertisement

These scams have been successful, and this year, the median loss per call has risen to $7,000.

SCAMMERS ARE USING FAKE NEWS, MALICIOUS LINKS TO TARGET YOU IN AN EMOTIONAL FACEBOOK PHISHING TRAP

Man on phone

Man on the phone (Kurt “CyberGuy” Knutsson)

MORE: HOW SCAMMERS USE AI TOOLS TO FILE PERFECT-LOOKING TAX RETURNS IN YOUR NAME

What the FTC and FBI are doing about it

Now, the FTC is catching onto these scams, warning and educating the public, and cracking down on perpetrators.

To start, the FTC said in a press release that it “will never send consumers to a Bitcoin ATM, tell them to go buy gold bars, or demand they withdraw cash and take it to someone in person.” The agency added that it will “never contact consumers to demand money, threaten to arrest or deport them, or promise a prize.”

Advertisement

Secondly, due to public outcry against AI and the increase in the number of complaints to the FBI, the FTC is working with the government to utilize tools to “detect, deter, and to catch these impersonators” while proposing strict legislation that would ban government and impersonation fraud, thereby “enabling the FTC to directly file federal court cases aimed at forcing scammers to return the money they made from government or business impersonation scams.”

How to keep yourself safe

1. Stay updated: The first step in protecting yourself from government impersonation fraud is simply being educated about it. Keep yourself updated about these scams and other scams by reading the news and subscribing to sites like Cyberguy. Again, the FTC reports that it will never call you and ask you to wire money in any way, shape or form, so if you get a call from them or another government agency asking for this, it’s a scam.

But due to the 22% increase in reported losses to online crime compared to 2022 based on the FBI’s 2023 Internet Crime Report, they’re proposing other strategies to keep yourself safe from scams by following the tips we list below:

2. Stay calm: If you think a scammer is calling you, it’s important to stay calm and not let yourself get emotional. It’s very easy to get worked up when you’re in the middle of this situation. If it happens, you can always say you will get back to them and hang up the phone. If they become aggressive and try to exploit you, tell them you will report them and hang up.

Advertisement

3. Cautionary note: Do not contact unknown telephone numbers provided in pop-ups, texts or emails.

4. Safety reminder: Do not download software at the request of unknown individuals who contact you.

5. Security precaution: Do not allow unknown individuals access to your computer.

6. Don’t click that link: Do not click on unsolicited pop-ups on your computer, links sent via text messages or email links and attachments.

7. Have good antivirus software: As scammers weaponize AI and deep-fake technologies, robust antivirus software becomes your first line of defense. The best way to protect yourself from clicking malicious links that install malware that may access your private information is to have antivirus protection installed on all your devices. This can also alert you of any phishing emails or ransomware scams. Get my picks for the best 2024 antivirus protection winners for your Windows, Mac, Android & iOS devices.

Advertisement
WOMAN talks on phone

Woman on the phone (Kurt “CyberGuy” Knutsson)

MORE: HOW A FLAW IN IPHONE’S SECURITY COULD LEAVE YOU LOCKED OUT

Kurt’s key takeaways

Scammers will continue finding new ways to prey on their victims, and AI has made it easier for them to do so. One of the major concerns about AI is that it will progress faster than agencies like the FTC and FBI can keep up with. Therefore, it is important to take measures to protect yourself and avoid falling victim.

Have you ever received a suspicious call or encountered an impersonation attempt? How did you handle it? Let us know by writing us at Cyberguy.com/Contact.

For more of my tech tips & security alerts, subscribe to my free CyberGuy Report Newsletter by heading to Cyberguy.com/Newsletter.

Advertisement

Ask Kurt a question or let us know what stories you’d like us to cover.

Answers to the most asked CyberGuy questions:

Copyright 2024 CyberGuy.com. All rights reserved.

Continue Reading
Advertisement
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Technology

Amazon says its Prime deliveries are getting even faster

Published

on

Amazon says its Prime deliveries are getting even faster

To me, Prime’s promise of two-day shipping is more of an added bonus to Prime Video and stuff like Fallout. But it’s become an expectation, leading other retailers like Walmart and Target to roll out faster shipping options of their own.

Now, Amazon says its deliveries are getting even faster, announcing that it delivered over 2 billion items the same or next day to Prime members during the first three months of 2024, breaking its record for 2023. The company says it delivered almost 60 percent of Prime orders the same or next day in 60 of the biggest metropolitan areas in the US.

If you buy from Amazon, have you noticed any differences lately? Same-day and next-day options seem to be more widely available, but it’s hard to tell how that applies to different items in different places or whether the associated costs are worth it based on reports of warehouse injuries and workers organizing for better conditions.

The control Amazon has over shipping and fulfillment has helped make it the target of a lawsuit from the Federal Trade Commission. The agency alleges Amazon engages in anticompetitive behavior by unfairly limiting which sellers are eligible for Prime shipping and coercing companies into using its fulfillment services. Amazon claims that the FTC’s efforts could result in “slower or less reliable” Prime shipping for customers.

In 2019, Amazon said it was spending billions to build up an in-house fulfillment operation covering planes, trucks, drones, and robots to rival FedEx and UPS and enable these one-day-or-less deliveries. And last year, it introduced a program that allows sellers to ship their products directly from factories — regardless of whether they’re going to one of Amazon’s many warehouses.

Advertisement
Continue Reading

Technology

How ‘Yahoo Boys’ use real-time face-swapping to carry out elaborate romance scams

Published

on

How ‘Yahoo Boys’ use real-time face-swapping to carry out elaborate romance scams

We’ve all heard of catfish scams – when someone pretends to be a lover on the other side of the screen, but instead, they aren’t who they say they are once their real face is revealed. Now, there’s a similar scam on the rise, and it’s much more sophisticated because scammers can fake the face, too. The scam is known as the “Yahoo Boys” scam, and it’s taking “catfishing” to a whole new level.

CLICK TO GET KURT’S FREE CYBERGUY NEWSLETTER WITH SECURITY ALERTS, QUICK VIDEO TIPS, TECH REVIEWS AND EASY HOW-TO’S TO MAKE YOU SMARTER

Woman on laptop making a heart sign (Kurt “CyberGuy” Knutsson)

How does deepfake technology work?

Deepfake technology uses AI to allow people to impersonate others over audio or video. The technology is essentially able to replicate someone’s face, facial expressions, gestures, voice, etc., so that the scammer can pretend to be someone they aren’t with almost perfect accuracy. Although deepfake technology has some intentions for good – like in the film industry or advertising (though there are some debates regarding the ethics of this) – it’s generally used for more malicious purposes than anything else.

Because it’s very difficult to tell whether or not a deepfake is the real person or a deepfake (and also because deepfakes are relatively new), deepfakes can have the potential to do a lot of damage. From the larger implications of it being used to spread inaccurate news stories that can sway public opinion and political processes to it being used to inflict damage on individuals with scams, it’s important to know what to watch out for.

Advertisement
How ‘Yahoo Boys’ use real-time face-swapping to carry out elaborate romance scams

Woman with a scan on her face (Kurt “CyberGuy” Knutsson)

MORE: EXPOSING THE TOP SCAMS TARGETING COSTCO SHOPPERS

What is the ‘Yahoo Boys’ scam?

The “Yahoo Boys” scam involves a group of sophisticated cybercriminals, primarily based in Nigeria, who are using this technology to conduct what’s otherwise known as romance scams. Like some catfishing attempts, they first build trust with victims through personal and romantic interactions over messaging and, eventually, video calls, where they then manipulate their appearances in real time to match the description and profile they may have shared with them up until that point.

By doing this, they can trick the victim into trusting them even more. This is because most of us still use video as a way to verify a person’s identity, when messaging isn’t convincing enough. Finally, when the timing is right, the victims are often persuaded into transferring money based on various fabricated scenarios, leading to significant financial losses. In fact, the FBI reported over $650 million lost to romance scams like these.

How ‘Yahoo Boys’ use real-time face-swapping to carry out elaborate romance scams

Man on his cellphone (Kurt “CyberGuy” Knutsson)

MORE: DON’T FALL FOR THESE SNEAKY TAX SCAMS THAT ARE OUT TO STEAL YOUR IDENTITY AND MONEY

How does it actually work?

In the case of the “Yahoo Boys” scam, the scammers do their dirty work by:

Advertisement

Step 1: The scammer will use two smartphones or a combination of a smartphone and a laptop. One device is used to conduct the video call with the victim, while the other runs face-swapping software.

Step 2: On a secondary device, the scammer activates face-swapping software. This device’s camera films the scammer’s face, and the software adds a digital mask over it. This mask is a realistic replica of another person’s facial features, which the scammer has chosen to impersonate. The software is sophisticated enough to track and mimic the scammer’s facial movements and expressions in real time, altering everything from skin tone and facial structure to hair and gender to match the chosen identity.

Step 3: For the video call, the scammer uses a primary device with its rear camera aimed at the secondary device’s screen. This screen shows the deepfake – the digitally altered face. The rear camera captures this and sends it to the victim, who sees the deepfake as if it’s the scammer’s actual face. To make the illusion more convincing, the devices are stabilized on stands, and ring lights provide even, flattering lighting. This setup ensures that the deepfake appears clear and stable, tricking the victim into believing they’re seeing a real person.

Step 4: Throughout the call, the scammer speaks using their own voice, although in some setups, voice-altering technology might also be used to match the voice to the deep-faked face. This comprehensive disguise allows the scammer to interact naturally with the victim, reinforcing the illusion.

Though each deepfake scam is different, having a basic level of understanding in terms of how scams like these work can help you recognize them.

Advertisement
How ‘Yahoo Boys’ use real-time face-swapping to carry out elaborate romance scams

Scammer typing on a keyboard (Kurt “CyberGuy” Knutsson)

How to stay safe from deepfake scams

To protect yourself from deepfake scams like the “Yahoo Boys” scam, here’s what you can do:

Verify identities: Always confirm the identity of individuals you meet online through video calls by asking them to perform unpredictable actions in real time, like writing a specific word on paper and showing it on camera.

Be skeptical of unusual requests: Be cautious if someone you’ve only met online requests money, personal information or any other sensitive details.

Enhance privacy settings: Adjust privacy settings on social media and other platforms to limit the amount of personal information available publicly, which can be used to create deepfake content.

Use secure communication channels: Prefer secure, encrypted platforms for communications and avoid sharing sensitive content over less secure channels.

Advertisement

Educate yourself about deepfakes: Stay informed about the latest developments in deepfake technology to better recognize potentially manipulated content.

Report suspicious activity: If you encounter a potential scam or deepfake attempt, report it to the relevant authorities or platforms to help prevent further incidents.

By following these guidelines, you can reduce your risk of falling victim to sophisticated digital scams and protect your personal and financial information from falling into the hands of these scammers.

How ‘Yahoo Boys’ use real-time face-swapping to carry out elaborate romance scams

Woman talking on her cellphone (Kurt “CyberGuy” Knutsson)

MORE: CAN AI HELP SOMEONE STAGE A FAKE KIDNAPPING SCAM AGAINST YOUR FAMILY

Protecting your identity in the age of deepfakes

As deepfake technology becomes more accessible and convincing, the risk of identity theft increases. Scammers can use stolen personal information to create more believable deepfakes, making it harder for you to detect fraud. Furthermore, the sophistication of deepfakes may allow criminals to bypass biometric security measures, potentially granting them unauthorized access to your personal accounts and sensitive financial information. This is where identity theft protection services become invaluable.

Advertisement

Identity Theft companies can monitor personal information like your Social Security Number, phone number and email address and alert you if it is being sold on the dark web or being used to open an account. They can also assist you in freezing your bank and credit card accounts to prevent further unauthorized use by criminals.

One of the best parts of using some services is that they might include identity theft insurance of up to $1 million to cover losses and legal fees and a white-glove fraud resolution team where a U.S.-based case manager helps you recover any losses. See my tips and best picks on how to protect yourself from identity theft.

By subscribing to a reputable identity theft protection service, you can add an extra layer of security, ensuring that your digital presence is monitored and protected against the ever-evolving tactics of cybercriminals like the “Yahoo Boys.”

MORE: UNFORGETTABLE MOTHER’S DAY GIFTS 2024

Kurt’s key takeaways

The main targets of the “Yahoo Boys” scam are individuals who establish online romantic relationships with the scammers. Because these victims are often emotionally invested and therefore more susceptible to manipulation, there’s not necessarily a specific age or demographic that can become a victim. And because we know the root of all these scams is the deepfake technology, truly anyone can become a target. So, we hope this helps, but also, be sure to spread the word to friends and family.

Advertisement

Have you heard of any other types of deepfake scams? Or do you know anyone who has been a victim of scams like this? Let us know by writing us at Cyberguy.com/Contact.

For more of my tech tips & security alerts, subscribe to my free CyberGuy Report Newsletter by heading to Cyberguy.com/Newsletter.

Ask Kurt a question or let us know what stories you’d like us to cover.

Answers to the most asked CyberGuy questions:

Copyright 2024 CyberGuy.com. All rights reserved.

Advertisement

Continue Reading

Technology

Financial Times signs licensing deal with OpenAI

Published

on

Financial Times signs licensing deal with OpenAI

The Financial Times has struck a deal with OpenAI to license its content and develop AI tools, the latest news organization to work with the AI company. 

The FT writes in a press release that ChatGPT users will see summaries, quotes, and links to its articles. Any prompt that returns information from the FT will be attributed to the publication. 

In return, OpenAI will work with the news organization to develop new AI products. The FT already uses OpenAI products, saying it is a customer of ChatGPT Enterprise. Last month, the FT released a generative AI search function on beta powered by Anthropic’s Claude large language model. Ask FT lets subscribers find information across the publication’s articles. 

Financial Times Group CEO John Ridding says that even as the company partners with OpenAI, the publication continues to commit to “human journalism.”

“It’s right, of course, that AI platforms pay publishers for the use of their material,” Ridding says. He adds that “it’s clearly in the interests of users that these products contain reliable sources.”

Advertisement
Continue Reading

Trending