The US is now leading a global surge in new gas power plants being built in large part to satisfy growing energy demand for data centers. And more gas means more planet-heating pollution.
Technology
Can AI help someone stage a fake kidnapping scam against you or your family?
You may feel confident in your ability to avoid becoming a victim of cyber scams. You know what to look for, and you won’t let someone fool you.
Then you receive a phone call from your son, which is unusual because he rarely calls. You hear a shout and sounds resembling a scuffle, making you take immediate notice. Suddenly, you hear a voice that you are absolutely certain is your son, screaming for help. When the alleged kidnappers come on the line and demand money to keep your son safe, you are sure that everything is real because you heard his voice.
Unfortunately, scammers are using artificial intelligence (AI) to mimic the voices of people, potentially turning these fake voices into things like kidnapping scams. This particular scam seems to be rare, but it’s happening.
CLICK TO GET KURT’S FREE CYBERGUY NEWSLETTER WITH SECURITY ALERTS, QUICK VIDEO TIPS, TECH REVIEWS AND EASY HOW-TO’S TO MAKE YOU SMARTER
An illustration of a scammer. (Kurt “CyberGuy” Knutsson)
How frequent are fake kidnapping calls enhanced with AI?
Such fake emergency scams occur frequently enough that the Federal Trade Commission (FTC) provided warnings and examples for consumers. Hard numbers that indicate the frequency of these calls aren’t readily available, though, especially for calls known to make use of AI.
Such scams are certainly possible with current AI technology. Fake video and audio of politicians and other famous people are appearing with regularity. Aided by AI, these clips are frighteningly believable.
You may recall the incident in late 2023 involving a fake dental plan advertisement that featured Tom Hanks. AI technology created the video. Hanks had to make a social media post calling out the fake advertisement.
Empty warehouse with a chair. (Kurt “CyberGuy” Knutsson)
MORE: THE ‘UNSUBSCRIBE’ EMAIL SCAM IS TARGETING AMERICANS
How does an AI fake call work?
The AI technology creates a fake by analyzing a sampling of an audio clip of the person it wants to mimic. It uses its ability to interpret incredible amounts of data to take note of multiple characteristics of the person’s voice, allowing it to make a highly realistic fake.
Once the AI is able to create the fake audio, programmers then tell it what to say, creating a personalized message designed to sell dental plans or to convince you that your loved one is in trouble with kidnappers.
Some AI programmers that use the fake audio for helpful purposes — such as for allowing people with medical problems like ALS to regain their “speech” — claim they can mimic a voice with as little as a few minutes of audio clips. However, the more audio that’s available, the more realistic the mimicked voice should sound. Twenty minutes of audio is far better than three, for example.
As AI’s capabilities continue to expand at breakneck speed, you can expect the time requirements to shrink in future years.
WHAT IS ARTIFICIAL INTELLIGENCE (AI)?
An illustration of artificial intelligence. (Kurt “CyberGuy” Knutsson)
MORE: HOW TO GUARD AGAINST BRUSHING SCAMS
Do I have to worry about falling for a fake AI audio kidnapping scheme?
Realistically, the vast majority of people don’t have to worry about a fake kidnapping scheme that originates from AI-generated audio. If your loved one has a lot of video and audio on social media, though, the scammers may be able to find enough source audio to create a realistic fake.
Even though AI makes this type of scam easier to perform, the setup process still remains too time-consuming for most scammers. After all, scammers in this type of scheme are relying on your rapidly expanding fear at receiving this type of call to cause you to miss obvious clues that would tell you it’s a fake.
The scammers may simply have a random child scream and sob uncontrollably, while allowing you to rapidly jump to the conclusion that it’s your child. This is far easier than using AI to try to source and generate audio … at least for now.
A woman surrounded by data. (Kurt “CyberGuy” Knutsson)
MORE: HOW SCAMMERS USE AI TOOLS TO FILE PERFECT-LOOKING TAX RETURNS IN YOUR NAME
Steps you can take to protect yourself from a fake kidnapping scam
Even though the scammers try to gain the upper hand with the suddenness of the fake kidnapping call and by catching you off guard, you have some steps you can take before and after you receive this type of call to prepare and protect yourself.
1. Ask your loved ones to keep you informed about trips: Fake kidnappers may try to convince you that the abduction is taking place outside your city. However, if you know that your loved one did not leave town, you can be confident that the call is probably a fake.
2. Set up a safe word or phrase: Set up a safe word that your loved ones should use if they ever are calling you because of a dangerous situation or because they are under duress. A scammer is not going to know this safe word. If you don’t hear the safe word, you know it’s probably a fake call.
3. Use privacy settings on social media: Ask your family members to limit who can see their social media posts. This would make it harder for a scammer to obtain source audio that’s usable in a fake kidnapping audio call. For more information on maintaining and protecting your online privacy, click here.
4. Try to text your loved one: Either during or immediately after the call, send a text message to your loved one without telling the caller. Ask your loved one to text you back immediately, so you can converse without tipping off the scammers. If you receive a text back, you can be confident the call is a fake. Consider creating a code word that you can use with the entire family. When you send this code word in a text, everyone knows it’s a serious situation that requires an immediate response.
5. Stay calm and think things through: Finally, although it is incredibly difficult to stay calm when you receive this kind of call, it’s important to keep thinking clearly. Do not panic. Regardless of whether it’s a real call or a scam call, panicking is never going to help. Listen for clues that make it obvious the call is a scam. Try to gather some information that can help you make a clear-headed judgment about the legitimacy of the call.
Kurt’s key takeaways
As AI continues to become more readily available and gains sophistication, scammers will be ready to take advantage of it. Perhaps by then, AI will even the playing field by coming up with ways to help us protect ourselves. Until then, taking steps to protect your family, such as by setting up a safe word, can give you some peace of mind.
Are you concerned about how scammers may take advantage of AI to create new scams? Let us know by writing us at Cyberguy.com/Contact
For more of my tech tips & security alerts, subscribe to my free CyberGuy Report Newsletter by heading to Cyberguy.com/Newsletter
Ask Kurt a question or let us know what stories you’d like us to cover.
Answers to the most-asked CyberGuy questions:
Copyright 2024 CyberGuy.com. All rights reserved.
Technology
It’s a new heyday for gas thanks to data centers
Gas-fired power generation in development globally rose by 31 percent in 2025. Almost a quarter of that added capacity is slated for the US, which has surpassed China with the biggest increase of any country. More than a third of that growth in the US is expected to directly power data centers, according to a recent analysis by the nonprofit Global Energy Monitor (GEM).
More gas means more planet-heating pollution
The rush to install more powerful hardware into expanding data centers used for generative AI has led to forecasts of skyrocketing power demand. There’s still a lot of uncertainty about whether AI will become as ingrained in everyday life as tech companies might like, and many proposed data centers could fall flat. Even so, plans to build out more gas plants in the name of AI are stalling efforts to transition to cleaner energy sources.
“There is a risk that this capacity could lock in future emissions and become stranded assets if anticipated electricity demand from AI never materializes,” Jenny Martos, project manager for GEM’s Global Oil and Gas Plant Tracker, said in a press release.
Already, 2026 is shaping up to be a record-smashing year for gas. If all of this year’s proposed projects cross the finish line, it would be an even bigger jump in added capacity than the record set in 2002. That’s pretty remarkable considering the 2000s ushered in America’s so-called “shale gas revolution,” when fracking suddenly unleashed previously hard-to-reach reserves. Gas is now a cheaper power source than coal and creates less carbon pollution when burned. But gas production releases methane, a more potent greenhouse gas than carbon dioxide even though it doesn’t persist in the atmosphere for as long.
Image: Global Energy Monitor
Ramping up electricity generation from gas is also a sharp pivot away from global climate goals. A decade ago, nearly every country on Earth — including its two biggest greenhouse gas polluters, China and the US — signed a historic deal in Paris to limit global warming. The only way to achieve the most ambitious goals set out in the Paris agreement would be to replace fossil fuels with less-polluting alternatives like renewable energy and slash greenhouse gas emissions to net zero by around 2050.
Technology
Thousands of iPhone apps expose data inside Apple App Store
NEWYou can now listen to Fox News articles!
Apple often promotes the App Store as a secure place to download apps. The company highlights strict reviews and a closed system as key protections for iPhone users. That reputation now faces serious questions.
New research shows that thousands of iOS apps approved by Apple contain hidden security flaws. These flaws can expose user data, cloud storage and even payment systems.
The issue is not malware; it’s poor security practices baked directly into the app code.
Sign up for my FREE CyberGuy Report
Get my best tech tips, urgent security alerts, and exclusive deals delivered straight to your inbox. Plus, you’ll get instant access to my Ultimate Scam Survival Guide – free when you join my CYBERGUY.COM newsletter.
APPLE WARNS MILLIONS OF IPHONES ARE EXPOSED TO ATTACK
Cybernews researchers found that many iOS apps store sensitive secrets directly inside app files, where they can be easily extracted. (Kurt “CyberGuy” Knutsson)
What researchers discovered inside iOS apps
Security researchers at Cybernews, a cybersecurity research firm, analyzed the code of more than 156,000 iPhone apps. That represents about 8% of all apps available worldwide.
Here is what they found:
- Over 815,000 hidden secrets inside app code
- An average of five secrets per app
- 71% of apps leaked at least one secret
These secrets include passwords, API keys and access tokens. Developers place them directly inside apps, where anyone can extract them. According to Cybernews researcher Aras Nazarovas, this makes attackers’ jobs much easier than most users realize.
What are hardcoded secrets in simple terms?
A hardcoded secret is sensitive information saved directly inside an app instead of being protected on a secure server. Think of it like writing your bank PIN on the back of your debit card. Once someone downloads the app, they can inspect its files and pull out those secrets. Attackers do not need special access or advanced hacking tools. Both the Cybersecurity and Infrastructure Security Agency and the Federal Bureau of Investigation warn developers not to do this. Yet it is happening at a massive scale.
Cloud storage leaks exposed huge amounts of data
One of the most serious problems involves cloud storage. More than 78,000 iOS apps contained direct links to cloud storage buckets. These buckets store files such as photos, documents, receipts and backups. In some cases, no password was required at all. Researchers found:
- 836 storage buckets are fully open to the public
- Over 76 billion exposed files
- More than 406 terabytes of leaked data
This data included user uploads, registration details, app logs and private records. Anyone who knew where to look could view or download it.
APPLE PATCHES TWO ZERO-DAY FLAWS USED IN TARGETED ATTACKS
This chart shows the most common types of hardcoded secrets found inside iOS apps, with Google-related keys appearing most often, according to Cybernews research. (Cybernews)
Firebase databases were also left open
Many iOS apps rely on Google Firebase to store user data. Cybernews found more than 51,000 Firebase database links hidden in app code. While some were protected, over 2,200 had no authentication. That exposed:
- Nearly 20 million user records
- Messages, profiles, and activity logs
- Databases that are mostly hosted in the U.S.
If a Firebase database is not locked down, attackers can browse user data like a public website.
Payment and login systems were at risk too
Some of the leaked secrets were far more dangerous than analytics or ads. Researchers discovered secret keys for:
- Stripe, which handles payments and refunds
- JWT authentication systems that control logins
- Order management tools used by shopping apps
A leaked Stripe secret key can allow attackers to issue refunds, move money or access billing details. Leaked login keys can let attackers impersonate users or take over accounts.
AI and social apps were among the worst offenders
Some of the apps with the largest leaks were related to artificial intelligence. According to VX Underground, security firm CovertLabs identified 198 iOS apps leaking user data. The worst known case was Chat & Ask AI by Codeway. Researchers say it exposed chat histories, phone numbers and email addresses tied to millions of users. Another app, YPT – Study Group, reportedly leaked messages, user IDs and access tokens. CovertLabs tracks these incidents in a restricted repository called Firehound. The full list of affected apps has not been publicly released, and researchers say the data is limited to prevent further exposure and to give developers time to fix security flaws.
MALICIOUS GOOGLE CHROME EXTENSIONS HIJACK ACCOUNTS
This example shows how sensitive keys like Google API credentials and Stripe payment secrets can be stored directly inside an iOS app’s files, where they are easy to extract. (Cybernews)
Why Apple’s App review can miss hidden security risks
Apple reviews apps before they appear in the App Store. However, the review process does not scan app code for hidden secrets. If an app behaves normally during testing, it can pass review even if sensitive keys are buried inside its files. This creates a gap between Apple’s security claims and real-world risks. Removing leaked secrets is not simple for developers. They must revoke old keys, create new ones and rebuild parts of their apps. That can break features and delay updates. Even though Apple says most app updates are reviewed within 24 hours, some updates take weeks. During that time, vulnerable apps can remain available.
CyberGuy contacted Apple for comment, but did not receive a response before publication.
Ways to stay safe right now
You cannot easily inspect an app for hidden secrets. Apple does not provide tools for that. Still, you can reduce your risk and limit exposure by being selective and cautious. These steps help reduce the risk if an app leaks data behind the scenes.
1) Stick to established app developers
Well-known developers tend to have stronger security teams and better update practices. Smaller or unknown apps may rush features to market and overlook security basics. Before downloading, check how long the developer has been active and how often the app is updated.
2) Review and limit app permissions
Many apps ask for more access than they need. Location, contacts, photos and microphone access all increase the risk of data leaks. Go into your iPhone settings and remove permissions that are not essential for the app to work.
3) Delete apps you no longer use
Unused apps still retain access to data you shared in the past. They may also store information on remote servers long after you stop opening them. If you have not used an app in months, remove it. Here’s how: Open Settings, tap General, select iPhone Storage, and scroll through the list of apps to see when each one was last used. Tap any app you no longer need and select Delete App to remove it and reduce ongoing data exposure.
4) Be cautious with personal and financial details
Avoid entering sensitive information unless it is absolutely necessary. This includes full names, addresses, payment details and private conversations. AI apps are especially risky if you share deeply personal content.
5) Use a password manager for every account
A password manager creates strong, unique passwords for each app and service. This prevents attackers from accessing multiple accounts if one app leaks data. Never reuse passwords tied to your email address.
Next, see if your email has been exposed in past breaches. Our No. 1 password manager pick includes a built-in breach scanner that checks whether your email address or passwords have appeared in known leaks. If you discover a match, immediately change any reused passwords and secure those accounts with new, unique credentials.
Check out the best expert-reviewed password managers of 2026 at Cyberguy.com.
6) Change passwords tied to exposed apps
If an app uses your email address for login, change that password immediately. Do this even if there is no confirmation of a breach. Attackers often test leaked credentials across other services.
7) Consider using a data removal service
Some leaked data ends up with data brokers that sell personal information online. A data removal service can help find and remove your details from these databases. This reduces the chance that exposed app data gets reused for scams or identity theft.
While no service can guarantee the complete removal of your data from the internet, a data removal service is really a smart choice. They aren’t cheap, and neither is your privacy. These services do all the work for you by actively monitoring and systematically erasing your personal information from hundreds of websites. It’s what gives me peace of mind and has proven to be the most effective way to erase your personal data from the internet. By limiting the information available, you reduce the risk of scammers cross-referencing data from breaches with information they might find on the dark web, making it harder for them to target you.
Check out my top picks for data removal services and get a free scan to find out if your personal information is already out on the web by visiting Cyberguy.com.
Get a free scan to find out if your personal information is already out on the web: Cyberguy.com.
8) Monitor your accounts for unusual activity
Watch for unexpected emails, password reset notices, login alerts, or payment confirmations. These can signal that leaked data is already being abused. Act quickly if something looks off.
9) Pause use of risky AI and chat apps
If you use AI apps for private conversations, consider stopping until the developer confirms security fixes. Once data is exposed, it cannot be pulled back. Avoid sharing sensitive details with apps that store conversations remotely.
Kurt’s key takeaways
Apple’s App Store still offers important protections, but this research shows it is not foolproof. Many trusted iPhone apps quietly expose data due to basic security mistakes. Until app reviews improve, you need to stay alert and limit how much data you share.
How many apps on your iPhone have access to information you would not want exposed? Let us know by writing to us at Cyberguy.com.
CLICK HERE TO DOWNLOAD THE FOX NEWS APP
Sign up for my FREE CyberGuy Report
Get my best tech tips, urgent security alerts, and exclusive deals delivered straight to your inbox. Plus, you’ll get instant access to my Ultimate Scam Survival Guide – free when you join my CYBERGUY.COM newsletter.
Copyright 2026 CyberGuy.com. All rights reserved.
Technology
Apple just had a record-breaking iPhone quarter
Apple says the iPhone had its “best-ever” quarter, with revenue hitting more than $85.3 billion over the past few months. The company announced the news as part of its Q1 2026 earnings report, which also revealed record-breaking revenue of $143.8 billion, up 16 percent when compared to the same time last year.
“iPhone had its best-ever quarter driven by unprecedented demand, with all-time records across every geographic segment, and Services also achieved an all-time revenue record, up 14 percent from a year ago,” Apple CEO Tim Cook says in the press release.
Apple’s services revenue, which includes subscriptions like Apple Music, iCloud, and Apple TV surged 14 percent year over year, while Mac and wearable revenue were down.
In the coming months, Apple will bring promised AI-powered personalization features to Siri. Apple is partnering with Google to power this upgrade, which will use a custom version of Google’s Gemini AI model.
The company is also acquiring an AI startup, called Q.ai, for $2 billion, according to the Financial Times. Though there aren’t any details on how Apple plans to use the startup’s technology, the Financial Times reports that Q.ai’s patents “show its technology being used in headphones or glasses, using ‘facial skin micro movements’ to communicate without talking.”
-
Illinois1 week agoIllinois school closings tomorrow: How to check if your school is closed due to extreme cold
-
Pennsylvania4 days agoRare ‘avalanche’ blocks Pennsylvania road during major snowstorm
-
Science1 week agoContributor: New food pyramid is a recipe for health disasters
-
Technology1 week agoRing claims it’s not giving ICE access to its cameras
-
Science1 week agoFed up with perimenopause or menopause? The We Do Not Care Club is here for you
-
Movie Reviews1 week ago
Movie Review: In ‘Mercy,’ Chris Pratt is on trial with an artificial intelligence judge
-
Politics1 week agoSupreme Court appears ready to keep Lisa Cook on Federal Reserve board despite Trump efforts to fire her
-
News1 week agoVideo: Jack Smith Defends His Trump Indictments During House Hearing