Connect with us

Technology

Third-party breach exposes ChatGPT account details

Published

on

Third-party breach exposes ChatGPT account details

NEWYou can now listen to Fox News articles!

ChatGPT went from novelty to necessity in less than two years. It is now part of how you work, learn, write, code and search. OpenAI has said the service has roughly 800 million weekly active users, which puts it in the same weight class as the biggest consumer platforms in the world. 

When a tool becomes that central to your daily life, you assume the people running it can keep your data safe. That trust took a hit recently after OpenAI confirmed that personal information linked to API accounts had been exposed in a breach involving one of its third-party partners.

Sign up for my FREE CyberGuy Report
Get my best tech tips, urgent security alerts and exclusive deals delivered straight to your inbox. Plus, you’ll get instant access to my Ultimate Scam Survival Guide — free when you join my CYBERGUY.COM newsletter.

The breach highlights how even trusted analytics partners can expose sensitive account details. (Kurt “CyberGuy” Knutsson)

Advertisement

What you need to know about the ChatGPT breach

OpenAI’s notification email places the breach squarely on Mixpanel, a major analytics provider the company used on its API platform. The email stresses that OpenAI’s own systems were not breached. No chat histories, billing information, passwords or API keys were exposed. Instead, the stolen data came from Mixpanel’s environment and included names, email addresses, Organization IDs, coarse location and technical metadata from user browsers. 

FAKE CHATGPT APPS ARE HIJACKING YOUR PHONE WITHOUT YOU KNOWING

That sounds harmless on the surface. The email calls this “limited” analytics data, but the label feels like PR cushioning more than anything else. For attackers, this kind of metadata is gold. A dataset that reveals who you are, where you work, what machine you use and how your account is structured gives threat actors everything they need to run targeted phishing and impersonation campaigns.

The biggest red flag is the exposure of Organization IDs. Anyone who builds on the OpenAI API knows how sensitive these identifiers are. They sit at the center of internal billing, usage limits, account hierarchy and support workflows. If an attacker quotes your Org ID during a fake billing alert or support request, it suddenly becomes very hard to dismiss the message as a scam.

OpenAI’s own reconstructed timeline raises bigger questions. Mixpanel first detected a smishing attack on November 8. Attackers accessed internal systems the next day and exported OpenAI’s data. That data was gone for more than two weeks before Mixpanel told OpenAI on November 25. Only then did OpenAI alert everyone. It is a long and worrying silent period, and it left API users exposed to targeted attacks without even knowing they were at risk. OpenAI says it cut Mixpanel off the next day.

Advertisement

The size of the risk and the policy problem behind it

The timing and the scale matter here. ChatGPT sits at the center of the generative AI boom. It does not just have consumer traffic. It has sensitive conversations from developers, employees, startups and enterprises. Even though the breach affected API accounts rather than consumer chat history, the exposure still highlights a wider issue. When a platform reaches almost a billion weekly users, any crack becomes a national-scale problem.

Regulators have been warning about this exact scenario. Vendor security is one of the weak links in modern tech policy. Data protection laws tend to focus on what a company does with the information you give them. They rarely provide strong guardrails around the entire chain of third-party services that process this data along the way. Mixpanel is not an obscure operator. It is a widely used analytics platform trusted by thousands of companies. Yet it still lost a dataset that should never have been accessible to an attacker.

Companies should treat analytics providers the same way they treat core infrastructure. If you cannot guarantee that your vendors follow the same security standards you do, you should not be collecting the data in the first place. For a platform as influential as ChatGPT, the responsibility is even higher. People do not fully understand how many invisible services sit behind a single AI query. They trust the brand they interact with, not the long list of partners behind it.

Attackers can use leaked metadata to craft convincing phishing emails that look legitimate. (Jaap Arriens/NurPhoto via Getty Images)

8 steps you can take to stay safer when using AI tools

If you rely on AI tools every day, it’s worth tightening your personal security before your data ends up floating around in someone else’s analytics dashboard. You cannot control how every vendor handles your information, but you can make it much harder for attackers to target you.

Advertisement

1) Use strong, unique passwords

Treat every AI account as if it holds something valuable because it does. Long, unique passwords stored in a reliable password manager reduce the fallout if one platform gets breached. This also protects you from credential stuffing, where attackers try the same password across multiple services.

Next, see if your email has been exposed in past breaches. Our #1 password manager (see Cyberguy.com/Passwords) pick includes a built-in breach scanner that checks whether your email address or passwords have appeared in known leaks. If you discover a match, immediately change any reused passwords and secure those accounts with new, unique credentials.

Check out the best expert-reviewed password managers of 2025 at Cyberguy.com.

2) Turn on phishing-resistant 2FA

AI platforms have become prime targets, so they rely on stronger 2FA. Use an authenticator app or a hardware security key. SMS codes can be intercepted or redirected, which makes them unreliable during large-scale phishing campaigns.

3) Use strong antivirus software

Another important step you can take to protect yourself from phishing attacks is to install strong antivirus software on your devices. This can also alert you to phishing emails and ransomware scams, helping you keep your personal information and digital assets safe. 

Advertisement

The best way to safeguard yourself from malicious links that install malware, potentially accessing your private information, is to have strong antivirus software installed on all your devices. This protection can also alert you to phishing emails and ransomware scams, keeping your personal information and digital assets safe. 

Get my picks for the best 2025 antivirus protection winners for your Windows, Mac, Android & iOS devices at Cyberguy.com.

PARENTS BLAME CHATGPT FOR SON’S SUICIDE, LAWSUIT ALLEGES OPENAI WEAKENED SAFEGUARDS TWICE BEFORE TEEN’S DEATH

4) Limit what personal or sensitive data you share

Think twice before pasting private conversations, company documents, medical notes or addresses into a chat window. Many AI tools store recent history for model improvements unless you opt out, and some route data through external vendors. Anything you paste could live on longer than you expect.

5) Use a data-removal service to shrink your online footprint

Attackers often combine leaked metadata with information they pull from people-search sites and old listings. A good data-removal service scans the web for exposed personal details and submits removal requests on your behalf. Some services even let you send custom links for takedowns. Cleaning up these traces makes targeted phishing and impersonation attacks much harder to pull off.

Advertisement

While no service can guarantee the complete removal of your data from the internet, a data removal service is really a smart choice. They aren’t cheap, and neither is your privacy. These services do all the work for you by actively monitoring and systematically erasing your personal information from hundreds of websites. It’s what gives me peace of mind and has proven to be the most effective way to erase your personal data from the internet. By limiting the information available, you reduce the risk of scammers cross-referencing data from breaches with information they might find on the dark web, making it harder for them to target you.

Check out my top picks for data removal services and get a free scan to find out if your personal information is already out on the web by visiting Cyberguy.com.

Get a free scan to find out if your personal information is already out on the web: Cyberguy.com.

6) Treat unexpected support messages with suspicion

Attackers know users panic when they hear about API limits, billing failures or account verification issues. If you get an email claiming to be from an AI provider, do not click the link. Open the site manually or use the official app to confirm whether the alert is real.

Events like this show why strengthening your personal security habits matters more than ever. (Kurt “CyberGuy” Knutsson)

Advertisement

7) Keep your devices and software updated

A lot of attacks succeed because devices run outdated operating systems or browsers. Regular updates close vulnerabilities that could be used to steal session tokens, capture keystrokes or hijack login flows. Updates are boring, but they prevent a surprising amount of trouble.

8) Delete accounts you no longer need

Old accounts sit around with old passwords and old data, and they become easy targets. If you’re not actively using a particular AI tool anymore, delete it from your account list and remove any saved information. It reduces your exposure and limits how many databases contain your details.

Kurt’s key takeaway

This breach may not have touched chat logs or payment details, but it shows how fragile the wider AI ecosystem can be. Your data is only as safe as the least secure partner in the chain. With ChatGPT now approaching a billion weekly users, that chain needs tighter rules, better oversight and fewer blind spots. If anything, this should be a reminder that the rush toward AI adoption needs stronger policy guardrails. Companies cannot hide behind transparent emails after the fact. They need to prove that the tools you rely on every day are secure at every layer, including the ones you never see.

Do you trust AI platforms with your personal information? Let us know by writing to us at Cyberguy.com.

CLICK HERE TO DOWNLOAD THE FOX NEWS APP

Advertisement

Sign up for my FREE CyberGuy Report 
Get my best tech tips, urgent security alerts and exclusive deals delivered straight to your inbox. Plus, you’ll get instant access to my Ultimate Scam Survival Guide — free when you join my CYBERGUY.COM newsletter. 

Copyright 2025 CyberGuy.com.  All rights reserved.

Advertisement
Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Technology

Betterment’s financial app sends customers a $10,000 crypto scam message

Published

on

Betterment’s financial app sends customers a ,000 crypto scam message

We’ll triple your crypto! (Limited Time)

Bryan: Betterment is giving back!

We’re celebrating our best-performing year yet by tripling Bitcoin and Ethereum deposits for the next three hours.

For example, if you send $10,000 in Bitcoin or Ethereum, we’ll send you right back $30,000 to your sending Bitcoin or Ethereum address.

Send deposits to these addresses:

Advertisement
Continue Reading

Technology

Fox News AI Newsletter: 10 showstopping CES innovations

Published

on

Fox News AI Newsletter: 10 showstopping CES innovations

NEWYou can now listen to Fox News articles!

Welcome to Fox News’ Artificial Intelligence newsletter with the latest AI technology advancements.

IN TODAY’S NEWSLETTER:

– CES 2026 showstoppers: 10 gadgets you have to see
– Construction giant unveils AI to help prevent job site accidents: ‘It’s essentially a personal assistant’
– Fox News gets exclusive look at company helping businesses nationwide harness AI-powered robots to boost efficiency and fill labor gaps

CES 2026 put health tech front and center, with companies showcasing smarter ways to support prevention, mobility and long-term wellness. (CES)

Advertisement

FUTURE IS NOW: Every January, the Consumer Electronics Show, better known as CES, takes over Las Vegas. It’s where tech companies show off what they’re building next, from products that are almost ready to buy to ideas that feel pulled from the future.

SAFER SITES: Construction equipment giant Caterpillar has unveiled a new artificial intelligence (AI) tool designed to improve job site safety and boost efficiency as the industry grapples with labor shortages.

FUTURE OF WELLNESS: The Consumer Electronics Show, better known as CES, is the world’s largest consumer technology event, and it’s underway in Las Vegas. It takes over the city every January for four days and draws global attention from tech companies, startups, researchers, investors and journalists, of course.

FUTURE OF WORK: As artificial intelligence is rapidly evolving, Fox News got an exclusive look at a company helping businesses nationwide harness AI-powered robots to boost efficiency and fill labor gaps. RobotLAB, with 36 locations across the country and headquartered in Texas, houses more than 50 different types of robots, from cleaning and customer service bots to security bots.

The LG CLOiD robot and the LG OLED evo AI Wallpaper TV are displayed onstage during an LG Electronics news conference at CES 2026 in Las Vegas, Jan. 5, 2026. (REUTERS/Steve Marcus)

Advertisement

COMPUTE CRUNCH: The price tag for competing in the artificial intelligence race is rapidly climbing, fueling demand for advanced computing power and the high-end chips that are needed to support it. Advanced Micro Devices (AMD) CEO Lisa Su said demand for AI computing is accelerating as industries rush to expand their capabilities.

AI GONE WRONG: A California teenager used a chatbot over several months for drug-use guidance on ChatGPT, his mother said. Sam Nelson, 18, was preparing for college when he asked an AI chatbot how many grams of kratom, a plant-based painkiller commonly sold at smoke shops and gas stations across the country, he would need to get a strong high, his mother, Leila Turner-Scott, told SFGate, according to the New York Post. 

DR CHAT: ‘The Big Money Show’ panelists weigh in on a report on people turning to ChatGPT for medical and healthcare questions.

‘FUNDAMENTALLY DEFLATIONARY’: OpenAI Board Chair Bret Taylor discusses artificial intelligence’s potential to change traditional work and its increasing use in healthcare on ‘Varney & Co.’

MIND TRAP ALERT: Artificial intelligence chatbots are quickly becoming part of our daily lives. Many of us turn to them for ideas, advice or conversation. For most, that interaction feels harmless. However, mental health experts now warn that for a small group of vulnerable people, long and emotionally charged conversations with AI may worsen delusions or psychotic symptoms.

Advertisement

A California teenager sought drug-use guidance from a ChatGPT chatbot over several months while preparing for college, his mother told SFGate, according to the New York Post. (Kurt “CyberGuy” Knutsson)

FOLLOW FOX NEWS ON SOCIAL MEDIA

Facebook
Instagram
YouTube
X
LinkedIn

SIGN UP FOR OUR OTHER NEWSLETTERS

Fox News First
Fox News Opinion
Fox News Lifestyle
Fox News Health

DOWNLOAD OUR APPS

Fox News
Fox Business
Fox Weather
Fox Sports
Tubi

WATCH FOX NEWS ONLINE

Fox News Go

Advertisement

STREAM FOX NATION

Fox Nation

Stay up to date on the latest AI technology advancements and learn about the challenges and opportunities AI presents now and for the future with Fox News here.

Advertisement
Continue Reading

Technology

Meta expands nuclear power ambitions to include Bill Gates’ startup

Published

on

Meta expands nuclear power ambitions to include Bill Gates’ startup

These AI projects include Prometheus, the first of several supercluster computing systems, which is expected to come online in New Albany, Ohio, sometime this year. Meta is funding the construction of new nuclear reactors as part of the agreements, the first of which may come online “as early as 2030.” These announcements are part of Meta’s ongoing goal to support its future AI operations with nuclear energy, having previously signed a deal with Constellation to revive an aging nuclear power plant last year.

Financial information for the agreements hasn’t been released, but Meta says that it will “pay the full costs for energy used by our data centers so consumers don’t bear these expenses.”

“Our agreements with Vistra, TerraPower, Oklo, and Constellation make Meta one of the most significant corporate purchasers of nuclear energy in American history,” Meta’s chief global affairs officer, Joel Kaplan, said in the announcement. “State-of-the-art data centers and AI infrastructure are essential to securing America’s position as a global leader in AI.”

Continue Reading
Advertisement

Trending