OpenAI’s ChatGPT Health Is Here
OpenAI presents a new Health space inside ChatGPT. The idea is simple: you talk about health in one place, and (if you choose) you connect your medical records and wellness apps so answers can use your own data. OpenAI says people already ask ChatGPT health and wellness questions at a massive scale: over 230 million users per week, based on its de-identified analysis.
That number explains the product move. OpenAI is turning an existing habit into a “real” feature with extra privacy controls, record connections, and clearer boundaries.
There’s a catch: health data is the most sensitive data most people have. Once you connect it, your history becomes part of the prompt context. That can be useful. It can also be risky.
What OpenAI Actually Launched
OpenAI describes ChatGPT Health as a dedicated Health experience inside ChatGPT. You can upload files, connect apps, and keep health chats separate from your other conversations.
OpenAI highlights a few key rules:
- health chats, connected apps, files, and Health “memories” are stored inside Health and separated from your main ChatGPT space
- OpenAI says Health chats, files, and memories are not used to train its foundation models
- OpenAI says ads are not shown in Health, and Health data is not used for ads
Rollout is limited at launch: OpenAI says Health supports the web and iOS, with Android “coming soon,” and it is not available in the EEA, Switzerland, or the UK for now (likely due to regulatory requirements around health data).
The “Connectors” Are the Point
OpenAI says Health can connect to both medical records and wellness sources, including Apple Health and partners such as MyFitnessPal and WeightWatchers, plus other lifestyle tools.
The biggest step is Medical Records, because it moves from “wellness” into clinical data. Not all connectors carry the same risk. Linking wellness apps (like activity or nutrition trackers) mainly shares lifestyle data, while connecting medical records can expose diagnoses, labs, and treatment history.
Medical Records via b.well (U.S.-Only at Launch)
OpenAI says Medical Records access is built with b.well, and it’s U.S.-only at launch (18+). Reporting notes that b.well can connect to records across a large provider network (OpenAI and press coverage describe it as broad access to U.S. healthcare providers).
This also means “connect my records” may still involve portal logins and uneven coverage, depending on where your care happens and what your providers support. Some setups may feel smooth; others may feel like normal healthcare IT: slow, fragmented, and full of passwords.
Privacy by Separation
OpenAI’s main privacy strategy is compartmentalization: Health stays inside Health.
That helps reduce “blast radius.” A leak of random chats is embarrassing. A leak of lab results, diagnoses, medications, and visit notes can be dangerous.
Still, there’s a key point many users miss: consumer tools are not automatically covered by health privacy rules the way hospitals and insurers are. Several reports stress that users should not assume HIPAA protections apply in the same way once data is shared with a consumer AI product. OpenAI also makes a distinction between consumer Health and a separate enterprise offering for healthcare organizations.
So the practical truth is: privacy here depends on product design, security controls, and OpenAI’s commitments, not on the same legal framework that applies inside many clinical systems.
“Not for Diagnosis” vs Real User Behavior
OpenAI repeats the line that Health is not meant to diagnose or treat.
Real users will still push it in that direction, because that’s what people do when they are anxious, short on time, or stuck waiting for an appointment.
The risk is not only wrong information. The bigger risk is confident wrong information.
A widely reported example: a case where ChatGPT suggested using sodium bromide instead of table salt, followed by bromism symptoms and hospitalization. Coverage highlights how fast a “helpful” answer can become a real safety issue when someone treats it like advice.
Another concern is “grading” or judging health trends in a way that feels medical, even when it is not. Reporting around ChatGPT Health testing shows doctors pushing back on overly confident interpretations of wearable data.
OpenAI’s Safety Pitch: HealthBench
OpenAI is trying to show it takes evaluation seriously. One big part of that story is HealthBench, an open-source benchmark for realistic health conversations.
OpenAI says HealthBench was built with 262 physicians across 60 countries and includes 5,000 multi-turn health conversations graded with physician-written rubrics.
This matters because health is not a quiz. It’s conversation, uncertainty, and knowing when to tell someone to seek real care. Benchmarks like this do not solve safety, but they make the product harder to hand-wave.
What ChatGPT Health Is Best For (And Where It’s Risky)
Use cases where ChatGPT Health can genuinely help:
- translate: explain lab results and visit notes in plain English, and define confusing terms
- organize: summarize long histories across PDFs, portals, and app data into a timeline you can bring to a doctor
- support habits: help plan sleep, workouts, or meals using wellness data, as long as the goal stays “wellness”
Areas where you should treat it as high-risk:
- new, severe, or fast-worsening symptoms (urgent triage belongs to clinicians and emergency services)
- medication changes, dosing, or interactions (too much downside if it guesses)
- any request where you want a “final answer” instead of a question list for a professional
If You Use It, Use It Like a “Second Reader”
Here’s the safest posture: treat Health as a helper for understanding and preparation.
- ask for a summary + questions to ask your clinician
- ask it to list unknowns and red flags
- turn on multi-factor authentication and protect the account like it’s a password vault
How to try ChatGPT Health
ChatGPT Health is rolling out gradually via a waitlist. To request early access, sign up at chatgpt.com/health/waitlist— once enabled, a dedicated “Health” space will appear in the ChatGPT sidebar (Web and iOS). Availability varies by country (it’s not currently available in the EEA, Switzerland, or the UK).
Final Thoughts
ChatGPT Health is part of a larger shift: the chat interface is becoming a front door for personal data, including health. OpenAI’s advantage is distribution: people already use ChatGPT for health questions at a huge scale.
The hard part is trust. In healthcare, trust is not branding. It’s what keeps people safe when the answer matters.
ChatGPT Health can make medical information easier to understand and easier to carry across time. The moment you connect records, it also becomes a new place where sensitive data lives, and a new place where confident mistakes can hurt.