

Tracking and understanding your health data shouldn’t require digging through multiple apps or dealing with complex integrations. Whether it’s steps, heart rate, sleep, or calories, the information is already on your phone, but it’s locked behind interfaces that aren’t exactly conversational or easy to connect with AI tools.
We wanted a way to ask natural questions like “How did my sleep affect my workouts this week?” and get a structured, AI-ready answer instantly, without sacrificing control over our own data.
That’s where we turned to the Model Context Protocol (MCP) — a lightweight, schema-based protocol — to bridge the gap. We built an open-source MCP server in Flutter that runs directly on Android and iOS, exposing health metrics in a structured format any compatible client (like Claude or Cursor) can query and understand.
In this post, we’ll walk you through the problem we set out to solve, the architecture we built, and where this could go next.
AI is incredibly powerful at understanding, analyzing, and summarizing information. However, most of the data it works with comes from the internet, not from you. Imagine if that same intelligence could analyze your own health records to give you quick and meaningful answers about your well-being.
Right now, that’s harder than it should be. Accessing your health data often means digging through apps like Google Fitbit or Apple Health, scrolling through charts, and piecing things together manually. With XLCare, our fully on-device health assistant, those insights already happen privately on your phone. This new project takes a different path: letting your health metrics be accessed directly from your favorite AI client, so you can combine your own data with the full analytical power of today’s most advanced tools.
We wanted something that:
Our idea was to make your phone’s health data accessible via MCP from a Cursor (or another MCP compatible client) client on your PC, so you can ask questions like:
“How many steps did I take yesterday?”
and get an answer from your own device, instantly and in a way AI can understand.
Our system is composed of four key layers that work together to serve structured health data via MCP:
For us, this isn’t just a neat tech demo — it’s a sneak peek at a whole new way of interacting with data.
Once we ran it on real devices, the potential became obvious:
What we’ve built is a first step toward making health insights easier, faster, and a lot more natural. You’re in control, the community can take it further, and the future of local + AI feels closer than ever.
If you want to see exactly how we wired everything together, from reading health data to serving it via MCP, the full implementation is open-source.
In the GitHub repo you’ll find:
Check the README for quick setup instructions.
If you choose to connect this project to a cloud-based AI model (like GPT-5, Claude Sonnet 4, or others), be aware that your health data will be sent to that provider for processing. Always review their privacy policies before sharing sensitive information. For a fully private experience, pair the MCP server with an on-device LLM model instead.
Connecting to a local MCP isn’t exactly plug-and-play. It takes a bit of technical wrangling. That’s the part we want to make vanish into the background.
We also have a few ideas in motion that could take this far beyond what we’ve shown here.
Can’t share all the details yet, but stay tuned.
This project started as an experiment to see if health data could be accessed, understood, and acted on through MCP, all from the device in your pocket. Along the way, we found it’s not just possible, it’s powerful.
By combining a local MCP server with AI clients, we open the door to instant insights, richer analytics, and entirely new ways to interact with personal health data. The potential goes well beyond tracking steps or logging workouts:t it’s about building AI-powered tools that are both personal and extensible.
We’re just scratching the surface. The code is open, the ideas are evolving, and the possibilities are wide open.