0:00
This audio is presented by Hacker Nune, where anyone can learn anything about any technology.
0:05
How Glucode turns 3B Plus data points, month into life-saving diabetes health care with Tiger Data,
0:11
by Tiger Data, Creators of TimeScaleDB. This is an installment of our Community Member Spotlight
0:17
series, in which winevite our customers to share their work, spotlight their success,
0:22
and inspire others with new ways to use technology to solve problems.
0:26
In this edition, for Grapadin, Vice President of Engineering at Glucode,
0:30
a diabetes monitoring solution used by 1M Plus patients shares how they migrated one of their
0:36
largest and most critical medical data workloads from a leading document database to Tiger Cloud,
0:41
unlocking lower storage costs, faster ingest, and a cleaner architecture for future growth.
0:47
About the company and team, I'm for Grapadin, Vice President of Engineering at Glucode,
0:52
a diabetes platform used by over 1M patients for glucose monitoring and chronic condition
0:57
tracking. We simplified diabetes care, integrate with EHR workflows, and deliver measurable health
1:03
outcomes. Glucode ingests over 100M new data points every day, primarily time stamp measurements
1:09
and patient events, to power real-time dashboards and analytics for patients and healthcare providers.
1:15
Older devices collect glucose values every 5 minutes, newer ones every minute.
1:20
To meet local data residency requirements, we utilize regional data centers in Germany,
1:25
Ireland, Canada, and the US for storage and analytics. The challenge, scaling time series workloads
1:32
with existing infrastructure, real-time time series workloads force ugly cost, performance
1:37
trade-offs. You can buy your way out of performance bottlenecks, but it gets expensive fast.
1:42
At Glucode, we hit the point where scaling our existing setup to keep up with rising demand
1:47
just wasn't worth the cost. Sluggish ingestion, 74B plus total CGM records and 100M plus new
1:55
readings per day were pushing our document database to its limits. Data bloat, existing
2:00
document database cluster had grown to 30 terabytes, largely due to index overhead.
2:06
Slow queries, 14 day and 90 day rollups were too slow for clinicians, and performance worsened
2:11
as data grew. No retention policies. Every new measurement was stored permanently, further
2:17
driving up costs and footprint. Replacing NoSQL with Tiger Data, as we evaluated alternatives
2:24
to our existing document database, Tiger Data stood out. We already trusted Postgres internally
2:30
and had considered building our OWN time series solution on it, but were uneasy about the risks
2:35
of self-hosting ATEC stack requiring HIPAA, GDPR compliance. In addition, we used a WS as the
2:42
backbone of our digital system, so Tiger Data's model running on EC2 with S3 storage made it easy
2:48
to snap into our existing architecture. With Tiger Cloud, we got the cost and performance benefits
2:54
we needed on a hosted Postgres platform, optimized for time series, managed Postgres with purpose
3:00
built time series extensions, including incremental materialized views and automatic partitioning.
3:06
Cost savings, policy-based compression and retention for automated data lifecycle
3:11
management, unified Postgres architecture resolved ingest and read bottlenecks. We unified
3:17
everything on Tiger Cloud and kept it in a single Postgres platform, with compute and storage
3:21
decoupled. That cleared our ingest, read bottlenecks without our costs exploding, and it gives
3:27
room to scale while shipping more functionality. We designed new time scale DB hyperdables around
3:33
how we actually query per patient per day. We kept only the fields we needed for analytics in the
3:38
hotpath and pushed the rest to cheaper storage. With this shift, query patterns now match the time
3:44
series partitioning relevant for clinicians and patients consuming the data, e, g, show me the last
3:50
90 days, etc. Late arriving data and continuous rights to recent time windows now occur in a normal
3:57
supported pattern rather than being managed on an ad hoc basis. In addition, we built a dedicated
4:03
medical data service, MDS, with direct access to Tiger Cloud. Legacy services and mobile clients
4:10
were updated to talk to MDS rather than with the database directly, simplifying and speeding up
4:14
the ingestion process. It also made it easier to update schema and performance characteristics
4:20
without rippling changes through the entire tech stack. Greater than the end result was excellent,
4:25
ingestion is faster than before, costs are lower greater than than before, and we ended up ahead
4:30
of our original timeline. Pergrapatin, greater than vice president of engineering at glucose
4:36
tiger data cuts costs 40%. Once the migration was done, we went back to the metrics and confirmed
4:42
that tiger data did reduce costs compared to our earlier architecture. Storage got cheaper because
4:47
we were storing less. Compute got cheaper because we could downsize instances more than we expected.
4:53
Compression ratios of approximately 95% and some compression ratios as high as 97%.
5:00
480X faster query speeds from 4 minutes to half a second. More stable ingestion for 100M
5:07
plus new readings per day across tens of billions of existing records. While we have finished
5:12
the core migration from a document database to a unified Postgres platform on AWS, we remain
5:18
focused on optimization. We plan to move our remaining medical data collections into tiger
5:23
clouds so our high volume telemetry and clinical datasets can live in one space, providing better
5:28
performance for patient and provider facing workloads. Thank you for listening to this Hackernoon
5:33
story, read by artificial intelligence. Visit Hackernoon.com to read, write, learn and publish.