<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.10.0">Jekyll</generator><link href="https://sagrd.github.io/feed.xml" rel="self" type="application/atom+xml" /><link href="https://sagrd.github.io/" rel="alternate" type="text/html" /><updated>2026-02-11T08:37:59+00:00</updated><id>https://sagrd.github.io/feed.xml</id><title type="html">Sagar Dhungel</title><subtitle>Sagar Dhungel - A Data Science and Engineering Consultant.
</subtitle><author><name>Sagar Dhungel</name></author><entry><title type="html">Large Scale Health Data Monitering With Databricks</title><link href="https://sagrd.github.io/large-scale-health-data-monitering-with-databricks.md.html" rel="alternate" type="text/html" title="Large Scale Health Data Monitering With Databricks" /><published>2025-12-21T00:00:00+00:00</published><updated>2025-12-21T00:00:00+00:00</updated><id>https://sagrd.github.io/large-scale-health-data-monitering-with-databricks</id><content type="html" xml:base="https://sagrd.github.io/large-scale-health-data-monitering-with-databricks.md.html"><![CDATA[<p>A wearable healthcare company was generating massive high-frequency data in Azure Blob Storage, tracking metrics like heart rate, respiration, oxygen levels, and activity every minute.</p>

<p><img src="../images/healthband.png" alt="healthband er diagram" class="ioda" /></p>

<p><strong>Scope</strong>:</p>
<ul>
  <li>Instantly alert users when heart rate crossed thresholds.</li>
  <li>Send personalized notifications and celebrate user milestones.</li>
  <li>Warn users of dangerously low oxygen levels.</li>
</ul>

<p><strong>Approach</strong>:</p>
<ul>
  <li>Leveraged scheduled Databricks notebooks to process terabytes of incoming data and deliver near real-time alerts.</li>
  <li>Performed aggregated analytics (daily/weekly) and stored results in PostgreSQL for rapid access to frequently used metrics.</li>
</ul>

<p><strong>Outcome</strong>:</p>
<ul>
  <li>95% of abnormal readings detected in real-time.</li>
  <li>Significantly decreased query times for common metrics.</li>
  <li>Scalable solution capable of handling terabytes of high-frequency data.</li>
</ul>

<hr />
<p>Think my experience aligns with your needs? or got any questions that I can help with? <a href="/contact">Let’s connect.</a></p>]]></content><author><name>Sagar Dhungel</name></author><summary type="html"><![CDATA[A wearable healthcare company was generating massive high-frequency data in Azure Blob Storage, tracking metrics like heart rate, respiration, oxygen levels, and activity every minute.]]></summary></entry><entry><title type="html">Database Modernization From Mysql To Snowflake</title><link href="https://sagrd.github.io/database-modernization-from-mysql-to-snowflake.md.html" rel="alternate" type="text/html" title="Database Modernization From Mysql To Snowflake" /><published>2025-12-20T00:00:00+00:00</published><updated>2025-12-20T00:00:00+00:00</updated><id>https://sagrd.github.io/database-modernization-from-mysql-to-snowflake</id><content type="html" xml:base="https://sagrd.github.io/database-modernization-from-mysql-to-snowflake.md.html"><![CDATA[<p>A US-based lead generation agency was operating on an on-premise MS SQL Server originally built for transactional workloads, which had increasingly been used for reporting and analytics.</p>

<p><img src="../images/migration.png" alt="healthband er diagram" class="ioda" /></p>

<p><strong>Challenge:</strong></p>
<ul>
  <li>Limited scalability and high fixed infrastructure costs</li>
  <li>Complex legacy schemas with undocumented SQL logic</li>
  <li>Slow turnaround for new analytics and reporting needs</li>
</ul>

<p><strong>Approach:</strong></p>
<ul>
  <li>Assessed existing SQL schemas, workloads, and critical reports</li>
  <li>Designed a Snowflake-based data warehouse optimised for analytics</li>
  <li>Migrated historical and incremental data from SQL Server to Snowflake</li>
  <li>Refactored SQL logic into Snowflake-native transformations</li>
  <li>Enabled role-based access and BI tool integration</li>
  <li>Ran systems in parallel to validate data accuracy before cutover</li>
</ul>

<p><strong>Outcome:</strong></p>
<ul>
  <li>Elastic, on-demand compute for reporting and ad-hoc analysis</li>
  <li>Reduced infrastructure and licensing costs through a pay-per-use model</li>
  <li>Faster and more reliable analytics delivery</li>
  <li>Scalable foundation for advanced analytics and data science</li>
</ul>

<hr />
<p>Think my experience aligns with your needs? or got any questions that I can help with? <a href="/contact">Let’s connect.</a></p>]]></content><author><name>Sagar Dhungel</name></author><summary type="html"><![CDATA[A US-based lead generation agency was operating on an on-premise MS SQL Server originally built for transactional workloads, which had increasingly been used for reporting and analytics.]]></summary></entry><entry><title type="html">Unified Data Platform For Marketing Analytics</title><link href="https://sagrd.github.io/unified-data-platform-for-marketing-analytics.md.html" rel="alternate" type="text/html" title="Unified Data Platform For Marketing Analytics" /><published>2025-11-12T00:00:00+00:00</published><updated>2025-11-12T00:00:00+00:00</updated><id>https://sagrd.github.io/unified-data-platform-for-marketing-analytics</id><content type="html" xml:base="https://sagrd.github.io/unified-data-platform-for-marketing-analytics.md.html"><![CDATA[<p>An entertainment agency faced fragmented and chaotic data across Social, Search, DSPs, and other third-party platforms.</p>

<p><img src="../images/dse.png" alt="healthband er diagram" class="ioda" /></p>

<p><strong>Challenge:</strong></p>
<ul>
  <li>Each platform (Meta, Google Ads, DV360, etc.) had its own schema, access method, and update schedule.</li>
  <li>ETL processes were hard to scale, and difficult to monitor.</li>
  <li>Business units waited days for performance insights due to disconnected tools and manual workflows.</li>
  <li>Data scientists lacked a centralized workspace to share ad-hoc analyses and notebooks.</li>
  <li>Ungoverned pipelines made audits and client reporting error-prone.</li>
</ul>

<p><strong>Approach:</strong></p>
<ul>
  <li>Built a modern end-to-end data pipeline using Airflow and AWS-native tools, containerized with Docker for scalability.</li>
  <li>Created a custom JupyterHub-based Data Science environment, enabling analysts to run, share, and collaborate on models and queries using unified warehouse data.</li>
</ul>

<p><strong>Outcome:</strong></p>
<ul>
  <li>Faster, data-driven campaign optimization.</li>
  <li>Analysts freed from manual data tasks, focusing on actionable insights.</li>
  <li>Real-time visibility for marketing leaders.</li>
  <li>Improved governance, audit readiness, and compliance.</li>
</ul>

<hr />
<p>Think my experience aligns with your needs? or got any questions that I can help with? <a href="/contact">Let’s connect.</a></p>]]></content><author><name>Sagar Dhungel</name></author><summary type="html"><![CDATA[An entertainment agency faced fragmented and chaotic data across Social, Search, DSPs, and other third-party platforms.]]></summary></entry></feed>