$ whoami
$ cat role.txt
$ ls skills/
$ echo $STATUS
Processing 10M+ records daily | 40% pipeline performance gains | Enterprise data architecture | ML system design
> MS Data Science @ Regis University (GPA: 3.9) - Completed: Dec 2025
> BS Computer Science - Completed 2021
> 6+ Years Python, SQL & Data Architecture Experience
> Location: Denver, CO
Data Architect with proven expertise designing enterprise-scale data infrastructure, distributed ETL pipelines processing 10M+ records daily, and production ML systems. Specializing in Apache Spark, AWS services (S3, Lambda, EC2, CloudFormation, Glue, Kinesis, Redshift), and modern data architectures with high accuracy and uptime standards.
Experienced in pipeline performance optimization, cloud infrastructure automation, query tuning, and proactive monitoring system design. Published researcher in predictive ML models for Scope 3 emissions data accuracy.
classGraduateResearchAssistant:def__init__(self):self.role="Graduate Research Assistant"self.company="Regis University"self.duration="Sep 2024 - Apr 2025"self.location="Denver, CO"defachievements(self):return["Designed end-to-end ETL pipeline for Scope 3 emissions (10M+ records)","Reduced processing time by 40% using Spark distributed computation","Developed optimized PySpark jobs for Snowflake data loading","Integrated scalable KNN imputation improving data quality by 15%","Co-authored published research on ML for emissions data accuracy"]deftech_stack(self):return["Apache Spark","Airflow","PySpark","Snowflake","Python"]
classDataEngineer:def__init__(self):self.role="Data Engineer"self.company="Appharu"self.duration="Jan 2020 - Jul 2023"self.location="Remote"defimpact(self):return{"focus":"ETL pipeline architecture","strengths":"Query optimization & monitoring","databases":"PostgreSQL, MySQL, MongoDB","scope":"Production data systems","automation":"Data quality & validation frameworks"}
Enterprise-grade ETL pipeline design for processing large-scale healthcare records using distributed computation. Implemented parallel processing with Apache Spark achieving 40% performance improvement.
Infrastructure as Code solution using AWS CloudFormation, automating provisioning of S3, Lambda, EC2, and RDS. Designed for rapid, repeatable environment deployment.
Real-time analytics dashboard integrated with Apache Airflow for automated reporting and Snowflake for analytics queries. Designed for fast query response and automated data refresh.
Open-source AI agent testing framework published on PyPI (pip install agentproof). Architected test harness, assertion engine, and reporting pipeline for production AI agent validation.
@inproceedings{dahal2024emissions,
title = {"Predictive Models for Scope 3 Emissions: Improving Accuracy with Machine Learning and Financial Data"},
author = {Dahal, S. and Pochampally, A. and Soraf, K.},
conference = {Marketing and Data Sciences},
institution = {Regis University},
year = {2024}
}
$ ping sarak-dahal --resolve
Connection established. Ready to receive transmission...
$ cat contact_info.txt
$ cat social_links.txt
$ compose_message --interactive