π» About Me: Senior Data Engineer
Hello! π
I’m Santosh Kumar Gouda, a Senior Data Engineer with 8.5+ years of experience architecting cloud-native data solutions. I transform complex data challenges into scalable, cost-efficient systemsβproven by 40%+ performance gains across multiple enterprises.
π οΈ Technical Arsenal
π§° Core Stack
Category | Technologies |
---|---|
Cloud Platforms | AWS (Glue, Redshift, S3, EMR, Lambda, RDS, ), Azure (Data Factory, Databricks, Synapse) |
Big Data | PySpark, Airflow, Kafka, Snowflake, Delta Lake |
Databases | PostgreSQL, MySQL, MongoDB, Azure SQL, Redshift |
DataOps | dbt, Great Expectations,, Kubernetes, Docker, CI/CD |
πΌ Professional Journey
π Data Analytic Tech Lead
π Sears Global Technologies | Nov 2022 - Present | Pune
- Engineered Snowflake ingestion pipelines from OLTP systems (MySQL/PostgreSQL/MongoDB) and S3/SFTP, achieving 30% faster data delivery
- Automated 200+ ETL workflows using Apache Airflow, reducing manual intervention by 40% with 99% on-time delivery
- Designed complex SQL queries for ad-hoc analytics, maintaining 98% stakeholder satisfaction
- Implemented data quality frameworks that reduced pipeline failures by 35%
π§ Data Engineer
π Enkash | May 2022 - Sep 2022 | Bangalore
- Built AWS Glue/PySpark pipelines to migrate on-prem MySQL β Redshift, cutting processing time by 40%
- Developed advanced Glue scripts optimizing transformation logic, reducing job runtime by 25%
- Created interactive QuickSight dashboards serving 200+ business users with real-time sales insights
π§ Data Engineer
π Tech Mahindra | May 2021 - May 2022 | Bangalore
- Consolidated 1TB+ disparate data sources using AWS Glue/S3, enabling unified analytics
- Automated data reconciliation with Python, achieving 90% accuracy improvement and saving 20+ hours/week
- Enhanced ETL performance reducing data wrangling time by 40%
π§ Data Engineer
π Alten India | Sep 2020 - Apr 2021 | Bangalore
- Migrated legacy SSIS packages β Azure Databricks, slashing ETL runtime by 35%
- Developed Azure Data Factory pipelines integrating Blob Storage β Data Warehouse, boosting efficiency by 50%
π§ Data Analytics Engineer
π Bristlecone | Nov 2019 - Sep 2020 | Bangalore
- Architected data management framework reducing retrieval time by 25%
- Processed 500K+ records/day into Azure Data Lakes, improving accessibility by 30%
π§ Data Analyst
π Accenture Solutions | Jan 2016 - Sep 2018 | Bangalore
- Automated data extraction with Python, cutting acquisition time by 35%
- Resolved 50+ critical script errors, ensuring system stability
ποΈ Signature Projects
π Real-Time Payment Analytics
Tech: Kafka β’ PySpark β’ Redshift β’ Airflow
- Processed 2M+ transactions/day with <100ms latency
- Reduced fraud detection time from hours β seconds
**π Cloud Data Migration **
Tech: Python β’ PySpark β’ SQL β’ Airflow β’ AWS Glue β’ Snowflake
- Cut ETL costs by 40% via auto-scaling and partitioning
- Documented as internal gold standard
π Blog Posts & Writing
- Handling Large Data Volumes (100GB β 1TB) in PySpark: Best Practices & Optimizations
(Practical handling of big data and optimization strategies for data warehouses) - Data Engineering Best Practices for Production Data Pipelines
(Best Practices of Data Pipelines)
π Knowledge Leadership
- π Certifications:
- Microsoft Certified Azure Data Engineer Associate (2025)
- AWS Certified Data Analytics Specialty (2025)
- π Education: B.Tech in CSE @ NIT Nagaland (7.74 CGPA)
π Resume
π Download Resume | π View Online
π Beyond Code
- π§ Mentor for early-career data professionals
- π Contributor to open-source data tools
π« Let’s Build Something Great
Platform | Link |
---|---|
s.k.gouda21@gmail.com | |
[linkedin.com/in/sgouda0412] | |
GitHub | [github.com/sgouda0412] |
Phone | +91-7619264934 |
Open to:
- Architectural consultations
- High-impact data engineering projects
- Speaking/mentorship opportunities
“Data systems should scale like bambooβstrong, flexible, and resilient.” π