Summary
Overview
Work History
Education
Skills
Timeline
Generic

Sri Harsha Vardhan Yendru

Charlotte

Summary

Data Engineer with 3+ years of experience in designing and optimizing scalable ETL/ELT pipelines, big data processing, and cloud-based data solutions. Proficient in Python, SQL, PySpark, AWS, Snowflake, DBT, and Azure, with strong expertise in building metadata-driven workflows and automating data pipelines. Currently working with Azure Data Factory (ADF) and Databricks to orchestrate and transform large datasets — ingesting CSV/Excel files from SFTP, performing complex transformations in Databricks, and exporting cleansed datasets back to SFTP as well as Delta tables. Skilled in integrating Snowflake for modern data warehousing and leveraging Streamlit for building lightweight analytics applications. Proven ability to enhance data quality, improve pipeline performance, and support analytics/BI teams through robust data modeling and modern warehousing practices.

Overview

4
4
years of professional experience

Work History

Data Warehouse Developer

Infoville
Charlotte
08.2024 - Current
  • Designed and orchestrated ETL workflows in Azure Data Factory (ADF) to automate the ingestion of CSV and Excel files from SFTP locations.
  • Implemented Databricks notebooks (PySpark/Python) for data cleaning, business rule transformations, and preparation of curated datasets.
  • Generated processed CSV outputs, and automated delivery back to SFTP for downstream systems.
  • Built and maintained Delta Lake tables in Databricks to enable incremental data loads, history preservation, and optimized query performance.
  • Integrated Delta Lake datasets with Snowflake for advanced analytics and reporting requirements.
  • Developed interactive Streamlit applications for data visualization and operational monitoring.
  • Ensured pipeline reliability and data quality through parameterization, validation checks, and error-handling logic in ADF and Databricks.

Product Engineer

Maybank
Bangalore
02.2023 - 07.2023
  • ECONOMIC CAPITAL: Worked on an R shiny-based application to calculate economic capital and expected loss of Maybank borrowers' portfolios using a Monte Carlo simulation model.
    Made various visualizations, like loss distribution, histogram, box plot, and scatter plot, to identify and share real-time trends, outliers, and new insights from the data generated through the Economic Capital Model.

Data Engineer Intern

Digipix Technologies
Bangalore
08.2021 - 12.2022
  • Built scalable pipelines for the Customer Engagement Project using AWS Glue to ingest data from FTP, APIs, and SharePoint into S3; optimized Spark-based ETL workflows, and improved data processing performance.
  • Migrated Pandas scripts to PySpark, improving performance by 80%.
  • Collaborated with stakeholders to translate business requirements into 3+ effective ETL pipelines.
  • Created stored procedures, views, and triggers to support SCD4 logic and reporting requirements

Education

Master of Science - Computer Science

University of New Haven
West Haven, CT
05-2025

Bachelor of Science - Computer Science And Engineering

Presidency University
Bangalore,India
06-2022

Skills

  • Programming: Python, SQL, R shinny and PySpark
  • Cloud: Azure(ADF,Blob Storage), AWS
  • Data Engineering: ETL, Data Pipelines, DBT, Data Modeling, Star/Snowflake Schema, Delta Lake
  • Big data: Hadoop, Hive, Spark, Kafka
  • Databases and warehouses: Snowflake, PostgreSQL, MySQL, Oracle SQL
  • Tools: Git, Azure DevOps, AWS CodeCommit, Jupyter Notebook, R studio, Databricks Notebook

Timeline

Data Warehouse Developer

Infoville
08.2024 - Current

Product Engineer

Maybank
02.2023 - 07.2023

Data Engineer Intern

Digipix Technologies
08.2021 - 12.2022

Master of Science - Computer Science

University of New Haven

Bachelor of Science - Computer Science And Engineering

Presidency University
Sri Harsha Vardhan Yendru