Summary
Overview
Work History
Education
Skills
Accomplishments
Internships
Certification
Languages
Interests
Timeline
Generic
Sai Charan  Anumolu

Sai Charan Anumolu

Atlanta,,GA

Summary

  • Results-driven Data Engineer with 3+ years of experience in ETL development, database optimization, and real-time data integration. Proficient in AWS Glue, Redshift, Lambda, and Airbyte, designing scalable ETL pipelines for seamless data ingestion, transformation, and validation. Optimized Redshift performance through indexing, partitioning, and query tuning, enhancing data retrieval speed.
  • Developed a real-time email synchronization system for Twenty CRM, integrating webhooks, OAuth authentication, and historical email processing. Built custom ETL workflows to process diverse datasets across Snowflake, PostgreSQL, and Airtable. Strong expertise in SQL, stored procedures, and advanced data transformations to drive analytics and reporting.
  • Collaborated with cross-functional teams to enhance data quality, governance, and compliance, delivering efficient, high-performance data solutions. Experienced in AWS services such as Glue, Redshift, Lambda, Step Functions, CodePipeline, and CloudWatch, ensuring scalability, security, and automation in cloud-based data architectures.

Overview

21
21
years of professional experience
1
1
Certification

Work History

Software Engineer

Kunai
05.2004 - Current
  • Developed a real-time email synchronization system for Twenty CRM using webhooks, enabling instant email ingestion and processing as soon as new messages arrive.
  • Engineered a complete email sync platform with a UI-based onboarding flow, enabling users to connect Google accounts, configure sync settings, and securely authenticate via OAuth.
  • Implemented historical email synchronization, ensuring systematic retrieval and storage of past messages within Twenty CRM.
  • Designed and implemented scalable ETL pipelines for ingesting, processing, and transforming large-scale data from multiple sources, including email data extraction, Snowflake, PostgreSQL, and Airtable.
  • Optimized Airbyte configurations for incremental data sync, ensuring efficient data flow while reducing processing overhead.
  • Developed custom workflows to standardize email data and align it with structured datasets for CRM integration.
  • Designed and optimized SQL queries in Snowflake, ensuring high performance for analytical queries.
  • Developed custom connectors and API integrations to synchronize data across multiple systems, including Twenty CRM, Snowflake, PostgreSQL, and Airtable.
  • Automated data enrichment workflows to improve CRM data quality and ensure seamless synchronization.
  • Engineered advanced data workflows for data transformations, API interactions, and conditional data routing.
  • Built dynamic GraphQL queries to interact with external services, enabling efficient data retrieval.
  • Refining data models and schema design for better maintainability and performance
  • Worked closely with leadership, engineers, and business teams to align data strategies with business goals.


Data Engineer

ExponentHR
11.2023 - 02.2024
  • Developed and maintained incremental ETL processes using Change Data Capture (CDC), ensuring efficient data updates and accurate integration
  • Implemented ETL solutions using AWS Glue and Lambda for complex data transformations, meeting specific business requirements
  • Designed ETL pipelines leveraging AWS services such as S3, Redshift, and Step Functions to handle data extraction, transformation, and loading processes
  • Developed data integration workflows using Python, SQL, and AWS Lambda to enhance data flow efficiency
  • Engineered ETL pipelines with AWS CodePipeline to manage data transformations, implementing CI/CD practices for seamless deployment
  • Monitored and addressed failed AWS Glue jobs by analyzing logs in CloudWatch, implementing retry mechanisms to improve pipeline robustness
  • Led troubleshooting and debugging efforts for ETL/ELT processes, creating comprehensive unit tests to ensure data transformation reliability

Assistant System Engineer (Data)

TATA Consultancy Services
11.2020 - 12.2021
  • Conducted extensive database performance analysis on AWS Redshift, leading to optimized performance and reduced query execution time
  • Managed and executed ETL processes using AWS Glue and Redshift, ensuring seamless integration, analysis, and validation of large datasets
  • Implemented effective indexing strategies within Redshift to enhance query performance and data retrieval speed
  • Developed ETL solutions leveraging AWS services, including Glue for data integration and Lambda for handling automated data tasks
  • Addressed client queries promptly, resolving design, performance, and technical issues while ensuring clear communication of solutions
  • Developed complex data transformations using Glue jobs, implementing logic for aggregations, joins, and conditional splits
  • Utilized SQL to maintain data integrity and optimize database structures within Redshift, executing advanced queries for data analysis

Data Engineer

Uber India Research and Development Center
09.2019 - 03.2020
  • Designed, implemented, and optimized stored procedures and SQL queries to extract insights, generate comprehensive reports, and support advanced analytics for strategic decision-making
  • Collaborated with cross-functional teams to understand data requirements and developed efficient data pipelines, ensuring seamless integration with downstream reporting systems
  • Enhanced data quality and reporting accuracy through rigorous validations, data audits, and adherence to privacy regulations, securing sensitive information while maintaining compliance
  • Implemented ETL processes that enabled the transformation and integration of diverse data sources into centralized repositories, facilitating faster and more reliable reporting
  • Designed and developed interactive dashboards and reports, providing stakeholders with clear visualizations and metrics to drive data-driven decisions
  • Optimized database architecture, including table indexing, partitioning, and storage strategies, improving data retrieval speed and enabling efficient access for complex reporting needs

Education

Big Data Analytics and Information Technology -

University of Central Missouri
Lee's Summit, MO
05.2023

Skills

  • SQL (Advanced)
  • API & Webhook Integration
  • Email Parsing & Data Structuring
  • Database & CRM Integration
  • Workflow Automation & Data Processing
  • Azure (DevOps, Data Factory, SQL Server, SSIS)
  • AWS (Glue, Redshift, S3, Lambda, Step Functions, IAM, RDS, EMR, Kinesis, QuickSight, CloudWatch, CodePipeline, CodeBuild, Data Pipeline, Aurora, DynamoDB, DMS, Athena)
  • Stored Procedures
  • Functions & Triggers
  • Data Pipelines (batch & real-time)
  • Incremental Data Processing
  • Data Validation & Quality Assurance
  • Database Performance Tuning
  • Complex Joins & Queries
  • Indexing & Partitioning
  • PostgreSQL, Snowflake
  • Python (Pandas, NumPy)
  • Interactive Dashboards ( Power BI)
  • Version Control (Git, GitHub, Azure DevOps)
  • JIRA
  • Agile Methodologies

Accomplishments

    Successfully integrated email data into the Twenty CRM tool, enabling seamless ingestion and synchronization of business communications. Designed a UI-driven onboarding platform for secure Google account connections and automated sync management, improving user experience and adoption. Engineered a real-time email ingestion pipeline using webhooks, n8n, and Windmill, significantly reducing latency in syncing new emails. Additionally, implemented historical email synchronization, ensuring businesses had complete and structured access to past communications. This solution enhanced data reliability, reduced manual intervention, and streamlined CRM workflows, enabling better business insights and decision-making.

Internships

 Software Developer Intern

 SysPro IT Solutions, Duluth, USA                                                          September 2023 - November 2023


  • Worked on Enhancement requests, Defects, and User stories as a developer in an agile methodology project.
  • Resolved metadata issues within ETL processes, ensuring accurate and consistent data integration.
  • Implemented incremental ETL, addressing, and resolving bugs related to incremental processes. Conducted thorough testing of incremental refresh loads to ensure accuracy and reliability.
  • Monitored and addressed failed SQL Server jobs by analyzing logs and identifying root causes, implemented retry mechanisms within Azure DevOps pipelines to automatically rerun failed SQL Server jobs, improving overall pipeline robustness.
  • Engineered ETL pipelines, including Azure DevOps pipelines, to execute and manage data transformations. Implemented Continuous Integration/Continuous Deployment (CI/CD) practices to ensure the seamless and efficient integration of information.
  • Configured Azure DevOps release pipelines to deploy SQL Server database changes across different environments, maintaining consistency and reliability.
  • Implemented Snowflake as a data warehousing solution, designing and implementing ETL processes to extract, transform, and load data seamlessly into Snowflake, ensuring high performance and reliability.
  • Integrated Python and Scala scripts into Azure Data Factory (ADF) pipelines, enabling seamless orchestration and execution of complex data workflows across hybrid environments.


 Software Developer Intern

 Merit American Technologies Inc, Irving, USA                                              February 2024 - May 2024


  • Implemented SSIS components such as Lookup, Merge Join, Conditional Split, and Derived Column to enhance data processing and quality
  • Conducted database performance analysis on SQL Server, implementing effective indexing strategies to improve query execution times
  • Managed ETL processes using traditional tools like SSIS alongside modern tools, ensuring effective integration, analysis, and validation of data
  • Utilized Variables and Parameters within SSIS packages to enable dynamic data transformations and process customization
  • Created Stored Procedures, Functions, and Packages to streamline ETL workflows and improve data processing efficiency
  • Designed and generated SSRS reports for key performance indicators (KPIs), facilitating data-driven decision-making
  • Integrated Power BI with existing SSIS packages to provide dynamic, interactive data visualizations that support business intelligence needs.


Certification

  • AWS Certified Data Engineer – Associate - Amazon Web Services (AWS).
  • Microsoft Certified: Azure Data Engineer Associate

Languages

English
Full Professional
Telugu
Native or Bilingual
Hindi
Professional Working

Interests

  • Tech enthusiast, passionate about exploring the latest advancements and innovations
  • Road Trips
  • Passionate about balancing physical health with mental and emotional wellness
  • Gym Workouts


Timeline

Data Engineer

ExponentHR
11.2023 - 02.2024

Assistant System Engineer (Data)

TATA Consultancy Services
11.2020 - 12.2021

Data Engineer

Uber India Research and Development Center
09.2019 - 03.2020

Software Engineer

Kunai
05.2004 - Current

Big Data Analytics and Information Technology -

University of Central Missouri
Sai Charan Anumolu