Timeline
Work History
Overview
Education
Skills
Work Preference
Summary
Websites
Languages
Software
Hi, I’m

VARSHITHA VEERAMACHANENI

Data Engineer
Frisco,TX
"Data is the new oil. It’s valuable, but if unrefined, it cannot really be used. It has to be changed into gas, plastic, chemicals, etc., to create a valuable entity that drives profitable activity."
Clive Humby
VARSHITHA VEERAMACHANENI

Timeline

Data Engineer

KHP Solutions
08.2024 - Current

Southern Arkansas University

Master of Science from Computer Science
08.2023 - 12.2024

Data Engineer

Centurion Software Solutions
07.2022 - 06.2023

Data Engineer

Fortune Technologies
10.2021 - 06.2022

Jawaharlal Institute of Technology University

Bachelor of Science from Computer Science
06.2018 - 07.2022

Work History

KHP Solutions

Data Engineer
08.2024 - Current

Job overview

  • Lead the development of scalable data pipelines utilizing Python and PySpark, focusing on streamlining the data flow from multiple sources into a centralized data warehouse.
  • Expert in crafting and executing advanced SQL queries for data manipulation and analysis.
  • Innovate in the automation of data flows, enhancing the efficiency and reliability of data processing.
  • Actively collaborate with cross-functional teams, including analysts, developers, and business stakeholders, to define and deliver on data requirements.
  • Spearhead initiatives to optimize data storage and retrieval processes, especially for handling large and complex datasets.
  • Independently execute a comprehensive web scraping project to collect and analyze market data from multiple digital media and entertainment platforms, enhancing competitive analysis and business strategy development.
  • Collaborate with cross-functional teams to integrate scraped data into the company’s central data warehouse, significantly improving the accessibility and usability of the data for various stakeholders.
  • Adhere strictly to legal and ethical standards in data collection and processing, ensuring compliance with relevant data privacy regulations and best practices.

Centurion Software Solutions

Data Engineer
07.2022 - 06.2023

Job overview

  • Worked closely with stakeholders across Financial Services departments to design, build and deploy various initiatives within the data platform.
  • Served as a SME for Finance Customer communication and transactional data.
  • Developed, deployed and maintained data services using Snowflake and Python.
  • Prepared and presented reports, analysis and presentations to various stakeholders including executives in Tableau.
  • Designed best practices to support continuous process automation for data ingestion and data pipeline workflows.
  • A solid experience and understanding of architecting, designing and operationalization of large-scale data and analytics solutions on Snowflake Cloud Data Warehouse.
  • Loaded various data pipelines into snowflake segregating into various stages and built various curated tables to feed the Tableau Dashboards to show various metrics to Business.
  • Extracted TBs of data from SQL Server and Teradata to the One Lake.
  • A solid experience and understanding of architecting, designing and operationalization of large-scale data and analytics solutions on Snowflake Cloud Data Warehouse.
  • Built event-driven and scheduled data pipelines to ingest data into operational and analytical DB from various sources

Fortune Technologies

Data Engineer
10.2021 - 06.2022

Job overview

  • Responsible for expanding, optimizing, and monitoring data collection, data flows and datasets.
  • Build and maintain data pipelines and datasets for data science and analytics experts to improve functionality in data systems.
  • Drive initiatives to analyze and acquire free and paid data from third parties.
  • Collaborate with data science and data services teams to develop best practices. Includes data ingestion, dataset creation, storage and updates, naming conventions and retention.
  • Work with stakeholders on data-related technical issues and data infrastructure needs.
  • Implement methods to improve data reliability and quality.
  • Assemble large and complex data sets that meet functional/non-functional business requirements. ldentify, design, and implement internal process improvements.
  • Includes automating manual processes, optimizing data delivery, and re-designing infrastructure for greater scalability.

Overview

3
years of professional experience

Education

Southern Arkansas University
Magnolia, AR

Master of Science from Computer Science
08.2023 - 12.2024

Jawaharlal Institute of Technology University
Kakinada, India

Bachelor of Science from Computer Science
06.2018 - 07.2022

Skills

  • Python
  • SQL
  • PySpark
  • DAX (Data Analysis Expressions)
  • Cobol (for legacy system migration)
  • Apache Spark
  • Hadoop (HDFS, Hive)
  • AWS Glue
  • Snowflake
  • DBT (Data Build Tool)
  • AWS (EC2, S3, RDS, Lambda, Step Functions, Glue, SNS, DMS, IAM)
  • Azure
  • Google Cloud Platform (Google Analytics)
  • MySQL
  • PostgreSQL
  • DB2
  • SQL Server
  • Oracle SQL
  • Teradata
  • NoSQL Databases (Document databases, in-memory caching)
  • Cloud Databases (AWS RDS, Snowflake)
  • Data Warehousing (Snowflake, Teradata)
  • Data Modeling (ERWIN, conceptual data flows)
  • Change Data Capture (CDC)
  • Data Lakes (AWS S3, Hadoop)
  • AWS Glue
  • DBT (Data Build Tool)
  • SSIS (SQL Server Integration Services)
  • Custom ETL workflows using Python and PySpark
  • Power BI (DAX, drill-through, hierarchies, role-level security)
  • Tableau (filters, parameters, calculated fields)
  • SSRS (SQL Server Reporting Services)
  • Qlik (migration to Power BI)
  • Excel (pivot tables, visualizations)
  • Git (version control)
  • CI/CD pipelines (automated software delivery)
  • AWS Step Functions
  • Workflow scheduling tools
  • Query optimization (SQL, Snowflake, SQL Server)
  • Performance tuning (ETL processes, SQL queries)
  • Process optimization (data delivery, scalability, reliability)
  • IAM (Identity and Access Management)
  • Vulnerability analysis
  • Compliance with industry standards (eg, retail, finance)
  • Mainframe to cloud migration (DB2, Cobol files)
  • Teradata to Snowflake migration
  • SAS to Power BI dashboard transition
  • Python scripts for data validation, cleaning, and transformation
  • Automated reporting (Power BI, SQL scripts)
  • AWS Lambda for serverless automation
  • Demand forecasting (Python)
  • Clickstream data analysis
  • SMART project analysis (revenue growth, customer segmentation)
  • API data extraction (10 APIs)
  • Flat file creation and loading
  • Multi-platform data architecture (Teradata, Big Data, SQL Server, Oracle SQL)
  • Leadership & Team Management
  • Communication
  • Problem-Solving
  • Project Management
  • Collaboration
  • Strategic Thinking
  • Retail Industry expertise
  • Finance Industry expertise
  • Supply Chain & Logistics expertise
  • Marketing & Sales expertise
  • Education Industry expertise
  • Power BI, Tableau, SSRS, Qlik, Excel
  • AWS (EC2, S3, RDS, Lambda, Glue, SNS, DMS, IAM)
  • Azure
  • Google Analytics
  • MySQL, PostgreSQL, DB2, SQL Server, Oracle SQL, Teradata, Snowflake
  • Hadoop, Hive, Spark, AWS Glue, DBT
  • Git, CI/CD pipelines
  • AWS Step Functions, workflow scheduling tools
  • Query optimization tools, performance tuning tools
  • Cobol, DB2, mainframe systems
  • Data Modeling (conceptual, logical, physical)
  • ETL/ELT Best Practices
  • Agile & Collaborative Development
  • Data Security & Compliance
  • Process Optimization
  • Automation

Work Preference

Work Type

Full TimeContract WorkInternship

Work Location

On-SiteRemoteHybrid

Summary

Data engineering professional with deep expertise in data architecture, pipeline development, and big data technologies. Proven track record in optimizing data workflows, enhancing system efficiency, and driving business intelligence initiatives. Strong collaborator, adaptable to evolving project demands, with focus on delivering impactful results through teamwork and innovation. Skilled in SQL, Python, Spark, and cloud platforms, with strategic approach to data management and problem-solving.

Languages

English

Software

Programming & Data Processing Python SQL PySpark DAX (Data Analysis Expressions) Cobol Big Data & Cloud Platforms Apache Spark Hadoop (HDFS, Hive) AWS Glue Snowflake DBT (Data Build Tool) AWS (EC2, S3, RDS, Lambda, Step Functions, Glue, SNS, DMS, IAM) Azure Google Cloud Platform (Google Analytics) Databases MySQL PostgreSQL DB2 SQL Server Oracle SQL Teradata NoSQL Databases (eg, MongoDB, Cassandra) AWS RDS Snowflake Data Warehousing & Modeling ERWIN (Data Modeling) Snowflake Teradata AWS S3 (Data Lakes) Hadoop (Data Lakes) ETL/ELT Tools AWS Glue DBT (Data Build Tool) SSIS (SQL Server Integration Services) Apache NiFi Talend Data Visualization & Reporting Power BI Tableau SSRS (SQL Server Reporting Services) Qlik Excel Version Control & CI/CD Git GitHub/GitLab/Bitbucket Jenkins (CI/CD pipelines) Azure DevOps Workflow Orchestration AWS Step Functions Apache Airflow Luigi Monitoring & Optimization SQL Query Optimization Tools (eg, SQL Profiler) Performance Monitoring Tools (eg, Datadog, New Relic) AWS CloudWatch Data Security & Compliance AWS IAM (Identity and Access Management) Vault by HashiCorp Compliance Tools (eg, AWS Config, Azure Security Center) Legacy Systems Mainframe Systems (eg, IBM z/OS) Cobol Compilers (eg, Micro Focus) Automation Python Scripts AWS Lambda PowerShell Ansible Advanced Analytics Google Analytics Clickstream Analysis Tools (eg, Adobe Analytics) Demand Forecasting Tools (eg, Prophet, ARIMA in Python) Data Integration API Management Tools (eg, Postman, Swagger) Flat File Processing Tools (eg, Alteryx, Informatica) Collaboration & Project Management Jira Confluence Trello Microsoft Teams Slack Other Tools Visual Studio (for ETL development) ERWIN (for Data Modeling) SAS (for legacy reporting) NetSuite (for KPI tracking)

VARSHITHA VEERAMACHANENIData Engineer