Summary
Overview
Work History
Education
Skills
Certification
Projects
Languages
Timeline
Generic

AKANKSHA BOMMIDEKA

DENTON,TX

Summary

Experienced Data Engineer with 4 years of proven expertise in designing, developing, and optimizing data pipelines across on-premises and cloud platforms. Proficient in Python for data processing and transformation, adept at building scalable pipelines with Scala in GCP environments. Skilled in translating business requirements into technical solutions and implementing robust data quality controls. Proficiency in SQL across multiple databases and creating interactive reports using Excel, Tableau, and other visualization tools. Demonstrated experience in cloud migration, real-time analytics with Scala, Spark, AWS Glue, and Hadoop stack. Contributed to Business Intelligence solutions for tracking KPIs and enhancing decision-making. Experienced with JSON structures and SQL Server technologies including SSIS, SSRS, and SSAS. Familiar with MongoDB and adaptable to emerging technologies.

Overview

6
6
years of professional experience
1
1
Certification

Work History

Data Engineer

DISH Network
TX
03.2023 - Current
  • Developed advanced ETL solutions to facilitate seamless data flow between systems, resulting in improved real-time data synchronization and exceptional communication performance
  • Developed Python code for enhancing operational and analytics data systems through solutions for data ingestion, curation, aggregation, integration, cleaning, transformation, and control
  • Created elaborate data sets to meet functional reporting demands, resulting in enhanced decision-making capabilities.
  • Implemented data pipeline solutions for efficient handling of massive datasets on Google Cloud Platform (GCP).
  • Streamlined ETL processes and automated data ingestion from diverse sources using Scala libraries and frameworks.
  • Collaborated on various code reviews, performance tuning, and troubleshooting exercises to uphold optimal standards for both Scala codebase and data pipelines.
  • Developed and implemented efficient processes to support data transformation, structure, and metadata while establishing data quality controls.
  • Applied TDD processes and tooling to enhance the effectiveness of established test suites
  • Documented and communicated deployment, maintenance, support, and business functionality information.
  • Optimized data acquisition by creating custom Python tools to automate text and image data crawling from various sources.
  • Worked closely with stakeholders to translate functional requirements into technical approach, design, and decisions
  • Streamlined data integration by leveraging GCP, maintaining adherence to high standards of data quality and accessibility
  • Implemented conditional and level metrics in SQL schema utilizing Python for improved reporting capabilities
  • Created Tableau dashboards to visualize data and generate insights.
  • Applied a range of technical skills including SQL, Python, Hadoop/Hive for data analysis and ETL solution design in order to enhance business processes.
  • Suggested optimal GCP tools and solutions for diverse project requirements and use cases.
  • Performed query optimizations for significant improvements in MicroStrategy, Tableau, RDBMS, Hadoop backend systems focusing on large datasets and aggregate structures.

Graduate Teaching Assistant

University Of North Texas
TX
01.2022 - 12.2022
  • Demonstrated expertise in utilizing Agile Methodology while overseeing projects for 35 students as a Teaching Assistant
  • Took charge of crafting, refining, and regularly maintaining content on the University graduate website to ensure efficient communication and widespread information exchange.
  • Implemented automated dissertation approval flow, reducing 10 hours of manual effort weekly.
  • Showcased efficiency, automation skills, and dedication to process improvement.

Data Engineer (Intern)

Otsuka pharmaceuticals
05.2022 - 08.2022
  • Gathered requirements, analyzed the system, and estimated development and testing efforts
  • Conducted analysis of enterprise data report integration and created functional specifications for enhancing Enterprise Reporting Systems
  • Migrated existing application to AWS, employing EC2 and S3 services for processing and storing small datasets. Managed maintenance of Hadoop cluster through AWS EMR.
  • Executed transformations with Spark, preserving the result in HDFS before loading it into the target database Snowflake.
  • Designed and developed ETL processes in AWS Glue, migrating data from external sources such as S3 and file formats including JSON, Parquet, and Text Files to AWS Redshift.
  • Processed real-time data analytics utilizing Spark Streaming, Kafka, and Flume
  • Configured Spark streaming to retrieve real-time data from Kafka and save it to HDFS.

Hadoop Developer

Info Sage
Hyderabad, India
11.2020 - 07.2021
  • Participated in Joint Application Development sessions with the business to analyze requirements and impact
  • Developed various ETL scripts (using database objects) and SSIS packages (using tasks) for data extraction, transformation, and loading between servers and client databases
  • Achieved improvements in the performance of Stored Procedures and long-running queries by implementing indexing strategies and query optimization techniques.
  • Implemented dynamic SQL for optimized performance and enhanced efficiency
  • Optimized and tuned Oracle PL/SQL procedures and SQL queries for improved performance.
  • Implemented SQL Loader to facilitate efficient data transfer.
  • Developed and deployed SSIS packages into various environments, utilizing package configuration to export package properties.
  • Created Tableau workbooks to analyze various types of data including year-over-year, quarter-over-quarter, YTD, QTD, and MTD
  • Assisted in the design, development, and implementation of a BI solution that tracked Sales, Product,and Customer KPIs.
  • Utilized Excel sheets and SQL Server to produce sophisticated Tableau dashboards for data analysis.
  • Designed and implemented application components using a test-driven development approach in an Agile environment
  • Utilized Excel pivot tables and complex formulas to effectively manipulate large data structures
  • Created distributed reports using SQL Server 2008 R2 Reporting Services (SSRS), producing multiple formats such as Excel, PDF, and CSV files.

Hadoop Developer (Intern)

ACS Solutions Pvt Ltd.
Hyderabad, India
06.2019 - 10.2020
  • Designed and executed a MapReduce program for efficient analysis of large datasets, showcasing problem-solving and coding expertise.
  • Streamlined data retrieval and analysis by designing and optimizing Hive queries
  • Organized and maintained Hadoop clusters for effective development and testing.
  • Utilized Apache Spark to collaboratively create and implement a data pipeline for real-time data processing.

Network Engineer (Intern)

ECIL
Hyderabad, India
05.2018 - 08.2018
  • Developed digital communication system for coursework project, showcasing expertise in modulation techniques and error control coding.
  • Executed a microcontroller-based project showcasing the practical uses of embedded systems.
  • Conducted electronic circuit and device experiments while analyzing and documenting results
  • Applied theoretical knowledge in previous internships or projects to real-life scenarios
  • Participated in a collaborative effort to identify and resolve electronic hardware complications
  • Contributed to the testing and calibration of electronic equipment alongside senior engineers.

Education

Master of Science - Computer and Information Sciences

University of North Texas
Denton, TX
12-2022

Bachelor of Science - Computer Science

AVN Institute Of Engineering And Technology
Hyderabad,India
10-2020

Skills

  • Python
  • SQL
  • Java
  • Hive
  • Spark
  • Scala
  • Shell Scripting
  • Oracle PL/SQL
  • SQL Server (2008 R2)
  • Snowflake (Data Warehouse)
  • Amazon Web Services (AWS)
  • GCP
  • EC2 (Elastic Compute Cloud)
  • S3
  • Redshift
  • Microsoft Azure (experience mentioned in relation to cloud)
  • API Development
  • Data Modeling
  • Machine Learning
  • Big data technologies
  • Relational databases
  • Data Analysis
  • Analytical Thinking
  • Team Collaboration
  • Continuous Improvement
  • Effective Communication
  • Time Management
  • Adaptability and Flexibility
  • Written Communication
  • Teamwork and Collaboration

Certification

  • AWS Certified Cloud Practitioner
  • Salesforce Certified Administrator

Projects

Internet-Based Sale and Purchase System: Created a project aimed at streamlining the operations of local general and mobile stores by providing an online platform for selling goods and enhancing accessibility for the public.

 iCare: Digital Health Records Platform, iCare serves as a platform dedicated to digitizing patients' health records, offering doctors the valuable ability to analyses patients' previous health data for more informed medical decisions and treatments. 

Currency Recognition System: Executed a currency recognition system as a vital component of academic projects, employing Java, Python, and Spring technologies.

Languages

English
Full Professional

Timeline

Data Engineer

DISH Network
03.2023 - Current

Data Engineer (Intern)

Otsuka pharmaceuticals
05.2022 - 08.2022

Graduate Teaching Assistant

University Of North Texas
01.2022 - 12.2022

Hadoop Developer

Info Sage
11.2020 - 07.2021

Hadoop Developer (Intern)

ACS Solutions Pvt Ltd.
06.2019 - 10.2020

Network Engineer (Intern)

ECIL
05.2018 - 08.2018

Master of Science - Computer and Information Sciences

University of North Texas

Bachelor of Science - Computer Science

AVN Institute Of Engineering And Technology
AKANKSHA BOMMIDEKA