Summary
Overview
Work History
Education
Skills
Languages
Timeline
Generic

Vijetha Baddam

Baltimore,MD

Summary

Dynamic Data Engineer with a proven track record at U.S. Bank, adept in Azure cloud components and big data processing. Excelled in developing efficient data ingestion pipelines and optimizing data flow, showcasing strong analytical skills and commitment to drive project success. Demonstrated leadership and problem-solving abilities, achieving a 99% reduction in outstanding receivables through strategic negotiations.

Overview

6
6
years of professional experience

Work History

Data Engineer

Ct Tech Services
02.2025 - Current
  • Developed data pipelines to streamline ETL processes, enhancing data availability for analytics.
  • Collaborated with cross-functional teams to define data architecture and improve system performance.
  • Implemented data quality checks and validation processes, ensuring accuracy in reporting and analysis.
  • Led initiatives to optimize database management, reducing query response times significantly.
  • Designed and maintained scalable data models aligned with business objectives and reporting needs.
  • Analyzed complex datasets to identify trends, driving strategic decisions for organizational growth.
  • Enhanced data quality by performing thorough cleaning, validation, and transformation tasks.

Rubik Systems
10.2023 - 11.2024
  • Involved in complete Big Data flow of the application starting from data ingestion from upstream to HDFS, processing and analyzing the data in HDFS
  • Working knowledge on Azure cloud components (Databricks, Data Lake, Blob Storage, Data Factory, Storage Explorer, SQL DB, SQL DWH, Cosmos DB)
  • Extract Transform and Load data from Sources Systems to Azure Data Storage services using a combination of Azure Data Factory, data bricks, PySpark, Spark SQL and U-SQL Azure Data Lake Analytics
  • Created Pipelines in ADF using Linked Services/Datasets/Pipeline/ to Extract, Transform, and load data from Azure SQL, Blob storage, and Azure SQL Data warehouse
  • Worked with Azure BLOB and Data Lake storage and loading data into Azure SQL Synapse analytics (DW)
  • Involved in developing data ingestion pipelines on Azure HDInsight Spark cluster using Azure Data Factory and Spark SQL
  • Build streaming ETL pipelines using Spark Streaming to extract data from various sources, transform it in real-time, and load it into a data warehouse such as Azure Synapse Analytics
  • Use tools such as Azure Databricks or HDInsight to scale out the Spark Streaming cluster as needed
  • Developed Spark API to import data into HDFS from Teradata and created Hive tables
  • Monitored Spark cluster using Log Analytics and Ambari Web UI
  • Developed Spark core and Spark SQL scripts using Scala for faster data processing
  • Worked on the Hadoop ecosystem in PySpark on HDInsight and Databricks
  • Extensively used Spark core – Spark Context, Spark SQL, and Spark Streaming for real time data
  • Performed data profiling and transformation on the raw data using Python.

Rolang technologies
06.2023 - 10.2024
  • Experience in designing and developing POCs in Spark using Scala to compare the performance of Spark with MapReduce, Hive
  • Hands - on experience in Azure Cloud Services, Azure Synapse Analytics, SQL Azure, Data Factory, Azure Analysis services, Application Insights, Azure Monitoring, Key Vault, and Azure Data Lake.
  • Created Batch & Streaming Pipelines in Azure Data Factory (ADF) using Linked Services/Datasets/Pipeline/ to Extract, Transform and load data.
  • Created Azure Data Factory (ADF) Batch pipelines to Ingest data from relational sources into Azure Data Lake Storage (ADLS gen2) & incremental fashion and then load into Delta tables after cleansing
  • Created Azure logic apps to trigger when a new email received with an attachment and load the file to blog storage
  • Implemented CI/CD pipelines using Azure DevOps in cloud with GIT, Maven, along with Jenkins plugins.
  • Build a Spark Streaming application to perform real-time analytics on streaming data.
  • Use Spark SQL to query and aggregate data in real-time, and output the results to various visualizations such as Power BI or Azure Data Studio.
  • Develop a Spark Streaming application that integrates with event-driven architectures such as Azure Functions or Azure Logic Apps.
  • Use Spark Streaming to process events in real-time, and trigger downstream workflows based on the results.
  • Involved in creating Hive Tables and loading and analysing data using hive queries
  • Designed and developed custom Hive UDF
  • Using the JSON and XMLSerDe's for serialization and deserialization to load JSON and XML data into HIVE tables

University of Findlay
08.2022 - 05.2023
  • Assisted patients with general inquiries and provided guidance to help them navigate the university's resources
  • Contacting Patients and Pharmacist regarding Copay Cards and resolve their problems
  • Demonstrated leadership by making improvements to work processes and helping to train others
  • Maintained updated knowledge through continuing education and advanced training
  • Promoted high customer satisfaction by resolving problems with knowledgeable and friendly service
  • Maintained HIPAA rules and regulations and accomplished task on given time
  • Helped Patients to answer their questions regarding medications, uses and side affects of the desired drug

Account Receivable Caller

JSD Medical and Business Management
01.2020 - 07.2021
  • Proven track record of effectively reducing outstanding receivables through strategic follow-up, negotiation, and issue resolution
  • Adept at maintaining accurate records, providing excellent customer service, and ensuring compliance with company policies
  • Strong communication and negotiation skills with a keen attention to detail
  • Managed end-to-end accounts receivable process, including invoice generation, distribution, and follow-up calls
  • Conducted negotiations with customers to establish payment plans and terms, resulting in a 99% reduction in outstanding receivables
  • Resolved payment issues and disputes promptly, ensuring customer satisfaction and timely resolution
  • Proficient in Microsoft Excel, adept at data analysis, and skilled in creating complex spreadsheets
  • Experience includes utilizing advanced functions, pivot tables, and charts to interpret and present data effectively

Medical Coder

Valliappa Medical Management
06.2019 - 12.2019
  • Ensured strict adherence to healthcare regulations, coding guidelines, and proficient use of medical coding software
  • Conducts coding audits, maintains high-quality standards, and implements corrective actions for accuracy enhancement
  • Collaborated effectively with healthcare professionals, resolving coding queries and discrepancies through clear communication
  • Improved coding productivity, streamlines processes, and contributed to revenue cycle efficiency
  • Provides training and mentorship as needed
  • Recognized for maintaining a high level of accuracy and attention to detail, minimizing errors in coding tasks

Education

Master of Science (M.S.) - Health Informatics

University of Findlay
Findlay, Ohio
05.2023

Bachelor of Science (B.S.) - Biotechnology

Aurora Degree and PG College
06.2019

Skills

  • ETL development
  • Data warehousing
  • Data modeling
  • Data pipeline design
  • Data migration
  • SQL expertise
  • Data quality assurance
  • Data pipeline control

Languages

English
Full Professional
Hindi
Full Professional
Telugu
Native or Bilingual

Timeline

Data Engineer

Ct Tech Services
02.2025 - Current

Rubik Systems
10.2023 - 11.2024

Rolang technologies
06.2023 - 10.2024

University of Findlay
08.2022 - 05.2023

Account Receivable Caller

JSD Medical and Business Management
01.2020 - 07.2021

Medical Coder

Valliappa Medical Management
06.2019 - 12.2019

Bachelor of Science (B.S.) - Biotechnology

Aurora Degree and PG College

Master of Science (M.S.) - Health Informatics

University of Findlay
Vijetha Baddam