Summary
Overview
Work History
Education
Skills
Certification
Projects
Websites, Portfolios, Profiles
Timeline
Generic

Nikhitha Maddineni

Houston,Texas

Summary

Experienced Senior Data Engineer with a strong background in designing and implementing scalable data solutions across multiple organizations. Expertise in Azure Data Lake, Azure Synapse, Power BI, and data pipeline orchestration to drive unified analytics and real-time performance metrics. Proven ability to enhance data quality, ensure compliance, and deliver actionable insights by collaborating effectively with cross-functional teams including analytics, product, and operations. Adept at optimizing data architecture and supporting business intelligence initiatives in healthcare and technology sectors.

Overview

7
7
years of professional experience
1
1
Certification

Work History

Sr. Data Engineer

SynergenX Health Holdings
Houstonia, Texas
05.2024 - Current
  • Integrated Dayforce HR and payroll data with enterprise data warehouse using ODBC and REST API connectors, enabling unified analytics across platforms.
  • Integrated Azure Blob and Data Lake Storage accounts with Databricks notebooks for scalable data transformation and advanced analytics, including patient outcomes, and financial metrics.
  • Built interactive Power BI dashboards using DAX and Power Query, presenting real-time KPIs for clinical performance, revenue cycle management, and patient scheduling, derived from EHR and PM systems.
  • Designed Power BI reports to visualize employee timesheets, scheduling, compensation trends, and compliance metrics, utilizing Dayforce and EHR data.
  • Established secure linked server configurations for access to on-premise AMD medical databases and EHR/PM systems, facilitating real-time data integration.
  • Collaborated with HR and Operations to develop actionable KPIs from Dayforce, EHR, and PM platforms, focusing on staffing coverage and performance metrics.
  • Ensured HIPAA and SOC 2 compliance while managing sensitive employee and patient data within Azure Storage and Power BI environments by enforcing RBAC policies.
  • Worked on GA4 data to streamline patients' marketing ROI and LTV data models.
  • Implemented data marts and snowflake data models in reporting.

Data Engineer

Datics/Walmart
Charlotte, NC
08.2023 - 04.2024
  • Employed HL7 message routing and message filtering techniques to efficiently distribute HL7 messages to the appropriate destination systems and applications.
  • Strong understanding of AI model lifecycle management, including versioning, monitoring, and performance optimization.
  • Implemented data retention and data deletion policies for PII and PHI data in compliance with data privacy laws, managing the lifecycle of sensitive information appropriately.
  • Utilized Snowflake to implement AI and machine learning (ML) solutions, effectively enhancing data engineering processes and enabling data-driven insights.
  • Collaborated with multiple teams to develop cutting-edge analytics-based solutions.
  • Demonstrated a deep understanding of GCP services, including Google BigQuery, Google Cloud Storage, Google Cloud Dataflow, and Google Cloud Pub/Sub, for data storage, processing, and streaming.
  • Successfully installed and configured Hadoop ecosystem components, ensuring the seamless integration of Big Data technologies with Snowflake.

Graduate Assistant

University of Houston
Houston, TX
03.2023 - 05.2023
  • Developed proficiency in using MS Power BI and DAX to create tabular models and visualize data effectively, enabling interactive and dynamic data exploration.
  • Revamped a legacy data pipeline for student academic data using AWS Glue, transforming student records for compliance with new regulations.
  • Developed custom ETL scripts using Python and Glue's dynamic frame API, ensuring accurate data anonymization and aggregation.
  • Configured and optimized the integration run times for data processing activities using Azure SQL DB, and Azure Synapse Analytics.
  • Built complex and external triggers to automate the scheduled events and data pipelines for student
    enrollment data.
  • Assisted in managing and maintaining data repositories, ensuring data quality, and developing data dictionaries to facilitate data understanding across the organization.
  • Orchestrated end-to-end data pipelines using Azure Data factory to extract customer interactions from
    multiple sources and transformed them into actionable insights.
  • Utilized Data Factory's scheduling and monitoring features to ensure timely data updates, resulting in a
    25% increase in personalized marketing campaign effectiveness.

Data Engineer

IBM India Pvt Ltd
Pune, Maharastra
06.2019 - 12.2021
  • Utilized Amazon Redshift for warehousing and parallel query processing while working larger SOX and currency related applications datasets.
  • Worked on of the Data modeling (Dimensional & Relational) concepts like Star - Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables.
  • Created roles and access level privileges and taken care of Snowflake Admin Activity end to end and IAM in cloud such as AWS and Azure.
  • Implemented data cleansing and data quality checks within ETL workflows, reducing data anomalies by 30% and improving the overall accuracy and reliability of business reports.
  • Used Snowflake SQL Writing SQL queries against Snowflake Developing scripts Unix, Python etc to Extract, Load and Transform data.
  • Utilized dimensional modeling techniques to structure data marts, ensuring optimal query performance and user accessibility.
  • Analyzed query execution plans and identified indexing gaps, leading to a 60% reduction in query execution times.
  • Revamped a legacy data pipeline for a healthcare provider data using AWS Glue, transforming patient records for compliance with PHI, PII standards.
  • Utilize Snowpipe, stage object, Streams & tasks to build optimal snowflake solutions.
  • Conducted ETL testing, including extracting data from databases, transforming data, and uploading it to data warehouse servers using SQL and NoSQL.
  • Responsible for designing and developing semantic layers that caters to high visibility Tablaeu reports.
  • Experience in developing Proof of Concepts using various snowflake features like SnowPipe, stage object, stream & tasks, snowflake procedures.
  • Analyzed business intelligence and trend data using tools such as Tableau and Power BI to provide actionable recommendations.
  • Recommended metrics and models based on observed trends, leveraging statistical analysis tools such as SPSS and SSAS.
  • Proficiently utilized Azure Synapse (formerly SQL Data Warehouse) for managing large-scale data processing and analytics workloads.

Data Analytics Intern

Infosys Pvt Ltd
Mysore, Karnataka
12.2018 - 05.2019
  • Applied data analysis techniques to financial and investment portfolios, providing valuable insights for risk assessment and portfolio optimization.
  • Collaborated with business analysts to define key metrics and implemented ETL pipelines to populate the data mart from various transactional systems.
  • Resulted in a 40% reduction in report generation time, enabling faster decision-making and reducing inventory holding costs.
  • Conducted exploratory data analysis to identify patterns, trends, and outliers, enabling the identification of business opportunities and risks.
  • Leveraged Synapse's integrated analytics capabilities to provide executives with near real-time insights into supply chain operations, leading to a 15% reduction in delivery times.
  • Integrated Google Cloud Platform Dataproc, dataflow and Amazon Web Service into programming tasks.
  • Implemented data quality checks in PostgreSQL and PL/SQL to ensure accuracy in financial reporting, integrating with cloud databases for seamless data management.
  • Utilized advanced Excel functions, including data formatting, Pivot tables, V-lookup, and dashboard creation to streamline data representation and enhance decision-making processes.
  • Designed and implemented data warehousing solutions on Azure Synapse, optimizing query performance and resource utilization.
  • Developed custom transformations and applied data enrichment techniques using AWS Glue to meet specific business requirements.
  • Integrated MDX queries with data visualization tools such as Excel PivotTables and Power BI reports, delivering interactive and user-friendly analytical solutions for business stakeholders.

Education

Master of Science - Computer Science With Data Analytics Core

University of Houston
Houston, TX, USA
05.2023

Bachelor of Science - Computer Science and Engineering

Koneru Lakshmaiah Education Foundation
Guntur, India
05.2019

Skills

  • Programming languages: C, Python, Java, R, and PySpark
  • Software tools: Power BI, Azure Data Studio, Azure Synapse Analytics, Azure Machine Learning Studio, SQL Developer, RStudio, Tableau, Excel, Snowflake, MATLAB, DBeaver
  • Analytical skills: SQL, Power BI, Python, R, Spark SQL, Excel
  • Azure data engineering: Azure Data Factory, Azure Data Lake (Gen1/Gen2), Azure Synapse, Azure Databricks
  • Database development: Stored procedures, SSIS, SSRS
  • Cloud databases: SQL Server, Oracle, MySQL, MongoDB
  • Data warehousing: Azure Synapse Analytics, Apache Spark
  • Python libraries: Pandas, NumPy
  • Cloud platforms: Azure Data Factory, GCP BigQuery
  • JavaScript frameworks: React, Angular
  • ETL tools: Informatica
  • Data modeling techniques: Regression analysis
  • Project management tools: JIRA
  • Mathematics and problem solving

Certification

IBM Agile explorer, IBM Certified Data Analyst, IBM certified Data science methodologies, Microsoft certified Data Analyst, SnowPro core certified. Microsoft Azure AI.

Projects

Construction Safety Analysis using OSHA Dataset. -Apr 2023 Conducted a Construction Safety Analysis utilizing a web-scraped dataset of 26,000 OSHA records. Preprocessed the data using advanced NLP techniques such as stemming and lemmatization to improve data quality. Utilized Principal Component Analysis (PCA) to reduce dataset dimensionality and applied TF-IDF Vectorization. Employed K-means Clustering technique to identify optimal K value and centroids, and named resulting clusters based on safety incident type or severity. Results were used to prioritize safety improvements and identify areas, Online Optimization of SVM Classifier using Kernel and Ensemble Techniques. -Aug 2022 Designed custom kernel functions for SVM using kernel tricks to train unbalanced data. Used Boosting and Bagging methods on custom kernel SVM models to improve accuracy. Increased accuracy of the models by more than 15% when compared with inbuilt SVM functions., Online class monitoring tool using facial recognition and emotion analysis. May 2022 Used Pandas, SciPy, Scikit-learn, PyTorch, and other libraries to process images from video frames. Developed modules using state-of-the-art technologies such as FaceNet and DeepFace for facial recognition and emotion analysis. Documented the emotions of students throughout a video lecture and presented a detailed report.

Websites, Portfolios, Profiles

  • Https://nikitha5463-cifar-10-classification-demo-ul3q97.streamlit.app/
  • Https://github.com/NikhithaM97

Timeline

Sr. Data Engineer

SynergenX Health Holdings
05.2024 - Current

Data Engineer

Datics/Walmart
08.2023 - 04.2024

Graduate Assistant

University of Houston
03.2023 - 05.2023

Data Engineer

IBM India Pvt Ltd
06.2019 - 12.2021

Data Analytics Intern

Infosys Pvt Ltd
12.2018 - 05.2019

Master of Science - Computer Science With Data Analytics Core

University of Houston

Bachelor of Science - Computer Science and Engineering

Koneru Lakshmaiah Education Foundation
Nikhitha Maddineni
Want your own profile? Create for free at Resume-Now.com