Summary
Overview
Work History
Education
Skills
Certification
Timeline
Generic

Saikrishna Gajjela

Summary

  • Over 4.5+ years of experience, you bring a wealth of knowledge and expertise in various aspects of data engineering, cloud computing, and big data technologies.
  • Proficient in Azure with expertise in various services including VMs, Data Factory, Databricks, SQL Database, Logic Apps.
  • Skilled in designing and implementing data warehouses like Azure Synapse Analytics, Snowflake.
  • Proficient in managing databases such as SQL Server 2018, PostgreSQL, and Azure SQL.
  • Experienced in CI/CD pipelines using Jenkins and Azure DevOps, with a strong grasp of source control tools like Git, GitHub, GitLab, and Bitbucket.
  • Well-versed in project management tools including JIRA, ServiceNow, and Confluence.
  • Proficient in Python programming, with experience in developing data solutions and automation scripts.
  • Adept in big data technologies such as Apache Spark, Hadoop, and Hive, for processing and analyzing large datasets.
  • Experienced in containerization with Docker and Kubernetes, enabling scalable and portable deployment of applications.
  • Adept in streaming data technologies like Apache Kafka and Apache Flink for real-time data processing and analytics.
  • Skilled in data quality frameworks and governance practices, ensuring data integrity and compliance with regulatory requirements.
  • Experience with Snowflake cloud data warehouse and ADLS Gen2 for integrating data from multiple source systems.
  • Proficient in loading nested JSON formatted data into Snowflake tables,ensuring efficient data ingestion and storage.
  • Proficient with container systems like Docker, enabling the creation, deployment, and management of containers for application development and deployment.

Overview

5
5
years of professional experience
1
1
Certification

Work History

Data Engineer

JPMorgan Chase
, NJ
01.2024 - Current
  • Designed and implemented data pipelines to maintain data integrity and generate key performance indicators (KPIs) reflecting data passage. Stored and manipulated configurations in MongoDB using PyMongo.
  • Developed and maintained Azure Analysis Services models for business intelligence and data analytics requirements.
  • Utilized Kafka functionalities for messaging systems and loaded data from REST endpoints to Kafka.
  • Created JSON scripts for deploying pipelines in Azure Data Factory (ADF) and implemented UNIX shell scripts for
    database connectivity and parallel job execution.
  • Worked on big data integration and analytics using Hadoop, PySpark, Kafka, Storm, and web Methods. Managed relational database services with Azure SQL and integrated various data storage solutions.
  • Implemented copy activity, custom Azure Data Factory pipeline activities.
  • Automated data quality checks and error handling processes to ensure the integrity and reliability of datasets.
  • Integrated Azure SQL's features for high availability, automated backups, and performance optimization.
  • Data ingestion to one or more Azure services (Azure Data Lake, Azure Storage, Azure SQL, Azure DW), and processing the data in Azure Databricks.
  • Integrated SQLite database for data storage and retrieval, implementing CRUD operations, and ensuring data integrity.
  • Utilized Java programming language and XML for backend logic and layout design respectively.
  • Importing and exporting databases using SQL Server Integration Services (SSIS) and Data Transformation Services (DTS) packages.
  • Managed relational database services utilizing Azure SQL, ensuring reliability, scalability, and maintenance of critical data stores.

Data Engineer

Capgemini
Pune, IN
01.2020 - 12.2022

· Developed scalable solutions on Azure to sustain backend infrastructure for mobile banking application using services such as Azure Data Factory, Azure Databricks, and Azure SQL Database.

· Created migration strategies for transitioning legacy banking systems to Azure, ensuring smooth integration with tools like Azure Lift and Shift and Azure Migrate.

· Employed Azure Blob Storage, Azure HDInsight, and Azure Data Lake Storage to securely store app data.

· Optimized application performance using Apache Spark's distributed processing capabilities on the Azure platform for large datasets.

· Constructed a system in Azure Data Lake to store, process, and analyze significant big data workloads.

· Implemented Azure Data Lake security controls, including access control lists (ACLs), and Azure Active Directory integration, to secure sensitive data assets.

· Automated execution and monitoring of data pipelines using integrated GitHub tools within Azure environments.

· Employed GitHub to manage versions of data transformation scripts, ensuring uniformity in different environments.

· Rigorously tested data pipelines to guarantee the accuracy and reliability of financial information, while seamlessly integrating with Azure services like Azure Data Lake Storage and Azure Cosmos DB.

· Created and implemented microservices-based applications on Microsoft Azure Service Fabric ensuring high availability.

· Developed microservices in Azure Service Fabric utilizing Reliable Actors and Reliable Services.

· Evaluated ETL process errors and data inconsistencies for troubleshooting.

· Implemented incremental ETL strategies to enhance efficiency.

· Developed, tested, and automated complex ETL pipelines using Azure Data Factory, ensuring efficient data ingestion, transformation, and loading from diverse data sources to Azure SQL Database, and Azure Data Lake Storage.

· Leveraged Azure Data Migration services for extensive data warehousing tasks.

· Leveraged Apache Spark on Azure Databricks for distributed computing, facilitating the processing and analysis of large banking data.

· Delivered secure and efficient storage solutions for structured and unstructured data using Azure Blob Storage, Azure Data Lake Storage, and Azure HDInsight.

· Implemented incremental ETL strategies in Azure Data Factory, reducing data processing times, and optimizing resource usage by only processing changed data.

· Developed comprehensive data models in Azure SQL Database, ensuring optimal database performance, and supporting complex queries and reporting needs.

· Deployed Azure Stream Analytics and Azure Event Hubs for real-time data processing and analytics, providing up-to-the-minute insights into business operations.

· Ensured data security and compliance with industry standards by implementing Azure Key Vault, Azure Policy, and Role-Based Access Control (RBAC) across Azure resources.

· Developed interactive dashboards and reports using Power BI, integrated with Azure services, providing stakeholders with real-time insights and actionable intelligence.

· Automated data workflows using Azure Logic Apps and Azure Functions reduce manual intervention and optimize data processing tasks for better efficiency.

· Collaborated with cross-functional teams to gather requirements and translate business needs into technical specifications for data solutions.

· Implemented data visualization tools like Tableau and Power BI to create dashboards and reports for business stakeholders.

· Managed version control and deployment of data applications using Git, Docker, and Jenkins.

· Created stored procedures for automating periodic tasks in SQL Server.

Education

Master of Science - Computer Science

University of Central Missouri
Warrensburg, MO
05-2024

Bachelor of Science - Computer Science

Lovely Professional University
Punjab,In
06-2020

Skills

  • Azure
  • DataFactory
  • Databricks
  • Synapse
  • SQL
  • T-SQL
  • Docker
  • Git
  • Snowflake
  • Python
  • Pyspark
  • Java
  • Jira
  • ServiceNow
  • Data Quality Frameworks
  • Data Governance Practices
  • Hadoop
  • Hive
  • ADLS
  • PostgreSQL
  • Azure DevOps
  • IAM
  • Azure Functions
  • Azure Logic Apps
  • PostgreSQL

Certification

  • Microsoft Certified Azure Data Engineer.
  • Microsoft Certified: Azure Fundamentals.
  • Microsoft Certified: Security, Compliance, and Identity Fundamentals.

Timeline

Data Engineer

JPMorgan Chase
01.2024 - Current

Data Engineer

Capgemini
01.2020 - 12.2022

Master of Science - Computer Science

University of Central Missouri

Bachelor of Science - Computer Science

Lovely Professional University
Saikrishna Gajjela