Summary
Overview
Work History
Education
Skills
Certification
Work Availability
Timeline
Hi, I’m

Balasiva Pindra

Staff Data Engineer
Palo Alto,CA
It is human nature to think wisely and act in an absurd fashion.
Anatole France
Balasiva Pindra

Summary

Profile:

Highly skilled Data Engineering with a proven track record in leading high-performing teams and architecting advanced data solutions. I bring extensive expertise in AWS, Databricks, Snowflake, Redshift, CI/CD automation tools, and programming languages such as Python, PySpark, and Java. Recognized for designing and implementing scalable data solutions, optimizing pipelines, and ensuring unmatched data quality. I am committed to delivering actionable insights and spearheading data-driven decision-making initiatives for organizations.Ability to lead cross-functional teams, foster collaboration, and architect robust data solutions that exceed organizational objectives. A results-driven professional with a passion for staying at the forefront of data engineering advancements. Ready to contribute skills and strategic vision to drive success in challenging and dynamic environments..

Overview

9
years of professional experience
5
years of post-secondary education

Work History

Rivian

Staff Data Engineer
01.2020 - Current

Job overview

Architected Enterprise Data Warehousing Solutions:

  • Lead an architecture of comprehensive data warehousing solutions for Rivian, employing cloud technologies including AWS, Databricks, and Bigdata (PySpark)

Developed Scalable Data Pipelines:

  • Engineered and implemented scalable data pipelines to efficiently extract, transform, and load large volumes of both structured and unstructured data with in Streaming and batch
  • Achieved a remarkable 30% increase in data processing speed, showcasing a commitment to optimizing performance and operational efficiency

Enterprise Data Warehouse Management:

  • Established and managed enterprise data warehouse comprising over 1000 tables and 100 pipelines, orchestrated seamlessly through AWS Airflow
  • Developed, implemented and maintained data analytics protocols, standards, and documentation

Custom Packages for Self-Service Applications:

  • Designed and implemented custom packages to empower business users with self-service applications, enabling them to effectively utilize data pipelines within platforms like SharePoint and AWS S3
  • Streamlined user accessibility and contributed to heightened organizational efficiency

Cross-Functional Collaboration:

  • Collaborated effectively with cross-functional teams including commercial, supply chain, product development, manufacturing, and service teams.Translated their analytical needs into actionable insights, leading to the development of models and rigorous data quality checks for enhanced accuracy and relevance across diverse business functions

CI/CD Implementation:

  • Designed and developed CI/CD processes using GitLab, Terraform, Pulumi, and CloudFormation for infrastructure as code

Migration of On-Premises Data Infrastructure:

  • Led the migration of on-premises data infrastructure for manufacturing data
  • Which results in a 25% reduction in infrastructure costs while simultaneously improving system availability and scalability.

Capital One

Sr Big Data Engineer
03.2018 - 01.2020

Job overview

AWS Server Configuration and Routing:

  • Configured ALE Server on EC2 instances within a VPC, employing ALB, NLB, ELB, and router configurations, seamlessly integrated with Route 53

CI/CD Deployment with Jenkins:

  • Implemented AWS infra deployment using Jenkins CI/CD, featuring automatic test report generation using PyUnit & PyTest.

Containerization for Critical Jobs:

  • Containerized critical jobs exceeding Lambda runtime, deploying them in AWS ECS
  • Executed Data Quality checks in Python & Spark on Snowflake data, with parallel processing for multiple queries and aggregate data verification against defined standards

Data Quality Visualization and Reporting:

  • Designed data quality jobs that visualize various data plots in a graphical format, sending daily email reports to responsible stakeholders

Real-Time Streaming with Kafka:

  • Developed Java producers for Kafka to subscribe to ZeroMQ streaming data in real-time
  • Utilized Java and multithreading concepts to ensure continuous streaming, running inside ECS containers

Datadog Metrics:

  • Sent metrics and traces to Datadog, while routing logs to CloudWatch for real-time producer monitoring

Self-Healing Streaming Containers:

  • Implemented self-healing streaming containers using a custom methodology
  • Ensured continuous streaming by automatically restarting containers in case of technical failures, with alerts sent to Slack, Email & PagerDuty

Data Serialization and Service Development:

  • Developed a generic de-serializer in Python for converting Protobuf data into JSON & Avro as part of all code activities
  • Created an ALE REST API service using masked data, deployed in dev & QA regions for automated testing.

Logipro Software

Data Engineer
12.2014 - 12.2016

Job overview

  • ETL Process Design with SQL Server:
    Designed ETL processes using Informatica to load data from Flat Files, SQL server, and Excel files into the target SQL Server Data Warehouse database.
  • Support and Collaboration:
    Provided support for code/design analysis, strategy development, and project planning.Collaborated with infrastructure, network, database, application, and BI teams to ensure data quality and availability.
  • Spark for Interactive Queries and Streaming:
    Leveraged Spark for interactive queries, streaming data processing, and seamless integration with NoSQL databases to handle massive data volumes.Developed bespoke ETL solutions, encompassing batch processing and real-time data ingestion pipelines, utilizing Python and shell scripting to facilitate smooth data movement in and out of Hadoop.
  • Hadoop Ecosystem Deployment:
    Contributed to streamlining business processes for a regional bank by developing, installing, and configuring Hadoop ecosystem components.

Education

Concordia University, St. Paul
Saint Paul, MN

Masters in Information Technology Mangement
12.2016 - 12.2017

Acharya Nagarjuna Univerity

Bachelor of Science from Electronics & Communications Engg
06.2010 - 06.2014

Skills

Database:Databricks, Snowflake,Redshift,Google BigQuery

undefined

Certification

AWS Big Data Specialty

Availability
See my work availability
Not Available
Available
monday
tuesday
wednesday
thursday
friday
saturday
sunday
morning
afternoon
evening
swipe to browse

Timeline

AWS Big Data Specialty

02-2022

Staff Data Engineer

Rivian
01.2020 - Current

Sr Big Data Engineer

Capital One
03.2018 - 01.2020

Concordia University, St. Paul

Masters in Information Technology Mangement
12.2016 - 12.2017

Data Engineer

Logipro Software
12.2014 - 12.2016

Acharya Nagarjuna Univerity

Bachelor of Science from Electronics & Communications Engg
06.2010 - 06.2014
Balasiva PindraStaff Data Engineer