Summary
Overview
Work History
Education
Skills
Timeline
Generic

Sarath Chandra Bandarupalli

Jamaica Plain,MA

Summary

Highly skilled and results-driven Data Engineer with over 4 years of hands-on experience in designing, developing, and maintaining data pipelines and infrastructure. Adept at leveraging cutting-edge technologies and best practices to ensure the efficient and secure flow of data throughout an organization. Proficient in data modeling, ETL processes, and data warehousing, with a strong ability to collaborate with cross-functional teams to meet business objectives. Demonstrated expertise in optimizing data architecture for performance, scalability, and reliability. Committed to staying current with industry trends and continuously enhancing data engineering skills to drive innovation and deliver valuable insights to drive informed decision-making. Seeking opportunities to contribute my expertise to a dynamic organization focused on harnessing the power of data to achieve strategic goals.

Overview

7
7
years of professional experience

Work History

Data Engineer

Indata Solutions
01.2023 - Current
  • I design, develop, and maintain data pipelines using technologies such as Apache Airflow and Apache NiFi, ensuring efficient data extraction, transformation, and loading (ETL) from sources like databases, APIs, and external data feeds
  • I've implemented Kafka as our real-time data streaming platform, facilitating data ingestion and message passing among microservices
  • I use Kafka Connect to integrate various data sources and sinks
  • I harness the power of PySpark and Scala in our data processing stack, enabling distributed and parallel data processing with Apache Spark, while Scala is used for complex data transformations and analytics tasks
  • Custom Python scripts are employed for automation tasks, leveraging libraries like Pandas, NumPy, and Requests to manipulate and manage data efficiently
  • Our data infrastructure resides on Azure, utilizing services such as Azure Data Factory for orchestration, Azure Databricks for big data analytics, and Azure Kubernetes Service (AKS) for containerized applications
  • I employ technologies like Apache Nifi and Great Expectations to establish data quality checks, validation rules, and anomaly detection mechanisms, ensuring high-quality data throughout the pipeline
  • Utilizing Apache Hive and Apache Pig for data transformation and SQL-based queries, I prepare data for reporting and analytics, enabling data-driven decision-making for stakeholders
  • Our codebase is managed using Git and GitHub, allowing for collaborative development, branching, and pull request workflows, ensuring code quality and traceability
  • I continuously optimize data pipelines with technologies like Apache Flink and Apache Kafka Streams, improving processing speed, resource utilization, and overall system performance
  • Tools such as Prometheus and Grafana are used for real-time monitoring and alerting, allowing for proactive issue identification and resolution in our data pipelines
  • I maintain comprehensive documentation using tools like Confluence, detailing data pipelines, processes, configurations, and best practices, facilitating knowledge sharing and onboarding of new team members
  • Collaborating with cross-functional teams, I work with data scientists who use Jupyter Notebooks, and analysts who use tools like Tableau and Power BI, ensuring their data requirements are met with tailored solutions that support their objectives.

Data Engineer

Cloud Infra IT Solutions
04.2022 - 07.2022
  • Employing big data technologies like Hadoop and Spark, with a strong emphasis on Scala and PySpark, to process and analyze vast volumes of healthcare data efficiently
  • This enabled the identification of trends and patterns for disease prevention and treatment optimization
  • Utilizing containerization and orchestration tools such as Docker and Kubernetes to build scalable and portable data applications, ensuring seamless deployment across diverse healthcare environments
  • Implementing data lakes and data warehouses using technologies like Amazon S3, Redshift, and Snowflake to centralize and store structured and unstructured healthcare data, facilitating comprehensive analytics and reporting
  • Collaborating with DevOps teams to establish continuous integration and continuous deployment (CI/CD) pipelines, ensuring that data solutions were always up to date and aligned with healthcare industry standards
  • Integrating natural language processing (NLP) libraries like NLTK and spaCy to extract valuable insights from unstructured clinical notes and medical literature, supporting research efforts and enhancing medical decision-making
  • Enhancing data security and privacy measures by implementing encryption and access controls in compliance with HIPAA (Health Insurance Portability and Accountability Act) and other regulatory requirements
  • Employing Scala and PySpark for data transformation, manipulation, and feature engineering, enabling advanced analytics on healthcare datasets
  • Using data visualization tools such as Tableau and Power BI to create intuitive dashboards and reports, enabling healthcare professionals and administrators to make data-driven decisions in real-time
  • Collaborating with data governance teams to establish data cataloging and metadata management practices, ensuring data lineage and compliance tracking for healthcare data assets
  • Conducting performance tuning and optimization of data pipelines to meet stringent healthcare data processing requirements, minimizing latency and ensuring high data availability for critical applications
  • Participating in cross-functional knowledge sharing sessions and mentoring junior data engineers to foster a culture of continuous learning and innovation within the organization.

Data Engineer

Tata Consultancy Services
05.2019 - 12.2020
  • Led the development of data pipelines for our e-commerce platform, utilizing cutting-edge technologies such as Apache Spark, Apache Kafka, and Hadoop
  • These pipelines efficiently extracted, transformed, and loaded (ETL) vast datasets, resulting in a 20% reduction in data processing time, thereby improving the real-time inventory management system
  • Designed and optimized data warehousing solutions on platforms like Amazon Redshift and Google BigQuery
  • This enabled timely access to critical sales and customer data, supporting personalized marketing efforts and resulting in a 15% increase in cross-selling and upselling
  • Implemented advanced performance tuning techniques to handle peak traffic during sales events, resulting in a 30% improvement in system responsiveness and customer experience
  • Developed robust data quality checks and monitoring processes, reducing data inconsistencies by 25%
  • This improved product recommendations and enhanced customer satisfaction by ensuring accurate product availability information
  • Collaborated closely with data scientists, analysts, and marketing teams to understand their data requirements
  • These collaborations led to the creation of targeted marketing campaigns, resulting in a 10% increase in conversion rates
  • Introduced automation scripts and CI/CD pipelines to streamline deployment processes, enhancing the efficiency of data pipeline deployments by 30%
  • This automation ensured rapid updates to the e-commerce platform, keeping it competitive and up-to-date
  • Implemented rigorous data security protocols to safeguard sensitive customer information and ensure compliance with industry regulations
  • This maintained customer trust and data integrity throughout the e-commerce project
  • Created comprehensive documentation for data engineering processes and conducted knowledge-sharing sessions with cross-functional teams
  • This facilitated seamless onboarding for new team members and fostered a culture of continuous learning, supporting the growth of our e-commerce project
  • My contributions as a Data Engineer on this e-commerce project not only optimized data operations but also played a pivotal role in the project's success, ensuring data-driven decision-making, innovation, and customer satisfaction.

Software Engineer Intern

Tata Consultancy Services
03.2017 - 03.2019
  • Collaborated closely with cross-functional teams, including developers, designers, and quality assurance engineers, to deliver high-quality software solutions that met client requirements and exceeded expectations
  • Wrote, tested, and debugged code in multiple programming languages, ensuring the reliability and performance of software applications
  • Developed effective problem-solving skills by identifying and resolving complex technical issues, contributing to project efficiency and reducing potential roadblocks
  • Utilized version control systems (e.g., Git) to manage codebase, enabling seamless collaboration and tracking changes across the development lifecycle
  • Created comprehensive documentation for software designs, code, and user manuals, ensuring clear communication and ease of maintenance for future development teams
  • Adapted to Agile development methodologies, participating in sprint planning, daily stand-up meetings, and retrospectives to improve team coordination and project delivery
  • Gained proficiency in a wide range of technologies and tools, expanding my technical expertise and adaptability
  • Demonstrated a commitment to ongoing learning and professional growth by attending training sessions, workshops, and staying updated with industry trends
  • Worked closely with clients to understand their specific needs, provided regular project updates, and ensured client satisfaction throughout the development lifecycle
  • Actively participated in knowledge-sharing sessions and mentored junior team members, fostering a collaborative and learning-oriented work environment
  • Assisted in the testing process by designing and executing test cases, identifying and reporting defects, and contributing to the overall quality assurance efforts
  • Collaborated with senior engineers to analyze and optimize code for improved performance, resulting in faster and more efficient software applications
  • Played a vital role in project coordination by assisting in project planning, tracking progress, and ensuring timely delivery of milestones
  • Actively sought feedback from senior team members and integrated their suggestions into my work, facilitating continuous improvement in coding practices and project execution
  • Demonstrated adaptability and a willingness to learn by taking on tasks outside of my core responsibilities, such as database management, server administration, or front-end development, when required
  • Conducted research on emerging technologies and software development trends, presenting findings to the team and proposing potential implementations for future projects
  • Exhibited strong interpersonal skills, fostering effective communication and collaboration within the team, resulting in smoother project workflows
  • Took the initiative to identify potential bottlenecks or challenges in project execution and proposed proactive solutions to mitigate risks
  • Assisted in enhancing project documentation processes, ensuring that project artifacts were organized, accessible, and useful for future reference
  • Presented the work completed during the internship to colleagues and management, showcasing project achievements, challenges overcome, and lessons learned.

Advanced Academic Tutor

Gokaraju Rangaraju, engineering & technology
09.2018 - 01.2019
  • Utilized learning management systems like Moodle and Blackboard to facilitate interactive and engaging tutoring sessions, particularly in computer science courses
  • These platforms enabled personalized learning experiences and easy access to course materials, contributing to improved student outcomes
  • Collaborated closely with faculty and staff to seamlessly integrate educational technologies into the computer science curriculum, ensuring the effective use of tools like to enhance supplemental instruction methods
  • Leveraged data-driven approaches and educational analytics tools, including learning management system(LMS) such as Moodle, Blackboard, programming environments like Java and Python, multimedia tools like Adobe Creative Suite, and education analytical tools to enhance supplemental instruction methods.", to track and document student progress
  • This data-driven approach helped identify areas of improvement and allowed for data-driven adjustments to tutoring techniques
  • Contributed to the development of digital study materials using Adobe Creative Suite, created interactive online exercises, and curated resources specifically tailored to computer science topics
  • These digital resources enhanced computer science students' self-guided learning experiences and fostered engagement
  • Maintained a high level of professionalism, fostering a positive and inclusive learning environment that encouraged open communication and active participation among computer science students while harnessing the power of educational technology to facilitate learning.

Tutor

Khammam institute of science and technology
01.2018 - 08.2018
  • Providing personalized one-on-one and group tutoring sessions to students across various subjects, focusing on mathematics, science, and technology-related courses
  • Assessing students' strengths and weaknesses to tailor tutoring approaches and develop customized study plans, resulting in improved comprehension and grades
  • Facilitating interactive and engaging learning experiences, utilizing innovative teaching methods to simplify complex concepts and promote critical thinking
  • Monitoring student progress and adapting teaching strategies as needed to ensure continuous improvement
  • Offering guidance on study techniques, time management, and test preparation strategies to empower students to become independent learners
  • Collaborating with faculty members to align tutoring sessions with classroom curriculum and reinforce classroom instruction
  • Maintaining accurate records of tutoring sessions and student progress to assess effectiveness and make data-driven improvements.

Education

Master of Professional Studies - Informatics

College of Professional Studies, Northeastern University
Boston, MA
12.2022

Bachelor of Technology - Computer Science & Engineering

Gokaraju Rangaraju Institute of Engineering & Technology
Hyderabad, TG, India
04.2019

Skills

  • Hadoop Ecosystem: MapReduce, HDFS, Hive, Spark, Kafka
  • Language: C, Python, Scala, UML
  • Web Technologies: JavaScript, JSP, Servlets, JDBC, Unix/Linux Shell Scripting, Python, HTML, XML
  • Methodologies: Waterfall, Agile/Scrum
  • Databases: Oracle, MySQL, HBase
  • Application/Web server: Apache Tomcat, WebSphere and JBoss
  • IDE’s: Eclipse, Netbeans, VS Code
  • ETL & Reporting Tools: Informatica, SAP Business Objects, Power BI
  • Cloud Infrastructures: Azure
  • Frameworks: AngularJS, ReactJS, MySQL, Mongo DB, Oracle DB, Node JS
  • Soft skills: Collaboration, Interpersonal Skills, communication Skills, Problem Solving, Critical Thinking

Timeline

Data Engineer

Indata Solutions
01.2023 - Current

Data Engineer

Cloud Infra IT Solutions
04.2022 - 07.2022

Data Engineer

Tata Consultancy Services
05.2019 - 12.2020

Advanced Academic Tutor

Gokaraju Rangaraju, engineering & technology
09.2018 - 01.2019

Tutor

Khammam institute of science and technology
01.2018 - 08.2018

Software Engineer Intern

Tata Consultancy Services
03.2017 - 03.2019

Master of Professional Studies - Informatics

College of Professional Studies, Northeastern University

Bachelor of Technology - Computer Science & Engineering

Gokaraju Rangaraju Institute of Engineering & Technology
Sarath Chandra Bandarupalli