Summary
Overview
Work History
Education
Skills
Certification
Accomplishments
Timeline
Generic

Goutham Bilakanti

Plano,TX

Summary

  • Senior Software Engineer with 15+ years of Expertise in Data Engineering, AWS Cloud Technologies and Business Intelligence, Specializing in ETL,Data Warehousing and SQL.
  • Led the design and deployment of cloud-based data warehouse solutions on AWS, ensuring seamless integration and real-time reporting.
  • Optimized data pipelines using partitioning,compression and Chance Data Capture(CDC) to enhance performance and reduce operational overhead.
  • Implemented and optimized Delta tables within Databricks environment for efficient data processing and management.
  • Terraform, Automated infrastructure deployment and management via terraform scripts. Deployed AWS resources via terraform scripts.
  • Good understanding on Teradata Performance tuning measures including collecting Stats, Primary Index, Primary Partitioned Index, Join Indexes, Secondary Indexes, analyzing explain plans.
  • Solid understanding and practical experience with Apache Spark for distributed data processing.
  • Strong DDL and DML writing skills as well as capable to write complex SQLs for data analysis.
  • Excellent problem-solving skills with strong technical background and good interpersonal skills.

Overview

4
4
years of professional experience
1
1
Certification

Work History

Senior Software Engineer

Cigna
Plano, TX
12.2020 - Current

Project 1: Symplr to Facets Integration.

Technology: API Gateway, AWS Glue, Step Functions, SNS, SQS, Lambda, Event Bridge, RDS (PostgreSQL), Cloud Watch.

Description:

The integration of Symplr to Facets involves utilizing Symplr as the single source of truth (SOT) for provider data, which needs to be seamlessly integrated into the Facets database. Symplr is utilized to centralize all provider data, ensuring that it is the most accurate and up-to-date information available. This eliminates discrepancies and inconsistencies that may arise from multiple data sources. The goal is to seamlessly transfer data from Symplr into the Facets database. This integration allows Facets to access reliable provider information, enhancing the quality of healthcare services, and administrative processes.

Responsibilities:

  • Design the complete application workflow, create the architecture diagram, and develop the solution.
  • Designed and implemented scalable AWS architecture solutions using services such as Step Functions, S3, RDS, and Lambda.
  • Identify and recommend continuous improvement opportunities to enhance customer satisfaction, internal work efficiencies, and cost reduction opportunities.
  • Developed ETL processes using AWS Glue, transforming and loading data into RDS PostgreSQL for Facets Integration.
  • Integrated AWS EventBridge to create event-driven architectures, enabling real-time data processing and notifications.
  • Conducted performance tuning and optimization of AWS services to improve application responsiveness and efficiency.
  • Collaborated with cross-functional teams to gather requirements and design effective cloud solutions tailored to business needs.
  • Monitored and maintained AWS environments using CloudWatch, ensuring high availability and performance of applications.

Project 2: FHIR Interoperability.

Technology: AWS Glue, AWS SSM, Lambda, API Gateway, Databricks, Delta Table, Auto Loader, Elastic Search, Athena, Cloud Watch.

Description:

Led the implementation of Fast Healthcare Interoperability Resources (FHIR) on AWS, utilizing various cloud services to enhance healthcare data. This project is a next-generation standards framework created by HL7. FHIR solutions are built from a set of modular components called 'Resources.' These resources can easily be assembled into working systems that solve real-world clinical and administrative problems at a fraction of the price of existing alternatives. FHIR is suitable for use in a wide variety of contexts – mobile phone apps, cloud communications, server communication in large institutional healthcare providers, and much more.

Responsibilities:

  • Designed and implemented AWS S3 resources, optimizing the data storage and retrieval process for efficient FHIR data handling.
  • Orchestrated API gateway and Lambda functions to create a seamless and secure interface for FHIR data exchange.
  • Implemented Step Functions to automate and streamline complex workflows.
  • Utilized AWS Glue for data integration and transformation.
  • Collaborated with cross-functional teams to provide technical guidance and strategic insights.
  • Integrated Elasticsearch to enable powerful and efficient search capabilities, enhancing data discovery and retrieval.
  • Designed and implemented IAM roles to enforce fine-grained access control, ensuring secure and least privilege principles within the AWS environment.
  • Integrated SQS and SNS with other AWS services, such as Lambda, EC2, and S3. Implemented dead-letter queues and message retries to handle failed messages.
  • GitHub: Collaborated on version control, managing repositories, and pull requests.
  • Terraform: Automated infrastructure deployment and management via Terraform scripts. Deployed AWS resources via Terraform scripts.
  • Splunk for effective error monitoring and troubleshooting.
  • Created a dynamic Teradata procedure for CDOs and SDOs to maintain SCD-type 2 logic.
  • Created Informatica mappings to generate JSON files for CDOs/SDOs and uploaded them to AWS S3 buckets using AWS CLI and Python scripts.

Software Engineer, Senior Advisor.

Plano, TX.

Bank of America (Client): Randstad Technologies (Full Time) (June 2018 - December 2020).

Project: Digital Marketing.

Technology: Teradata, Hadoop, Sqoop, Hive.

Description:

The scope of the project is to maintain the customer information in the MDS system. We receive the data from different source systems on various subject areas, such as Home Loans, Auto Loans, Sales, and Promotion.

Email campaign, etc. The campaign program is maintained by a third-party vendor, and True Value will extract the sales data and share it with the vendor for campaign programming.

Responsibilities:

  • Worked extensively on BTEQ and UNIX Shell scripting.
  • Resolved issues related to the semantic layer, or reporting layer.
  • Worked on different subject areas, like Campaign, Digital Marketing, Promotion, and Item.
  • Written complex SQLs using joins, subqueries, and correlated subqueries. Expertise in SQL queries for cross-verification of data.
  • Involved in data ingestion into HDFS using Sqoop from a variety of sources, like Teradata, using connectors such as JDBC and import parameters.
  • Designed and developed the scripts to load the data into Hive.
  • Performed various performance optimizations, such as using distributed cache for small datasets, partitions, and bucketing in Hive.
  • Implemented partitioning, dynamic partitions, and buckets in HIVE, and analyzed the partitioned and bucketed data to compute various metrics for reporting.
  • Involved in creating Hive tables, loading them with data, and writing HIVE queries that will run internally in a MapReduce way. Involved in using HCatalog to access Hive table metadata from MapReduce.
  • Responsible for migrating tables from traditional RDBMS to Hive tables using Sqoop.
  • Developed the Teradata macros, stored procedures to load data into worktables, and then move data from worktables into base tables.
  • Reviewed the SQL for missing joins, join constraints, data format issues, mismatched aliases, and casting errors.
  • Responsible for design, data mapping analysis, and mapping rules.
  • Responsible for development, coding, and testing.
  • Responsible for implementation and post-implementation support.
  • Extensively used loader utilities to load flat files into Teradata RDBMS.
  • Used BTEQ and SQL Assistant (Queryman) front-end tools to issue SQL commands matching the business requirements to the Teradata RDBMS.
  • Created TPT to transfer the data from SQL Server to Teradata.
  • Collected statistics periodically on tables to improve system performance.
  • Performed tuning and optimization of application SQL using query analyzing tools.

Senior Software Engineer.

Plano,TX

CVS Pharmacy, Inc. (Client) (Apr 2017 to Jun 2018)

Company: Cognizant Technology Solutions (Full-Time)

  • Involved in data ingestion into HDFS using Sqoop from a variety of sources, like Teradata, using connectors such as JDBC and import parameters.
  • Designed and developed the scripts to load the data into Hive.
  • Performed various performance optimizations, such as using distributed cache for small datasets, partitions, and bucketing in Hive.
  • Implemented partitioning, dynamic partitions, and buckets in HIVE, and analyzed the partitioned and bucketed data to compute various metrics for reporting.
  • Involved in creating Hive tables, loading them with data, and writing Hive queries that will run internally in a map-reduce way. Involved in using HCATALOG to access Hive table metadata from MapReduce.

Senior Software Engineer.

Hyderabad,INDIA

American Express (Client) - April 2012 to April 2017.

Company Cognizant (Full Time)

  • Prepare test cases, test plans, and strategies to test the code that will be developed.
  • Work with the source system analysts to understand the windows available for data extraction.
  • Increased performance by 35% to 40% in some situations. With the in-depth expertise in the Teradata cost-based query optimizer, I identified potential bottlenecks with queries from the aspects of query writing, skewed redistributions, join order, optimizer statistics, and physical design considerations (PI, USI, NUSI, and JI, etc.). etc.
  • The scope of his responsibilities includes discussing with business partners to understand the requirements, making the required code changes as per the requirements, and testing the code using the given source files.
  • Once the code is tested, it will be moved to the repository for version maintenance, and it will be moved to production through the defined code deployment process.
  • Include interacting with different teams, such as the Source team, Reporting team, and Business owner, to ensure a smooth flow of the process in implementing the enhancements.
  • Parallel testing between Teradata and Sybase.
  • Converting Sybase procedures to Teradata procedures.
  • Has been involved in stored procedure conversion, history data migration of tables, and stored procedure testing.
  • Handled a team and effectively carried out node configuration, job scheduling of the nodes associated with the stored procedures in the Event engine.
  • Resolved complex stored procedure issues, such as spool space issues, target row updated by multiple source rows errors, etc. And also guided the team in debugging errors so that stored procedures can be deployed on time.
  • Gained thorough exposure to the MDF (file load) framework, which was suggested as a replacement in Teradata for all the direct file load Sybase stored procedures by the client.

Software Engineer

Hyderabad,India

Teradata/Dec 2008 to Apr 2017

  • Converting the existing DDL scripts into temporal/column partitioning (CP) tables.
  • Need to run existing SQL scripts on converted temporal/CP tables, and need to verify the behavior of these queries.
  • Tested the above scripts on multiple platforms.
  • Need to log a DR if any problem appears in this process.
  • Move the converted test scripts to the Parallel Test Environment (PTE).
  • Develop test scenarios for new TD13 and TD14 features and their interaction with existing Teradata Database features.
  • Creating test scripts for the above-mentioned test scenarios and assembling them into test suites.
  • Managed high-volume databases including all database objects, triggers, procedures, user right, and functions, packages, indexes, views, procedures in test cases.
  • Used Fast Load, Multi Load, and Fast Export utilities efficiently to import and export data to Teradata and from Teradata.
  • The problems faced by the customers are reproduced in-house and are added for repeated testing in the future.
  • Worked on utilities to load and export data to reproduce the problems using Fast Load, MLoad, etc. And worked on tools like SQL Assistant, TSET, and Viewpoint, etc.
  • Scripts to automate the testing of different features were developed as per requirements, and performance of customer queries was analyzed using DBQL and the Viewpoint portlet.
  • Involved in the analysis of the requirements and creating the design documents.
  • Responsible for finding which snippet of C code is creating the problem during the build generation process, and need to identify the associated code developer.
  • Responsible for upgrading all systems to the current CIBT build.
  • Responsible for monitoring sessions with the help of QrySession.
  • Need to promote new test results as correct control files for test scripts, if necessary.

Education

Master of Science - Computer Science

Nirvana Institution of Engineering
Osmania University,Hyderabad,India
08-2008

Skills

  • Cloud Platforms & Services: AWS (AWS Glue, Lambda, API Gateway, S3, IAM, SNS, SQS, Athena, Glue Catalog, RDS, CloudWatch, VPC, Subnets, KMS)
  • ETL Tools & Data Engineering: Teradata Utilities (FastLoad, TPT, MultiLoad), Informatica, DataBricks, PySpark
  • Programming Languages: Python
  • Business Intelligence Tools: Tableau
  • Database and Data Warehouse: Teradata, Postgres, Sybase, Oracle, MySQL, SQL Server
  • Other Tools & Technologies : Jira,Github,Jenkins,Splunk,Terraform,Elastic Search

Certification

  • Teradata Certified Professional (Version: TD12)
  • Databricks Certified Associate Developer for Apache Spark 2.4

Accomplishments

  • Received the EIM Ekalavya 2015 award from Cognizant for contributing to positive client feedback or satisfaction, and delivering high-quality results consistently, ensuring that the project's objectives are met efficiently and effectively
  • Received the EIM Maestro 2012 award from Cognizant for developing unique solutions to complex challenges that improve project outcomes, or streamline processes

Timeline

Senior Software Engineer

Cigna
12.2020 - Current

Master of Science - Computer Science

Nirvana Institution of Engineering
Goutham Bilakanti