Summary
Overview
Work History
Education
Skills
Certification
Languages
Timeline
Generic

Chetan Paunikar

Irving,TX

Summary

Dynamic data professional with 10+ years of experince with robust background in data architecture and pipeline development, committed to delivering impactful contributions. Expertise in streamlining data processes and enhancing data integrity through innovative solutions, utilizing advanced proficiency in SQL and Python to empower cross-functional teams and facilitate data-driven decision-making. Proven ability to design, build, and optimize complex data pipelines and ETL processes, ensuring seamless integration and robust data solutions. Thrives in collaborative environments while swiftly adapting to evolving needs, consistently driving results that enhance overall team success.

Overview

13
13
years of professional experience
1
1
Certification

Work History

Data Engineer

Synechron Technologies
10.2023 - Current
  • Collaborated on ETL (Extract, Transform, Load) tasks, maintaining data integrity and verifying pipeline stability.
  • .Fine-tuned query performance and optimized database structures for faster, more accurate data retrieval and reporting.
  • Enhanced data quality by performing thorough cleaning, validation, and transformation tasks.
  • Streamlined complex workflows by breaking them down into manageable components for easier implementation and maintenance.

Data Engineer

Synechron Technologies Pvt. Ltd.
04.2022 - Current
  • NDW (Nuveen Data Warehouse) is a common target state reporting platform across TIAA and Nuveen. It consists of IBOR, ABOR and AUM data. NDW was created to retire, consolidate and provide functionality of existing data warehouse that support Market and Distribution(IRD), Front Office (EQ/FISRP) and Shared services. NDW was migrated from Oracle Exadata to Snowflake.
  • Responsibilities:
  • 1. Converting the ETL DataStage jobs which connect to Oracle initially to connect to Snowflake for data warehousing.
  • 2. Conversion of Oracle sql , stored procedure and functions to Snowsql and Stored procedures supported by Snowflake
  • 3. Designing and building etl pipelines to bring data from existing datalake.
  • 4. Deploying and maintenance of the code.
  • Environment: Datastage, Oracle Exadata, Snowflake, Hive, Autosys etc.

Data Engineer

Synechron Technologies Pvt. Ltd.
08.2021 - 04.2022
  • ESP (Enterprise staging platform) is an enterprise level datalake for TIAA, Data from all the LOB and products are reported and consolidated at the ESP and is further used for reporting and compliance purposes.
  • Responsibilities:
  • 1. Requirement analysis
  • 2. Building generic frame work to onboard new tables as a part of migration from the existing DataMarts in Teradata and oracle.
  • 3. Designing and building etl pipelines to bring data from various sources.
  • 4. Deploying and maintenance of the code.
  • Environment: Hive, sqoop, Spark, Scala, Oracle, CDP distribution for hadoop, shell scripting, Autosys etc.

Data Engineer

Synechron Technologies Pvt. Ltd.
06.2020 - 08.2021
  • The accelerator demonstrated the functionality of showing the ESG score at portfolio, It performs Portfolio Analysis against different ESG metrics and benchmarking using different ESG scores and index. It also improves the portfolio ESG score by proposing an “alternative portfolio” with minimal changes and considering client preferences. Finally, it provides support for extracting report overviews of the recommendation/investment.
  • The data was coming from sources like RavenPack, Refinitive and scrapping from Yahoo Finance.
  • Responsibilities:
  • Writing python code to get data from various data sources like RavenPack and Refinitive Apis and Scrape data from yahoo finance.
  • Creating
  • Creating managing the datalake and schemas for this data.
  • Environment: Python, Mysql, Hbase , Airflow, py-spark, Hive, Hue, Apache Atlas

Data Engineer

Synechron Technologies Pvt. Ltd.
06.2020 - 08.2021
  • Digital Compliance spawns regulatory site and downloads all the published notices and its rules. And based on user subscription for notices show them in dashboard. Also, provide chat bot kind of functionality for Q&A which can help you to define, design and execute technology implementation that drives business growth. Analyze long regulatory documents and get required information
  • Responsibilities:
  • Writing python Utilities to bing/download all the data from various regulatory sites
  • Ingesting the scraped data to data lake and applying pre-processing, labelling and Segmentation of data by topic mapping
  • Extracting and generating additional insights from the data scraped using NLP and machine learning algorithms
  • Managing and scheduling the daily jobs for data scraping and pre-processing.
  • Environment: Azure cloud, python, hbase, mongodb, Airflow Hadoop

Hadoop and Spark Developer

Synechron Technologies Pvt. Ltd.
11.2017 - 03.2020
  • Wells fargo has a variety of foreign exchange products which were traded and maintained in various Environment, Having a 360 deg view of all the customer and Products was not possible. All the reporting for these products were done from different SOR. FX Dataservices moved all the data for Trades to one single datalake so that it is a single and Primary point of reporting and Analytics. In Dataservices we maintained records for more than 14 different FX products and sis the primary level or reporting to the downstream and business users.
  • Responsibilities:
  • Analysed requirements for different products and their attribute
  • Designed the workflow for project , and creating new solution for the project
  • Writing Java, shell spark scripts for various ETL process, analytics and reporting
  • Had to coordinate with the onshore and the offshore team.
  • Building various validation and recon processes and tools.
  • Environment: Java, Hive, Hbase, MaprDB, Autosys, NDM, Spark, Shell Scripting, Solace Messaging

Hadoop and Java Developer

Synechron Technologies Pvt. Ltd.
02.2017 - 10.2017
  • This project was a product design for banking clients for Analysing NASDAQ and NYSE Data for all active stocks and predictive analytics on the trend for these stocks for better investment perspective for the clients The data was coming from static web pages in JSON format and had to be consumed by Kafka as Streaming source and stored on HDFS to be processed further by SSIS. News and feeds related to every stocks had to be collected and analysed to predict the trend and movements of the stocks
  • Responsibilities:
  • Analysed requirements for the project
  • Designed the workflow for project
  • Single handily wrote all the Java and Kafka code and implemented the project
  • Maintained the Hadoop Cluster
  • Environment: Core Java, Kafka, HDFS, Cron tab, Scala Spark

Hadoop Developer

Cognizant Technology Solution
10.2016 - 01.2017
  • This project was a POC project for retail data analysis in Spark using Scala for building various analytics on the retail data. The source for data was MySQL and the data was imported on HDFS and data transformations were done in pig and the analytics were performed on Spark using Scala
  • Responsibilities:
  • Analysed requirements for the project
  • Designed the workflow for project

Hadoop Developer

Sears Holdings India
09.2012 - 08.2016
  • Worked as a Hadoop developer for Mainframe migration project which involved in migration multiple retail business process from mainframe-based system to Hadoop ecosystem.
  • Designed and maintained Hive-based data warehouse with partitioning and bucketing along with optimizations
  • Integrated data from DB2 , Oracle and Teradata using Sqoop and Custom ETL workflows
  • Built UDFs in Java and Map-reduce code.
  • Carried out Hadoop Upgradation (CDH3 to CDH4)

Education

Bachelor of Engineering - Information Technology

R.T.M University

C-DAC - undefined

Infoway Technologies

Skills

  • Data pipeline creation
  • Data modeling and analysis skills
  • Proficient in scripting languages: Python, Java, Shell scripting
  • Proficient in Spark framework
  • Cloud Technology: AWS , Azure, Snowflake
  • Experience with HDFS, Hive, HBase, and Sqoop
  • SQL performance tuning
  • Data warehousing expertise
  • Proficient in multiple database technologies (Snowflake, Oracle, Hive, HBase, Cassandra, MySQL, Redshift, DynamoDb)
  • Proficient in scheduling automation: Control-M, Cron Tab, Airflow, Autosys

Certification

  • SnowPro Core certified, Snowflake - Jan'25 - Jan'27
  • Microsoft Certified: Azure Data Engineer Associate
    Credential: https://learn.microsoft.com/en-in/users/chetanpaunikar-7463/credentials/eca826611cf7aec8
  • AWS Certified Solutions Architect – Associate
    Issued: Jul 28, 2023
    Registration Number: 455595675
  • AWS Certified Solutions Architect – Professional
    Issued: Jan 20, 2025 – Expires: Jan 20, 2028
    Verification: AWS Certification Verification

Languages

English
Full Professional

Timeline

Data Engineer

Synechron Technologies
10.2023 - Current

Data Engineer

Synechron Technologies Pvt. Ltd.
04.2022 - Current

Data Engineer

Synechron Technologies Pvt. Ltd.
08.2021 - 04.2022

Data Engineer

Synechron Technologies Pvt. Ltd.
06.2020 - 08.2021

Data Engineer

Synechron Technologies Pvt. Ltd.
06.2020 - 08.2021

Hadoop and Spark Developer

Synechron Technologies Pvt. Ltd.
11.2017 - 03.2020

Hadoop and Java Developer

Synechron Technologies Pvt. Ltd.
02.2017 - 10.2017

Hadoop Developer

Cognizant Technology Solution
10.2016 - 01.2017

Hadoop Developer

Sears Holdings India
09.2012 - 08.2016

C-DAC - undefined

Infoway Technologies

Bachelor of Engineering - Information Technology

R.T.M University
Chetan Paunikar