Summary
Overview
Work History
Education
Skills
Websites
Timeline
Generic

Sai Sandeep Bejjanki

Glen Ellyn,IL

Summary

Business data analyst with 5 years of experience in the creation of functional requirements, process flow analysis, and design, conducting functional and regression testing, coordinating user acceptance testing, content management analysis, detailed data analysis and proof of concept coordination. Experience in data analysis, data visualization using Tableau/ Microsoft Power BI, model building with Machine learning algorithms for prediction and forecasting using data (historical or time series with regression techniques), using statistical / econometric tools like SAS and R, data mining using Python, SQL, Hadoop, Spark and Hive. Strong experience in Business and Data Governance, Data Integration, MDM, NoSQL and Metadata Management Services and Configuration Management. Expertise and vast knowledge in Enterprise Data Warehousing including Data Modeling, Data Architecture, Data Integration (ETL/ELT) and Business Intelligence. Develop and implement data governance policies, procedures, and standards to ensure data quality, integrity, and security. Establish and maintain a data governance framework, including data classification, data ownership, and data stewardship. Experience performing multi-function management and business analysis for various projects. Lead data analytics projects from requirement stage, scope analysis, developing models, deployment and support. Developed automated scripts to do the migration using UNIX Shell scripting, Python, Oracle, T-SQL, Macros and Procedures. Expertise in transforming business requirements into analytical models, design algorithms, develop reporting solutions that scale across massive volumes of structured and unstructured data. Experience interpreting and analyzing large data sets for driving business solutions, building and evaluating analytical models, examining a range of disparate data sources, performing statistical analysis, creating reports/dashboards/graphs/storytelling to provide actionable insight for decision making. Skilled in Statistical analysis, Qualitative analysis, predictive modelling and forecasting, hypothesis testing, data mining, data cleaning, data quality assessment data integration and validation, data modeling, ETL and data visualization. Develop and deliver data governance training programs to educate employees on data governance principles and best practices. Monitor and enforce compliance with data governance policies and procedures. Proficient in MySQL (RDBMS, SQL Server, PL/SQL), Python (NumPy, pandas, scikit-learn), R programming (ddply, ggplot2, random forest), SAS EG, SAS/ BASE Core Java, Web scrapping, A/B Testing, Google analytics, Web analytics, marketing analytics. Highly proficient in writing, testing and implementing triggers, stored procedures, functions, packages and cursors using PL/SQL. Developed Python scripts to parse the Flat Files, CSV, XML, JSON Files and extract the data from various sources and load the data into the data warehouse. Experience in designing Business Intelligence Solutions with Microsoft SQL Server and using MS SQL Server Integration Services (SSIS), MS SQL Server Reporting Services (SSRS), SQL Server Analysis Services (SSAS). Expertise in Advanced MS-Excel, PowerPoint, spreadsheets (Pivot tables, Pivot Charts, Macros, VLOOKUP), Jupyter Notebook, Version Control Software (GitHub, Jira). Expertise in Master Data Management concepts, Methodologies and ability to apply this knowledge in building master Worked with AWS Cloud platform and its features which includes EC2, VPC, RDS, EBS, S3, CloudWatch, Cloud Trail, Cloud Formation and Autoscaling. Involved in System Integration Testing, (SIT), Regression Testing, Performance Testing and User Acceptance Testing (UAT).

Overview

6
6
years of professional experience

Work History

Business Data Analyst

Sripadha
05.2023 - 05.2024
    • Worked with BI Team in gathering the report requirements and Sqoop to export data into HDFS and Hive
    • Involved in all the phases of Analytics using R, Python and Jupyter Notebook
    • Analyzed existing internal data and external data, worked on entry errors, classification errors and defined criteria for missing values
    • Wrote Base SAS Programs for loading data from flat files to Teradata tables
    • Applied simple statistical procedures such as PROC MEANS, PROC FREQ for analyzing the data
    • Developed forms, reports, queries, macros, VBA code and tables to automate data importation and exportation to a system created in MS Access
    • Extract Business Glossary metadata from Informatica analyst and create linking between Business metadata and Technical metadata extracted from all the source systems within organization
    • ETL process to store and retrieve data from SQL Server Database Management, data warehouse and Troubleshoot test scripts, SQL queries, ETL jobs, data warehouse/ data mart/ data store models
    • Designing the SSAS Cubes, SSRS Reports and Ad hoc Querying facilities
    • Developed and implemented a comprehensive data governance framework, including data classification, ownership, and stewardship, resulting in a 30% improvement in data quality and security
    • Collaborated with IT teams to implement data governance tools and technologies, leading to a 20% increase in data governance automation and efficiency
    • Expert in dimensional data modeling using SSAS and Erwin, cube partitioning, optimization and creating aggregations
    • Creation of multiple, very complex reports in SSRS and PowerBI, that run on high volume of data, with low response time and which pulls data from SQL Server as well as Mongo DB
    • Created Tableau Dashboard for the top key performance indicators for the top management by connecting various data sources like Excel, Flat Files and SQL Database
    • Used python packages (Pandas, NumPy, Scikit- Learn) for feature engineering and machine learning models
    • Used machine learning algorithms (Logistic Regression as Baseline Model) to analyze the user demographics and trends
    • Python scripts were written to read CSV, JSON and parquet files from S3 buckets and load them into AWS S3, Dynamo DB and Snowflake
    • Used AWS EMR to transform and move large amounts of data into and out of other AWS data stores and databases, such as Amazon Simple Storage Service (S3) and Amazon Dynamo DB
    • Designed and developed security framework to provide fine-grained access to objects in AWS S3 using AWS Lambda and Dynamo DB
    • Performed End-to-End architecture and implementation of various AWS services like Amazon EMR, EC2, Redshift and S3
    • Load data into Amazon Redshift and use AWS Cloud Watch to collect and monitor AWS RDS instances
    • Implemented AWS Step Functions to automate and orchestrate the Amazon Sage Maker-related tasks such as publishing data to S3, training the ML model, and deploying it for prediction.
    • Generated ad-hoc reports to evaluate specific business requirements.

Business Data Analyst

Kaseya
01.2023 - 04.2023
  • Extracted data from multiple data sources, wrote python scripts to scrape the data and images
  • Performed Data Profiling, Data Pipelining and Data Mining, validating and analyzing data (Exploratory analysis/ Statistical analysis) and generated reports
  • Creation of multiple SSIS packages to import data from the legacy (mainframe) system, Oracle, and Mongo DB to target SQL Server DB for report consumption and other use
  • Designed and developed an end-to-end data warehousing and OLAP solution using SSIS, SSAS, SSRS and SQL Server
  • ETL process to clean and load large data extracted from CSV/ JSON files to the MYSQL server
  • Created reports and dashboards by connecting multiple data sources to help track user trends, behavior, engagement and performance
  • Extract and load the metadata from disparate source systems into MM warehouse using informatica metadata manager packaged models as well as custom models and create data lineage among these systems
  • Extensively used SAS procedures like IMPORT, EXPORT, SORT, FREQ, MEANS, FORMAT, APPEND, UNIVARIATE, DATASETS and REPORT
  • Experience in providing SQL to tollgate the data quality check to ETL to test the data inserted by the Java rules engine into staging table as per the requirement and logic to automate the process
  • Worked on AWS hosted Data bricks environment and used Spark structured streaming to consume the data from Kafka and perform merge operations on delta lake tables
  • Developed Spark/ Scala, Python for regular expression(regex) project in the Hadoop/ Hive environment for big data resources
  • Assess and implement advanced web analytics tracking, monitor KPI's using Google analytics, create google analytics reports and dashboards
  • Developed Tableau Data Visualization using Pareto's, Combo charts, Heat Maps, Box and Whisker plots, Scatter plots, Geographic Map, Cross tabs and Histograms.

Data Analyst

Arche Group INC
09.2022 - 12.2022
  • Created database and database objects such as tables, indexes, constraints, stored procedures, views, and triggers, and maintained relationships between tables
  • Transformed the data using AWS Glue dynamic frames with PySpark; cataloged the transformed data using Crawlers and scheduled the job and crawler using workflow feature
  • Allotted permissions, policies, and roles to users and groups using AWS Identity and Access Management (IAM)
  • Analyze, improve and manage ETL architecture and provide strategy related to data cleansing, data quality, and data consolidation
  • Executed the program by using Python API written in Python to support Apache Spark or PySpark
  • Experience in using Spark SQL to handle structured data from Hive in the AWS EMR Platform
  • Implemented Job Monitoring, health checks, and job streaming of daily user production data volumes in the HDFS system using Scala, Python, Spark SQL, HBase, and Phoenix
  • Create/Modify shell scripts for scheduling various data cleansing scripts and ETL loading processes
  • Developed Informatica mappings to load data files sourced from Project plan, requirement analysis, low-level design specification, development, preparation of test plans and test specs for QA/UAT, Preparation of release notes
  • Developed Data Mapping, Data Governance, Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP
  • Created Tableau scorecards, dashboards using stack bars, bar graphs, scattered plots, geographical maps, Gantt charts using show me functionality
  • Propose ideas to improve the process efficiency and quality through weekly and monthly status reports/meetings.

Data Analyst

Vaihuti Infotech Pvt. Ltd
06.2018 - 06.2021
  • Developed and modified Sybase packages, wrote and optimized in-application SQL, PL/SQL statements
  • Created DB2 database objects such as schema, tables, indexes, constraints, functions, stored procedures, views, and triggers, and maintained relationships between tables
  • Used Informatica Data Validation to create ETL testing situation and conducted Data Validation before and after migration
  • Performed daily integration and ETL tasks by extracting, transforming and loading data to and from different RDBMS
  • Used informatica to create repeatable tests that validate PowerCenter mapping data
  • Worked actively in ETL process and writing complex SQL queries for querying against different databases (SQL Server and Netezza) for data verification process
  • Created complex SQL queries and scripts to extract and aggregate data to validate the accuracy of the data
  • Used advanced Excel features (Pivot Tables, Index, VLOOKUP, Reporting) to analyze the data and performed end to end data validation in the Historical Migration process from SQL Server to Netezza
  • Prepare high level analysis reports with Excel and Tableau
  • Tuned and optimized the performance of queries with modification in DB2 queries, T-SQL queries, normalized tables, established joins, and created indexes wherever necessary
  • Worked on Normalization and De-Normalization concepts and design methodologies like Ralph Kimball and Bill Inmon's Data warehouse methodology.

Education

Master's - Business Analytics

Texas A&M University-Commerce
Commerce, TX
08.2022

Bachelor's - Computer Science

Acharya Nagarjuna University
Guntur, India
04.2018

Skills

  • Machine learning
  • Advanced Excel
  • Business intelligence
  • Natural language processing
  • Data warehousing
  • ETL development
  • Big data analytics
  • Report generation
  • Predictive modeling
  • Cluster analysis
  • SQL programming
  • Hadoop ecosystem
  • Data cleaning
  • Python programming
  • Data quality assurance
  • Tableau software
  • R programming
  • NoSQL databases
  • Apache Spark
  • Decision trees
  • Data integration
  • Dashboard creation
  • SAS programming
  • Time series analysis
  • Data mining
  • Data analysis
  • Requirements gathering
  • Cost-benefit analysis
  • Stakeholder management
  • Workflow Analysis
  • Operations analysis
  • Database management
  • Pivot tables
  • Business process improvement
  • HTML and CSS
  • Microsoft Office Suite
  • Business planning
  • Revenue development
  • IIBA methodologies
  • CRM systems
  • Quality assurance
  • KPI tracking
  • SAP
  • Data modeling
  • Data pipeline design
  • SQL expertise
  • Data governance
  • Data security
  • Real-time analytics
  • Metadata management
  • Data pipeline control
  • SQL and databases

Timeline

Business Data Analyst

Sripadha
05.2023 - 05.2024

Business Data Analyst

Kaseya
01.2023 - 04.2023

Data Analyst

Arche Group INC
09.2022 - 12.2022

Data Analyst

Vaihuti Infotech Pvt. Ltd
06.2018 - 06.2021

Bachelor's - Computer Science

Acharya Nagarjuna University

Master's - Business Analytics

Texas A&M University-Commerce
Sai Sandeep Bejjanki