Innovative and results-driven technology leader with extensive experience in project management, data management, and software development. Proven track record of delivering high-impact solutions in complex business environments, with a strong focus on automation, data security, and efficient system design. Proficient in leveraging a wide range of technologies to drive business transformation and operational excellence.
• Led the automation of critical document generation and data processing projects, significantly improving operational efficiency and process standardize without compromising on security or compliance, resulting in annual savings of over $3.1 million across multiple sites.
• Designed and set up an Enterprise Data Lake, supporting analytics, processing, storage, and reporting of voluminous, rapidly changing data, enhancing data accessibility and quality for business decision making.
• Maintained and ensured data integrity in a relational environment by implementing cleaning, transformation operations, and developing a Security Framework for fine-grained access control in AWS S3, using AWS Lambda and DynamoDB.
• Conducted architecture and implementation assessments for AWS services (S3, Redshift, DynamoDB, AWS Batch, AWS Glue), optimizing data movement between AWS Redshift and MySQL databases with AWS Glue, Spark, and Python, ensuring high-quality data provision.
• Automated numerous Amazon manual processes, saving more than 200K manual man hours, and developed QuickSight dashboards for system and operations departments, improving transparency and operational deliverables for 500+ staff members and decision makers across the US.
• Implemented AWS Step functions to automate Amazon tasks (e.g., data publishing to S3), and created Lambda functions with Boto3 for serverless code execution, demonstrating leadership in leveraging cloud technologies to drive significant cost savings and operational efficiencies
• Designed and implemented data strategies and tools for various data platforms, including relational, dimensional, and NoSQL databases.
• Experience with cloud computing, particularly in the context of AWS.
• Proficient in using PySpark and SparkSQL for data modeling, transformation, and cleaning in the Hadoop and Hive ecosystem.
• Knowledgeable in working with big data tools and technologies, such as Python, Spark, Hive, and Airflow.
• Skilled in building end-to-end ETL pipelines to fulfill customer requests for new data models and ingesting data from various sources using SQOOP.
• Experienced in data warehousing projects using tools like Talend, Informatica, and Pentaho, and proficient in developing reports and analytics using Tableau, Pentaho BA Reporting Tool, and other statistical tools.
• Designed and developed data flow/solution architecture with flexible schema to handle structured/ unstructured data for real time and batch processing application using Oracle / python.
• Designed relational/dimensional database (logical/physical data model) and ETL processes.
• Installed, configured, scheduled and monitored ETL process on Pentaho.
• Programmed views, stored procedures and functions.
• Did query performance and code optimization using AWR reports.
• Provided database support by coding utilities, responding to user questions and resolving problems.
• Created python scripts for Disk space monitoring on DB servers.
• Created daily reports for business users using python
• Performed tables tuning, created Hints, partitioning
• Developed Pl/Sql triggers, stored procedures, functions and packages from moving data from staging area to data mart.
• Developed and designed Oracle PL/SQL batch and real time processes for loading data from internal and external systems using SQL Loader and External tables.
• Developed complex SQL queries with sub-queries, analytical functions, and inline views.
• Designed and created entity relationship diagrams (ERD) for Oracle based applications and interfaces .
• Design, develop and maintain oracle database schemas, tables, standard views, materialized views, synonyms, unique indexes, non unique indexes, constraints, triggers, sequences, implicit cursors, explicit cursors, cursor for loops, reference cursors and other database objects.
• Created PL/SQL packages, stand alone procedures, and stand alone functions.
• Design, and develop, Oracle PL/SQL ETL processes for loading data from internal systems using SQL Loader and External tables
• Manage User Acceptance Testing of new applications and new versions of existing applications.
• Served as technical lead for the design and development of new enterprise web based applications as well as new versions of existing ones written in Oracle PL/SQL.
• Facilitate requirements gathering and design sessions.
Amazon EC2 Oracle Database Amazon S3 Amazon QuickSight Amazon Redshift AWS Glue Amazon Relational Database Service (RDS) Apache Spark AWS Lambda DynamoDB AWS Batch MySQL databases Python PySpark Amazon SQS Amazon kinesis Airflow Pentaho Hadoop Data Modelling Database Architecture Database Normalization