Summary
Overview
Work History
Education
Skills
Accomplishments
Personal Information
Timeline
Generic
Akshay Bodke

Akshay Bodke

Data Engineer @ IBM India
Pune

Summary

Experienced Data Engineer with over 4+ Years in the field, currently making significant contributions at IBM. Adept at crafting and implementing sophisticated data pipelines, Specialize in translating intricate business requirements into scalable and efficient solutions. Proficiency extends to leading and inspiring teams in Snowflake and ETL DBT and Azure Cloud, ensuring the seamless integration of data for optimal outcomes. With a focus on optimizing Data Pipelines and upholding accuracy and integrity, Strategic approach to Data Pipelines has consistently driven organizational efficiency and long-term objectives. Actively seeking a challenging role where Expertise in data engineering, coupled with leadership in Snowflake and ETL DBT and Azure cloud, can be applied to make substantial contributions.

Overview

4
4
years of professional experience
4
4
years of post-secondary education

Work History

Data Engineer

IBM India
08.2021 - Current
  • Demonstrated expertise in the development and deployment of both incremental and full-load data ingestion pipelines, leveraging the robust capabilities of Confluent Kafka and Snowflake.
  • Successfully conducted proof-of-concept projects utilizing Fivetran and HVR technologies to explore and evaluate options for the enhancement of future data pipeline implementations.
  • Successfully executed the migration of petabyte-scale historical datasets from an on-premises Hadoop environment to the Snowflake Data Warehouse, employing Azure ADLS Gen2 Blob storage as an intermediary to ensure data integrity and transfer efficiency.
  • Executed the establishment of Azure Data Factory and Azure Databricks Python & Pyspark pipelines, specifically designed for robust balance and control mechanisms, to bolster data quality assurance across the processing framework.
  • Orchestrated the development of data pipelines using DBT (Data Build Tool) with Snowflake, adept at handling both structured and semi-structured data transformations. Possess hands-on experience in flattening complex XML and JSON datasets to facilitate efficient data transformation.
  • Engineered a comprehensive project by skillfully integrating multiple models within DBT to transform data seamlessly through to the presentation layer, aligning with specific business requirements and objectives.
  • Contributed to the configuration and maintenance of Splunk Cloud queries for efficient data logging, and implementation of a sophisticated alert system to enhance monitoring and incident response capabilities.
  • Collaborated with various teams in performing rigorous data load testing to ensure system robustness and data integrity across multiple platforms.
  • Possess extensive experience in handling various data source types, including Flat-Files, Azure Service Bus, JDBC connectors, SQL SERVER Change Data Capture (CDC), XML, and JSON, among others.
  • Provided post-production support for operational issues as a Level 3 Developer, taking part in a rotational shift schedule to ensure continuous technical assistance and swift resolution of complex problems.

Associate Business Analyst

Infosys Limited
06.2021 - 08.2021
  • Designed and implemented an automated testing pipeline using Python to verify the authenticity Dynamic Nested JSON data, focusing on the validation of pre-defined field keys, values, and data types to ensure data quality and adherence to schemas.

Data Engineer (Graduate Apprentice Trainee)

Rishabh Instruments Pvt Ltd
06.2020 - 06.2021
  • Crafted and executed a real-time data pipeline capable of processing semi-structured data by seamlessly integrating 100 million raw records daily, sourced from IoT devices, utilizing Apache Kafka for data ingestion and Apache Spark for data processing, ultimately storing the refined data in Cassandra for optimized access and analysis.
  • Developed a customer-centric event processing alert system capable of handling thousands of user interactions, employing a robust tech stack including Apache Kafka for messaging, Spark for stream processing, MQTT for lightweight messaging, WSO2 for integration solutions, and K-SQL for processing Kafka streams.
  • Enhanced the performance of current ETL processes and optimized Cassandra queries pertaining to CRM summary data by strategically scheduling and managing CRON job triggers Apache Spark Jobs to streamline workflow and data transformation procedures.
  • Deployed a data pipeline on an AWS EC2 server to accommodate new architectural changes, utilizing K-SQL (Kafka SQL) for stream processing. This infrastructure was efficiently scaled to support an upsurge from zero to 10,000 daily active node users.

Data Science Intern (Software Trainee)

Epic Technology Research Pvt Ltd
12.2019 - 05.2020
  • Streamlined disaster management logistics for the State Government of Florida by engineering automated ETL (Extract, Transform, Load) workflows using Python. This enhancement optimized the allocation of necessary resources during emergencies, significantly reducing manual efforts during critical periods.
  • Developed a Machine Learning-based Time-Series Model to predict road pavement wear and tear, achieving a reduction in manual error and lowering maintenance-related financial expenditures by 7% for the client. This innovation has ensured a more accurate allocation of resources and proactive infrastructure maintenance.
  • Designed and developed a prototype chat-bot utilizing the RASA framework to elevate the user experience for the company's proprietary software product I-3. This initiative demonstrated the potential to enhance customer interaction and streamline in-house support.
  • Engaged in several proof of concept projects to discover client solutions utilizing a variety of Machine Learning algorithms. Leveraged extensive Python libraries including Pandas, NumPy, Seaborn, Matplotlib, and Scikit-learn to explore and implement data-driven strategies.

Education

PG-Diploma - Big Data Analytics

C-DAC
Pune, India
02.2019 - 08.2019

Bachelor of Engineering - Electrical Engineering

K K Wagh College of Engineering
Nashik,India
09.2011 - 07.2015

Skills

Data Warehousing: Snowflake

Accomplishments


  • Results-driven Data Engineer proficient in Snowflake & DBT, contributing to seamless ETL processes and optimizing data structures for efficient data processing within the banking sector.
  • Azure Data Engineer with expertise in designing and implementing intricate ETL workflows, emphasizing the importance of robust data pipeline for effective data transformation.
  • Proven Data Implementation relatively focused on Snowflake, advanced clustering features for optimized storage and retrieval of large-scale datasets in insurance scenarios.
  • Proficient in Snowflake & DBT, applied lateral flattening methods within data models to handle complex nested structures, ensuring streamlined query performance for banking data analytics.
  • Dynamic Spark Developer with a strong emphasis on Azure Databricks & Azure Data Factory .
  • Tableau & Power-BI Development, creating visually compelling dashboards that derive insights from meticulously designed data structures.
  • Kafka Developer with expertise in implementation of confluent Kafka connects scripts, producer & consumers, K-SQL.





Personal Information

Date Of Birth: 08-Oct-1992

Language Known: English, Hindi, Marathi

Interests: Modern Data Stack, Artificial Intelligence



Timeline

Data Engineer

IBM India
08.2021 - Current

Associate Business Analyst

Infosys Limited
06.2021 - 08.2021

Data Engineer (Graduate Apprentice Trainee)

Rishabh Instruments Pvt Ltd
06.2020 - 06.2021

Data Science Intern (Software Trainee)

Epic Technology Research Pvt Ltd
12.2019 - 05.2020

PG-Diploma - Big Data Analytics

C-DAC
02.2019 - 08.2019

Bachelor of Engineering - Electrical Engineering

K K Wagh College of Engineering
09.2011 - 07.2015
Akshay BodkeData Engineer @ IBM India