Cloud Based Developer - SantoshKedar (4y - 0m)

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 3

Santosh Kedar

Big Data Developer


: sankedar15@gmail.com : Santosh Kedar | LinkedIn
: (+91) 9028286701 : Aurangabad, Maharashtra OBJECTI

Career Objective:

To be associate with a forward-thinking organization offering a challenging and rewarding career


opportunity and to leverage my strong Data Engineering skills to contribute to a dynamic, data-
driven company. I am a result-oriented Data Engineer/Developer with 4 years of experience in Big
Data technology, specializing in designing, developing, ETL tools and maintaining data pipelines
for organizations

Professional Summary:

Dedicated Big Data Engineering professional with 4 yrs. of experience and proficient in Spark,
PySpark, Snowflake, AWS services, Python, and SQL with a strong focus on processing large data
volumes. Skilled in building optimized big data pipelines with AWS Step Functions and processing
structured, semi-structured and unstructured data using AWS Glue with various input data formats,
such as CSV and JSON. Hands-on experience with AWS services, including AWS S3, Glue,
Lambda, CloudWatch, SQS and IAM roles. Also previously worked as Business Process Lead in
SAP deployment project (Mfg. Module-Healthcare) and Mfg. Ops. Dept.

Professional Skills:

➢ Programming Languages: Core Python, SQL


➢ Big Data Technology: Apache Spark, PySpark
➢ Cloud Services: AWS (S3, EC2, RDS, EMR, IAM, CloudWatch, Glue, Lambda, Athena)
➢ Database Management: SQL, MySQL
➢ Data Warehousing: Snowflake
➢ Version Control: Git, GitHub
➢ ETL Tools: Apache Nifi
➢ IaC tool: Terraform
➢ Proficiency in CI/CD practices
➢ Agile Methodologies
➢ JIRA technology for efficient project management

Certification:

Databricks Certified Associate Developer for Apache Spark 3.0


Professional Experience:

Organization : Grace Infosoft Private Ltd. Oct.2019 – Present


Nagpur, Maharashtra

Job Summary:

➢ Designed and implemented ETL processes to extract, transform, and load data from diverse
sources into a centralized data warehouse.
➢ Automated data ingestion and transformation processes, reducing manual data handling, leading
to improved data quality and operational efficiency.
➢ Assisted in the development of ETL pipelines and data warehousing solutions, gaining hands-on
experience with tools and technology such as Python, SQL, PySpark, Terraform, SnowFlake and
AWS Glue, Lambda, S3 buckets, CloudWatch, SQS, IAM roles.
➢ Supported the team in troubleshooting and resolving data-related issues and contributed to on-
call support as needed.

Project: "Strategic Insights: Data-Driven Trends for Business Excellence"


Domain : Retail Data Analytics

Project Summary:

➢ The primary objective of this project is to create a seamless and efficient workflow where data
is continuously updated in the Snowflake data warehouse, eliminating the need for manual
interventions and ensuring the availability of up-to-date data for analysis and reporting.
➢ Project focuses on establishing a streamlined data pipeline from raw JSON data that stored
finally in an AWS S3 bucket.
➢ It utilizes an AWS Glue ETL job written in PySpark to transform the data and store it in a
designated S3 landing bucket.
➢ The transformed data is seamlessly integrated with Snowflake, a cloud-based data warehousing
platform, to ensure high performance and scalability.
➢ To maintain real-time data synchronization, an SQS (Simple Queue Service) queue is set up to
receive notifications when new data files arrive.
➢ These SQS notifications trigger Snowpipe, a native Snowflake feature, to automate the data
loading process into the Snowflake table.
➢ Utilize Git and GitHub for version control to manage and collaborate on the codebase
effectively.

Project : "Cinephile's Delight: Leveraging TMDB API for Media Applications"


Domain : Entertainment and media

Project Summary:

➢ Establishing an end-to-end data pipeline for movie and TV show data retrieval and processing
from The Movie Database (TMDB) API.
➢ Developed Python script and pandas for efficient data extraction, data processing and
transformation.
➢ Deployment on Amazon Web Services (AWS) infrastructure using Terraform for infrastructure
as code.
➢ Utilizing a serverless architecture on AWS lambda function for streamlining infrastructure
management.
➢ Key technologies used include Python and Pandas for data processing, AWS Lambda for
serverless computing, AWS S3 for data storage, AWS CloudWatch for monitoring and
Terraform for infrastructure management.

Professional Experience:

Organization : Pfizer Ltd. Designation : Assist. Manager


Aurangabad, Maharashtra.

Duration : 3.5 Yrs. (SAP/ERP domain as end user) and 16 Yrs. in Healthcare Mfg. Ops.
Project : Pfizer SAP-ERP deployment project with IBM team for implementation of SAP
at the site location including following phases.

➢ Plan & Analyze phase ➢ Integration Testing


➢ Solution Confirmation phase ➢ Best & Final Load
➢ Mock Data Load ➢ Cut Over & Ramp- Down Phase
➢ Integration Data Load ➢ Go-Live & Ramp up Phase
➢ Business Verification Testing ➢ Hyper-care Support

Project Summary:

➢ Collected business requirements data across multiple Manufacturing plants, specifically


related to the SAP PP module.
➢ Conducted a successful Fit-Gap analysis, leading to the identification of enhancement
opportunities.
➢ Provided essential support to various stakeholders, including Manufacturing Leads, the Core
Solution Team, Centre of Excellence, Business Process Leads, and Super Users.
➢ Collaborated closely with cross-functional business teams (APO, Finance, WM) to optimize
integration points with the PP module.
➢ Took a central role in formulating key design aspects, including Recipe, BOM, Production
Version, Phase Relationships, and Handling Unit functionality, contributing significantly to
the project's success.

Educational Qualification:

Bachelor degree in Engineering Graduate July-1999


Amravati University, Maharashtra

You might also like