Download as pdf or txt
Download as pdf or txt
You are on page 1of 3

Rohit Ranjan

Address: Gaya, Bihar


Mobile: +91 9163044154
Email: rojuieebe@gmail.com
Current Employer : PricewaterhouseCoopers(PWC)

Experience Overview:
• 4 years of experience in Data analytics & Engineering.
• Exposure of multiple domains like Pharma and Hospitality sector in designing and development.
• Exposure to the ETL Pipelines, Data structures, Data management and processing systems using
Spark, Azure Data factory (ADF), PySpark, SNAPLOGIC.
• Expertise in SQL and data modelling.
• Experienced in working on Google Cloud Platform (GCP), Azure cloud and data Analysis using
Dremio.
• Possessing excellent creative, analytical, verbal, and written communication skills.
• Committed, result oriented, hardworking, quest to learn new technologies.
• Ability to meet demanding schedules and various work/process environments.
• Have worked in requirement gathering, development, Implementation, enhancement & production
support.

Technical Skills:
Relational Database MySQL
ETL Tool Snaplogic , Azure Data Factory (ADF)
Programming Python , SQL, PySpark
Technologies Spark - SparkSQL and PySpark ,Hadoop – HDFS
,Hive, Google Cloud Platform (GCP) , Big Query,
Airflow Dag, Google Cloud Function,
Dataproc,Databricks

Certifications: Training/Courses:
• Azure Fundamentals developer by Azure • Machine Learning with Python
• Cloudera Spark Developer by Cloudera • Data Science Fundamentals with
• Google Associate Cloud Engineer Python and SQL
• Snaplogic Developer

Professional Experience Summary:


Current Employer PwC India
Title Associate
Role Developer
Date of Employment 24, July 2019

Education Summary:
• Bachelor of Engineering in Electronics & Telecommunication Engineering from Jadavpur University,
Kolkata (2015 - 2019)
• Higher Secondary School Certificate (CBSE) from DAV Gaya in 2014.
• Senior Secondary School Certificate (CBSE) from DAV Gaya in 2012.

Experience Profile – Key Projects

Project: Generic Market Analysis system for the US Based Pharmaceutical Industry

Development of a system which gives total understanding of the Pharma market including
competitors market share and standing. It works by transforming extracted raw data which we get from
various US based Pharma regulatory bodies in different formats into a common data model according
business rules which comprises a series of entity related data tables that together model transactional
and sales data of various HCP and HCO.
As Spark developer, developed spark scripts to automate ETL process, making it easier to wrangle data
and reducing time by 40%.

Roles and Responsibilities:


• Performed various performance tuning logic to load data efficiently on target and reduce loading
time.
• Performed the data ingestion, CDC (Change data capture) and transformations using PySpark
scripts, which also included performing aggregations on the data as part of the data preparation
activities.
• Developed various reusable templates and custom defined scripts to meet complex business
requirements to increase re usability.
Technologies:
Spark - SparkSQL and PySpark, Dremio,SQL,Wasbs.

Project: Centralized Data lake Development on google cloud for an Indian pharma company
Development of centralized data lake, so that data can be used by various downstream system from a
centralized location , Here I got the opportunity to understand the business perspective of pharma
sector along with detailed understanding google cloud platform .
Roles and Responsibilities:
• Get the data from different sources where data had been receding in different formats (like xml,
csv, excel, json) using Various ETL tools , Python Scripts ,DAG and loaded to the big query tables
after doing all the required transformations
• Performed the data ingestion and transformations using google functionality which also include
performing aggregations on the data as part of the data preparation activities.
• Developed various Airflow DAG script using Python to make the pipelines more efficient while
processing larger files.
• Developed the pipelines that loads data to various downstream system like Salesforce, Mendix,
Magento
• Preformed Web Scrapping from website using python scripts.
Technologies:
Google Cloud Platform (GCP), Google cloud function, Virtual Machines , DataProc , Airflow DAG ,Big
Query

Project: Source and Target agnostic common data model to accelerate migration of data from various
ERP systems.
As ETL developer was involved in the development of a system works by transforming extracted raw
ERP data into a common data model which comprises a series of entity related data tables that together
model transactional & accounting data.
As sub part to this, was also involved in the development of system which would do the finance liquidity
analysis for the client by using common data model.
As Snaplogic developer was involved in Data Ingestion and Data Preparation in this assignment for
Oracle Track, solely responsible for ETL development using Snaplogic and SQL and was able to be the
reliable point of contact.
Roles and Responsibilities:
• Functional understanding of the processes across different teams.
• Functional requirement gathering across various functional areas.
• Developing data flow as part of ETL standards.
• Implementing Business Logic as per client requirement using Snaplogic.
• Working with Business Object team to deliver the client finance requirement.
Technologies:
Snaplogic, SQL, Oracle EBSR12,Wasb,Oracle

Personal Information:
Date of Birth 2-April-1997
Gender Male
Marital Status Single
Languages Known English and Hindi

You might also like