Download as pdf or txt
Download as pdf or txt
You are on page 1of 4

Valluri Sunitha

Mobile: +91- 9640670104 Email: valluri.sunitha04@gmail.com

Professional Summary
Having Total 4.6 years of Professional experience in IT industry out of which 3 years in
Microsoft Azure Services.

Summary:

 Strong working knowledge on Azure Data bricks , Azure SQL, Azure Data
Factory.
 Hands-on experience in Azure Data factory and its Core Concept's like Datasets,
Pipelines and Activities, Scheduling and Execution.
 Excellent knowledge of ADF building components –Integration Runtime, Linked
Services, Data Sets, Pipelines, Activities.
 Experience in Managing and storing confidential credential in Azure Key vault.
 Good experience in Python code for data loads, extracts, in Azure Data bricks.
 Knowledge on Data Extraction from On-Premise Sources and Delta Extraction
methods from Source Systems to ADLS.
 Hands on experience in integrating Hive with Spark to perform HQL in spark.
 Excellent communication, interpersonal, analytical skills, and strong ability to
perform as part of team.
 Hands-on experience in python programming and Spark components like Spark-
core and Spark-SQL.
 Worked on creating RDD's, DF's for the required input data and performed the data
transformations.
 Experience on Spark core and creating RDD to perform aggregations and grouping
etc. in spark.
 Manage data recovery for Azure Data Factory Pipelines.
 Knowledge on Azure Data bricks, creation of Notebooks.
 Good team player with excellent communication and interpersonal skills.
 Play a key role in development team.

Professional Experience

Working as Software Engineer at Mouriya IT Solutions Private Limited from June 2018 – Till date.
Education Details

Completed Graduation from JNTU-H in 2015.


Technical Skills

Cloud : Microsoft Azure


ETL Tools : Azure Data Factory
Azure Services : Azure Data Bricks, ADLS, Azure Key Vault, Azure Gen2Lake
Azure Devops : Repos, Boards
Programming Language : Python , SQL
Operating System : Windows

PROJECT-1:

Project Title : Building Data Lake in Azure


Role : Azure Data Engineer
Client : CAQH
Environment : Azure Cloud
Components Used : Azure Data Factory V2, Azure SQL, DW, ADLS

Description :

CAQH was formed by a number of the nation’s largest health insurance companies with the goal of
creating a forum for healthcare industry stakeholders to discuss administrative burdens for
physicians, patients, and payers.Its mission is to accelerate the transformation of business
processes in healthcare through collaboration, innovation, and a commitment to ensuring value
across stakeholders, including healthcare providers, trade associations, and health plans.

Roles and Responsibilities:

 Coordinate with external teams to reduce data flow issues and unblock team members.
 Always actively participate in four ceremonies: Sprint planning meeting, Daily Scrum,
Sprint review meeting, and Sprint retrospective meeting.
 Working with Source team to extract the data and it will be loaded in the ADLS.
 Creating the linked service for source and target connectivity Based on the requirement.
 Creating the pipelines and datasets which are deployed in ADF non-restricted.
 Performing metadata insert scripts for the pipelines which is available in the logging
framework.
 Once it’s created pipelines and datasets will be triggered based on LOAD (HISTORY/DELTA)
operations.
 Based on source (big or small) data loaded files will be processed in AZURE DATABRICKS
by applying operations in spark SQL. Which will be deployed through AZURE DATA
FACTORY pipelines.
 Involved in deploying the solutions to QA, DEV and PROD.
 Involved in setting up the environments for QA, DEV and PROD using VSTS.
PROJECT-2:

Project Title : EDW to Azure Migration


Client : Hindustan Unilever Pvt., Ltd
Environment : Azure Cloud
Components Used : Azure Data Factory, Azure SQL, DW, ADLS, Azure Data Bricks

Description :

Unilever is a multinational corporation selling consumer goods including foods, beverages,


cleaning agents and personal care products. Unilever is a dual-listed company consisting of
Unilever NV in Rotterdam and Unilever PLC in London. Unilever owns more than 400 brands
including 11 "billion - dollar brands", which each achieve annual sales in excess of €1 billion. GDS
is the data science project where data resides on-premises and it is been pulled via ADF to ADLS
and ingested to Azure SQL Data ware house. From there it has been consumed by Azure Analysis
Cube then to Power BI Reports.

Roles and Responsibilities:

 Experience delivering quality in Line of Business applications.


 Passion for product quality, customer satisfaction and a proven track record for delivering
quality.
 Design and document database architecture.
 Actively involved in the design phase of the development, conducted detailed data
modeling for the database to be developed.
 Involved in Optimization and Tuning.
 Experience in migration of data from excel, flat file, and others using Azure ADF V2.
 Designed tabular cubes for the data in Azure SQL DW.
 Involved in migrating data from various countries data files to azure cloud using ADF V2.

PROJECT-3:

Project Title : Hadoop, Hive


Role : Hadoop developer
Client : Lowell
Environment : Hadoop, Hive

Description :

Lowell is debt management company, they purchase portfolios from banks and they will reach out
to customers for debt collections. As part of that Lowell team will create different plans for
collecting debts from customers, customers will make payments against plans which created by
Lowell team. Lowell also do prediction analysis for understanding the collection in next 3
years .What are the new accounts joined in this month and which customers completed their
debts in this month.
Roles and Responsibilities:

 Responsible to manage data coming from Oracle source.


 Responsible for building scalable distributed data solutions using Hadoop.
 Developed Spark scripts by using python shell commands as per the requirement.
 Used Spark API over Cloud era Hadoop YARN to perform analytics on data in Hive.
 Loaded the data into Spark RDD and do in memory data Computation to generate the
Output response.
 Experienced in handling large datasets using Partitions, Spark in Memory capabilities,
Broadcasts in Spark, Effective & efficient Joins, Transformations and other during ingestion
process itself.
 Involved in creating Hive tables and loading and analyzing data using hive queries.
 Developed Hive queries to process the data and generate the data cubes for visualizing.
 Implemented Partitioning, Bucketing in HIVE.
 Developed Python scripts using Data frames/SQL and RDD for Data Aggregation, queries
and writing data back into OLTP system through Sqoop.
 Experience in Job management using Fair scheduler and Developed job processing scripts
using Oozie workflow.

You might also like