Professional Documents
Culture Documents
Ram Madhav Resume
Ram Madhav Resume
PROFESSIONAL SUMMARY:
To secure a challenging role as a Data Engineer, leveraging my 9+ years of software industry experience, with a focus
on Azure cloud services and Big Data technologies like Spark, MapReduce, Hive, Yarn, and HDFS, using programming
languages such as Scala and Python. With my 4 years of experience in Data Warehouse, I possess a deep understanding
of ETL processes, data modeling, and data warehousing. I am committed to delivering efficient and scalable data
solutions that drive business growth and support strategic decision-making.
✔ In depth knowledge on Big Data with Hadoop, Hive, Map Reduce, Spark, Spark Core, Spark SQL, and Data
Frames/Data Sets/RDD API
✔ Developed applications using Spark with PySpark, Scala for data processing
✔ Experience in writing Spark Jobs for data cleansing and transformations
✔ Good knowledge on Spark architecture and real-time streaming using Spark
✔ Experience writing in house UNIX shell scripts for Hadoop&Big Data Development
✔ Have good experience working with Azure BLOB and Data lake storage and loading data into Azure SQL Synapse
analytics (DW).
✔ Hands on experience on Python programming PySpark implementation azure data farctory building data
pipelines infrastructure to support deployments for Machine Learning models.
✔ Proficient writing complex spark (pyspark) User defined functions (UDFs), Spark SQL and HiveQL.
✔ Experience working on Azure Services like Data Lake, Data Lake Analytics, SQL Database, Synapse, Data Bricks,
Data factory, Logic Apps, Functional App and EventHub services.
✔ Experience in developing data pipeline using Hive and Sqoop, to extract the data from weblogs and store in
HDFS and developing HiveQL for data analytics
✔ Extensively dealt with Spark Streaming and Apache Kafka to fetch live stream data
✔ Experience in converting Hive/SQL queries into Spark transformations using Java and experience in ETL
development using Kafka, and Sqoop
✔ Developed Spark scripts by using Scala shell commands as per the requirement
✔ Good experience in writing Sqoop queries for transferring bulk data between ApacheHadoop and structured
data stores
✔ Substantial experience in writing Map Reduce jobs in Java.
✔ Experience in Developing Spark applications using PySpark, and Spark-SQL in Databricks for data extraction,
transformation, and aggregation from multiple file formats(structured/unstructured) for analyzing and
transforming the data to uncover insights into the customer usage patterns.
✔ Experience in Data Warehousing, Data Mart, Data Wrangling using Azure Synapse Analytics
✔ Experience in understanding business requirements for analysis, database design&development of applications
✔ Worked with Kafka tools like Kafka migration, Mirror maker and Consumer offset checker
✔ Experience with realtime data ingestion using Kafka.
✔ Experience with CI/CDpipelines with Jenkins, Bitbucket, GitHub etc.
✔ Strong expertise in troubleshooting and performance fine-tuning Spark, and Hive applications
✔ Hands on experience in developing SPARK applications using Spark tools like RDD transformations, Spark core,
SparkStreaming and SparkSQL
✔ Extensive experience in developing applications that perform DataProcessing tasks using Teradata, Oracle, SQL
Server and MySQL database
✔ Worked on data warehousing and ETL tools like Informatica and PowerBI,
✔ Acquaintance with Agile and Waterfall methodologies. Responsible for handling several clients facing meetings
with great communication skills
EDUCATION
TECHNICAL SKILLS
✔ Big Data: Hadoop 3.0, Spark 2.3, Hive 2.3, Cassandra 3.11, MongoDB 3.6, MapReduce, Sqoop.
✔ Programming Languages: Python, Scala, SQL, Pyspark
✔ Big Data Technologies: Spark, Hadoop, HDFS, Hive, Yarn
✔ Databases: PostgreSQL, MySQL, Oracle, MongoDB, DynamoDB
✔ Other Tools: Eclipse, PyCharm, GitHub, Jira
✔ Cloud: Azure data Factory, Azure Data Bricks, Logic Apps, Functional App, Synopses, EventHub, Azure DevOps.
✔ Methodologies: Agile, Waterfall Model.
WORK EXPERIENCE
Environment: Azure, Hadoop, HDFS, Yarn, MapReduce, Hive, Sqoop, Oozie, Kafka, SparkSQL, Spark Streaming, Eclipse,
Informatica, Oracle, CI/CD, PL/SQL UNIX Shell Scripting, Cloudera.
Environment: Spark SQL, HDFS, Hive, Pig, Apache Sqoop, Java (JDK SE 6, 7), Scala, Shell scripting, Linux, MySQL Oracle
Enterprise DB, IntelliJ, CI/CD, Oracle, Subversion, and Agile Methodologies.
Environment: Spark, Azue SQL, Python, HDFS, Hive, Sqoop, Scala, Kafka, Shell scripting, Linux, Eclipse, Git, Oozie,
Informatica, Agile Methodology.