Naukri Vijay (3y 0m)

You might also like

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 2

Vijay K

+91-9941119812
As a Cloud Technician, aim to apply my expertise in designing and implementing robust data
architectures, optimizing ETL processes, and managing data infrastructure. I strive to enhance vijaykumar06899@gmail.
the seamless flow, storage, and analysis of data within an organization, utilizing skills in com
technologies like Spark, Hadoop, AWS, and various programming languages Bengaluru

PROFILE SUMMARY CORE COMPETENCIES


 Skilled in technologies, such as Spark, Hadoop framework, Snowflake, AWS
Data Architecture Design
GCP, Pyspark, Autosys, Splunk, ETL, Data warehousing, and Spark
Streaming.
Data Pipeline Development
 Proficient in real-time data analysis employing Spark Streaming and Kafka,
adeptly extracting live feeds through Kafka, and converting them seamlessly into
Database Management
RDD for subsequent processing and analysis in structured data frames.
 Substantial experience in developing Spark applications using Python and Scala.
Data Integration
 Designing and implementing scalable and efficient data architectures that meet the
organization's business needs Data Modeling
 Creating and maintaining robust ETL (Extract, Transform, Load) processes
and data pipelines to move and transform data from source systems to data Quality Assurance
warehouses or other storage solutions to ensure data quality, consistency, and
availability. Cross-Functional
 Integrating data from different sources within the organization, ensuring that
data is accurate, consistent, and accessible for analysis. Collaboration
 Developing and maintaining data models that support business
requirements, ensuring data structures align with the organization's goals and Monitoring and Optimization
facilitate efficient data retrieval.
 Implementing data quality and validation processes to ensure accuracy and Documentation
reliability of the data; developing and enforcing data quality standards.
 Working closely with data scientists, analysts, and other stakeholders to Technology Evaluation
understand their data needs and providing the necessary infrastructure and tools
for analysis.
 Monitoring system performance, identifying bottlenecks, and optimizing data
processes for efficiency.

CERTIFICATIONS
 Certified Spark Professional
EDUCATION  Certified Devops
2020: Bachelor of Science (Computer Science)  Certified Data Management
Madras University, Chennai with 75%  certified Big Data Developer
 Certified Python developer
 Certified Snowflake Developer
PROJECTS
Neural Machine Translation:
Machine translation using sequence-to-sequence architecture

Covid Forecasting:
Predict covid effect and its increase and decrease rate, each state-wise statistics and
prediction to take precaution

IT SKILLS
 Languages: Python, Java, Scala, SQL, Hive, Spark, Pyspark

 Libraries: NumPy, Sci-Kit, Pandas, STL, Apache Kafka, Apache Spark

 Skills: DBMS, Linux, Spark Streaming, Hadoop, Data Analytics, ETL, Kafka
Streams, Microservices, Quantexa

 Software and Developer Tools: GIT, Datalab, Snowflake, Teradata, MYSQL


workbench Tableau, Pycharm, Intellij, MS Excel, MS Powerpoint, MS Word,
BBedit.

 Cloud Technology: AZURE- ADF, Databrick; AWS - Snowflake, S3, AWS, GCP
Glue
WORK EXPERIENCE
Since April’21 with TCS

Growth Path:

Client: Apple
Sep’22 – Present as Specialist
Role:
 Engaged in active dialogues with the business team to collaboratively devise the architecture and assess the feasibility of the solution.
 Formulated end-to-end ETL pipelines encompassing diverse source and destination combinations such as Snowflake, Teradata, HDFS,
Kafka, and AWS S3. Generated and supervised production jobs, addressing any failed jobs and conducting error analyses on platforms like
Splunk, Airflow, Autosys, and Spark UI.
 Enhanced Spark job performance in AWS through meticulous optimization and fine-tuning of Spark applications, including memory
adjustments based on specific requirements.
 Utilized Scala & Hive queries within Spark ecosystem, developing proof-of-concepts with Spark SQL to advance data processing capabilities.
 Conducted data analysis in Hive, strategically creating partitioned and bucketed tables to maintain operational efficiency.
 Responded to inquiries from agents and business users by crafting Teradata, Snowflake, and Datalake queries, facilitating prompt and
accurate information retrieval.
 Implemented scripts for Datalab and Snowflake tables, leveraging Python for data delay monitoring.
 Took the lead in designing and implementing a robust data engineering pipeline, optimizing SQL queries, creating efficient data models,
and implementing orchestration workflows for enhanced overall efficiency.

Client:
Jul’21 Apple
– Sep’22 as System Engineer
Role:
 Crafted an interactive data analysis dashboard through Streamlit, providing stakeholders with illuminating visualizations to facilitate well-
informed decision-making.
 Streamlined system efficiency by transitioning code from Scala to Python, and instituted Python scripts for backfill, daily operations, and
automated job triggering.
 Conducted thorough data analysis and manipulation using Pandas and Numpy, extracting valuable insights to inform decision-making.
 Managed Snowflake SQL objects, optimizing tables, stored procedures, functions, and views for enhanced system performance.
 Monitored SLA Azure Data Factory pipelines, promptly addressing issues to rectify failures and ensure smooth operations.
 Employed code coverage tools such as Jacoco, code vulnerability tools like Sonar Cube & code security tools like Coverity and Source Clear
 Spearheaded implementation of CI/CD pipelines, showcasing proficiency in GIT, and actively contributed to Agile projects using JIRA,
Confluence, and Git for seamless project delivery.
 Demonstrated proficiency in database management (SQL and NoSQL), displaying adeptness in interacting with APIs, including the
creation and documentation of endpoints, with a strong understanding of data formats like JSON and XML.

Role:
Apr’21 – July’21 as System Engineer-Trainee
 Worked on data warehousing concepts and ETL processes; performed data analysis and manipulation using Pandas and Numpy.
 Collected, analyzed, and visualized large datasets; analyzed data using Hadoop and Spark.

PERSONAL DETAILS
Address: Bangalore Electronic City
Languages Known: English, Hindi, Kanada, Tamil

You might also like