Professional Documents
Culture Documents
Naukri Vijay (3y 0m)
Naukri Vijay (3y 0m)
Naukri Vijay (3y 0m)
+91-9941119812
As a Cloud Technician, aim to apply my expertise in designing and implementing robust data
architectures, optimizing ETL processes, and managing data infrastructure. I strive to enhance vijaykumar06899@gmail.
the seamless flow, storage, and analysis of data within an organization, utilizing skills in com
technologies like Spark, Hadoop, AWS, and various programming languages Bengaluru
CERTIFICATIONS
Certified Spark Professional
EDUCATION Certified Devops
2020: Bachelor of Science (Computer Science) Certified Data Management
Madras University, Chennai with 75% certified Big Data Developer
Certified Python developer
Certified Snowflake Developer
PROJECTS
Neural Machine Translation:
Machine translation using sequence-to-sequence architecture
Covid Forecasting:
Predict covid effect and its increase and decrease rate, each state-wise statistics and
prediction to take precaution
IT SKILLS
Languages: Python, Java, Scala, SQL, Hive, Spark, Pyspark
Skills: DBMS, Linux, Spark Streaming, Hadoop, Data Analytics, ETL, Kafka
Streams, Microservices, Quantexa
Cloud Technology: AZURE- ADF, Databrick; AWS - Snowflake, S3, AWS, GCP
Glue
WORK EXPERIENCE
Since April’21 with TCS
Growth Path:
Client: Apple
Sep’22 – Present as Specialist
Role:
Engaged in active dialogues with the business team to collaboratively devise the architecture and assess the feasibility of the solution.
Formulated end-to-end ETL pipelines encompassing diverse source and destination combinations such as Snowflake, Teradata, HDFS,
Kafka, and AWS S3. Generated and supervised production jobs, addressing any failed jobs and conducting error analyses on platforms like
Splunk, Airflow, Autosys, and Spark UI.
Enhanced Spark job performance in AWS through meticulous optimization and fine-tuning of Spark applications, including memory
adjustments based on specific requirements.
Utilized Scala & Hive queries within Spark ecosystem, developing proof-of-concepts with Spark SQL to advance data processing capabilities.
Conducted data analysis in Hive, strategically creating partitioned and bucketed tables to maintain operational efficiency.
Responded to inquiries from agents and business users by crafting Teradata, Snowflake, and Datalake queries, facilitating prompt and
accurate information retrieval.
Implemented scripts for Datalab and Snowflake tables, leveraging Python for data delay monitoring.
Took the lead in designing and implementing a robust data engineering pipeline, optimizing SQL queries, creating efficient data models,
and implementing orchestration workflows for enhanced overall efficiency.
Client:
Jul’21 Apple
– Sep’22 as System Engineer
Role:
Crafted an interactive data analysis dashboard through Streamlit, providing stakeholders with illuminating visualizations to facilitate well-
informed decision-making.
Streamlined system efficiency by transitioning code from Scala to Python, and instituted Python scripts for backfill, daily operations, and
automated job triggering.
Conducted thorough data analysis and manipulation using Pandas and Numpy, extracting valuable insights to inform decision-making.
Managed Snowflake SQL objects, optimizing tables, stored procedures, functions, and views for enhanced system performance.
Monitored SLA Azure Data Factory pipelines, promptly addressing issues to rectify failures and ensure smooth operations.
Employed code coverage tools such as Jacoco, code vulnerability tools like Sonar Cube & code security tools like Coverity and Source Clear
Spearheaded implementation of CI/CD pipelines, showcasing proficiency in GIT, and actively contributed to Agile projects using JIRA,
Confluence, and Git for seamless project delivery.
Demonstrated proficiency in database management (SQL and NoSQL), displaying adeptness in interacting with APIs, including the
creation and documentation of endpoints, with a strong understanding of data formats like JSON and XML.
Role:
Apr’21 – July’21 as System Engineer-Trainee
Worked on data warehousing concepts and ETL processes; performed data analysis and manipulation using Pandas and Numpy.
Collected, analyzed, and visualized large datasets; analyzed data using Hadoop and Spark.
PERSONAL DETAILS
Address: Bangalore Electronic City
Languages Known: English, Hindi, Kanada, Tamil