Professional Documents
Culture Documents
Arpit Ashok Patel Resume
Arpit Ashok Patel Resume
Sr.Python Developer
patel2020arpit@gmail.com
(762) 250-1723
Professional Summary: -
Technical skills:
Work Experience:
Description:
Principle Healthcare Europe GmbH is a private label supplier of vitamin and mineral food supplements,
supporting and servicing Central Europe’s largest pharmacy and retail chains. The Company's offerings
include pharmacy benefit management services, mail order, retail and specialty pharmacy, disease
management programs, and retail clinics.
Responsibilities:
Experience with Web Development, Web Services, Python, and the Django framework.
Working knowledge ofIPython Notebook, Pandas, Numpy, Scipy
Experience with modernjavascript front-end frameworks like vue.js, React.
Implemented the micro-service via Scala Play Framework and provide RESTful APIs for other
services in the pipeline to consume it.
Involved in web application development for backend system using AngularJS and Node.js with
cutting edge HTML5 and CSS techniques.
Developed frontend and backend modules using Python on Django including Tastypie Web
Framework using Git. Implemented pre- and post-processing pipelines for Machine Learning
algorithms.
Written with object-oriented Python, Flask, SQL, Beautiful Soup, httplib2, Jinja2, HTML/CSS,
Bootstrap, jQuery, Linux, Sublime Text, git.
Developed Restful microservices using Flask and Django and deployed on AWS servers using EBS
and EC2.
Designed a python script to load transformed data into cloud services (AWS, Azure, and GCP ).
Integrated web applications with Adobe Experience Manager (AEM)
Used Apache Spark Scala as machine learning language for model creation and predictions using
Logistic Regression algorithm.
Developed a Progressive Web App, which more suited for mobile devices.
Design UX interface and UI components using HTML and CSS
Design back schema, and database JSON structures
Used web applications development using Django/Python, Flask/Python, and, JQuery, Ajax while
using HTML/CSS/JavaScript for server-side rendered application.
Wrote a Python module to connect and view the status of an Apache Cassandra instance
Developed MVC prototype replacement of current product with Django.
Developed dashboard using HighCharts JavaScript library.
Familiar with JSON based REST Web services and Amazon Web services (AWS) and Responsible
for setting up Python REST API framework and spring framework using DJANGO.
Implemented Backbone JS for creation of networking applications using JavaScript.
Used Test driven approach (TDD) for developing services required for the application and
Implemented Integration test cases and Developing predictive analytic using Apache Spark Scala
APIs.
Used MySQL database on simple queries and writing Stored Procedures for normalization and
denormalization.
Developed Sybase Stored procedures for the back end processing of the proposed data base
design
Azure cloud Merge with python to store the data into cloud with High security.
Developed and tested many features for dashboard using Python, Java, Bootstrap, CSS,
JavaScript and JQuery.
Integrated GIT into Jenkins to automate the code check-out process.
Developing applications using RESTFUL architecture using Node.js and PHP as backend
languages and Used Numpy for Numerical analysis and Used Spring Framework to support the
Hibernate tool and Struts.
Environment: Python 3.7, PyQT, PyQuery, MVW, HTML5, CSS3, DOM, Angular.js, Shell Scripting, JSON,
Rest, Apache Web Server, Django, SQL, UNIX, Windows, AWS, MongoDB, PostgreSQL, and python
libraries such as Numpy, IPython, sqlalchemy, Django Tastypie
Description:
NetJets Inc. operates as a private jet company. The Company provides aircraft management, charter
management, on-demand charter, and private flight services. NetJets serves its customers worldwide.
Responsibilities:
Involved in building database Model, APIs and Views utilizing Python, to build an interactive
web-based solution.
Responsible for gathering requirements, system analysis, design, development, testing and
deployment.
Worked on UI using HTML5, CSS3 and Java Scripts.
Experience working with Azure SQL Database Import and Export Service.
Created User Controls and simple animations using Java Script and Python.
Used Pandas library for statistical Analysis.
Perform quality control checks of GIS data and non-spatial databases to ensure data integrity
Used Django Database API's to access database objects.
Scalable, database-driven web application development using a variety of frameworks: ASP.NET
on C#, Flask on Python, and PHP.
Designed and implemented Model-View-View Model pattern for many components in the
project (Contracts, Products, Search Sales, Employees, BackOffice Admin).
Developed tools using Python, Shell scripting, XML to automate some of the menial tasks.
Interfacing with supervisors, artists, systems administrators and production to ensure
production deadlines are met.
Responsible for managing and deployment of product WSGI server and Apache Server on
Microsoft Azure Centos Server.
Worked on ETL tasks like pulling, pushing data from and too various servers.
Tested and evaluated results for inclusion into software product.
Designed and deployed machine learning solutions in Python to classify millions of previously
unclassified Twitter users into core data product.
Worked on testing and Debugging the Embedded C code and handling various Interrupts in
various environments and writing test scripts in C for testing requirements in Auto-tester.
Exploring with the Spark improving the performance and optimization of the existing algorithms
in Hadoop using Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.
Responsible for Plug-in Management, User Management and Backup/Disaster Recovery
Plan/Implementation (BDR) on Jenkins.
Part of team implementing REST APIs in Python, Flask and SQLAlchemy for management of data
center resources on which OpenStack cluster is deployed
Performed automation of testing using Quick Test Pro and programmed dynamic VB Scripts in
Expert view for testing GUI Functionality.
Automated continuous integration delivery workflows to deploy Micro services applications via
Docker containers.
Experience working with large datasets and Deep Learning class using TensorFlow and Apache
Spark. Worked with Spark Core, Spark ML, Spark Streaming and Spark SQL.
Enabled speedy reviews and first mover advantages by using Oozie to automate data loading
into the Hadoop Distributed File System and Pig to pre-process the data.
Installed, configured and managed the ELK (Elastic Search, Log Facilitated Scrum ceremonies like
Sprint planning, retrospectives, Daily stand-ups, etc. Stash and Kibana) for Log management
within EC2/ Elastic Load balancer (ELB) for Elastic search
Met client audit requirements by developing python scripts to retrieve AWS RDS logs using
Boto3
Used JQuery and Ajax calls for transmitting JSON data objects between frontend and controllers.
Loaded and transformed large sets of structured, semi-structured and unstructured data using
Hadoop/Big Data concepts.
Developed SQL Queries, Stored Procedures, and Triggers Using Oracle, SQL, PL/SQL.
Created PHP/MySQL back-end for data entry from Flash.
Using SQL Server Reporting Services (SSRS) delivering enterprise, web-enabled reporting to
create reports that draw content from a variety of data sources.
Used JQuery for selecting DOM elements when parsing HTML.
Used GitHub for version control.
Environment: Python, Django, Couchbase Python, PHP, C++, HTML, JQuery, AJAX, XHTML, JavaScript,
XML, JSON, GitHub, Flash, SQLite, MYSQL, SQL, PLSQL, Hadoop, Oracle and Windows.
Description: - This project is to migrate B2C and Pro customers to B2B platform and provide Order
History, Order Details for Online and In Store Orders / Transaction and Calculate Accruals (Loyalty
Rewards) for B2B customer.
Responsibilities:
Description: - Capital One Financial Corporation is a bank holding company specializing in credit cards,
auto loans, banking and savings products headquartered in McLean, Virginia.
Responsibilities:
Planned, coordinated, and monitored project levels of performance and activities to ensure
project completion in time.
Created new reports based on requirements. Responsible in Generating Weekly ad-hoc Reports
Worked in a Scrum Agile process& Writing Stories with two-week iterations delivering product
for each iteration.
Automated and scheduled recurring reporting processes using UNIX shell scripting and Teradata
utilities such as MLOAD, BTEQ and Fast load
Involved in defining and Constructing the customer to customer relationships based on
Association to an account & customer.
Worked on transferring the data files to vendor through sftp &Ftp process
Experience in performing Tableau administering by using tableau admin commands.
Created action filters, parameters and calculated sets for preparing dashboards and worksheets
in Tableau.
Interacted regularly with SMEs to understand requirements and incorporate them into the
application.
Used Haystack for Text Search in the application.
UI interactions is done with the help of JQuery.
UsedGIT for Source code control and JIRA for Bug tracking.
Worked with project team representatives to ensure that logical and physical data models were
developed in line with corporate standards and guidelines.
Worked with architects and, assisting in the development of current and target state enterprise
level data architectures
Created Excel charts and pivot tables for the Ad Hoc data pull
client and server logstash nodes for caching/persistence to enable real-time online updates to
the logstash configuration.
Worked in a Linux environment for implemented the application.
Written Scripts to establish continuous workflows from different teams providing data.
Used SQLite database for development and SQL server for production.
Responsible for defining the key identifiers for each mapping/interface.
Involved in defining the source to target data mappings, business rules and data definitions.
Migrated three critical reporting systems to Business Objects and Web Intelligence on a
Teradata platform
Performed data analysis and data profiling using complex SQL on various sources systems
including Oracle and Teradata.
Environment: Teradata 13.1, Informatica 6.2.1, Ab Initio, Business Objects, Oracle 10g/9i, PL/SQL,
Microsoft Office Suite (Excel, Vlookup, Pivot, Access, Powerpoint), Visio, VBA, Micro Strategy, Tableau,
UNIX Shell Scripting ERWIN).
Description: - Kentucky Insurance, Inc. is an independent insurance agency offering all types of
insurance products from top rated, financially sound, insurance companies.
Responsibilities:
Developed Views and Templates with Django view controller and template language to create a
user-friendly website interface.
Implemented user interface guidelines and standards throughout the development and
maintenance of the website using the HTML, CSS, JavaScript and JQuery.
Created APIs, database Model and Views Utilization Python to build responsive web page
application.
Stored the data in the form of JSON structure-based documents, stored in a collection using
MongoDB.
Developed GUI using Django for dynamically displaying the test block documentation and other
features of Python code using a web browser.
Built various graphs for business decision making using Python matplotlib library.
Worked in developing Python/Django based web application, MySQL DB and integrations with
3rd party email, messaging, Storage services.
Worked in developing a fully automated continuous integration system using Git, Gerrit, Jenkins,
MySQL.
Developed and executed User Acceptance Testing portion of test plan.
Used Amazon Elastic Beanstalk with Amazon EC2 to deploy project into AWS and good
experience with AWS storage services (S3) .
Responsible for testing methodologies like unit testing, Integration testing, web application
testing.
Performed user validations on client side as well as server side.
Created Python Django forms to record data of online users and used PyTest for writing test
cases.
Efficient delivered code based on principles of Test Driven Development (TDD) and continuous
integration to keep in line with Agile Software Methodology principles.
Participated with QA to develop test plans from high-level design documentation.
Developed web-based applications using Python 2.7/2.6, Django 1.4/1.3, PHP, Flask, Webapp2,
Angular.js, VB, C++, XML, CSS, HTML, DHTML, JavaScript and jQuery
Developed automated process for builds and deployments by using Jenkins, Ant, Maven, Shell
Script.
Documented company Restful API's using Swagger for internal and third part use and worked on
Unit testing and Integration testing.
Deployed the project into Jenkins using GIT version control system.
Worked on changes to open stack and AWS to accommodate large-scale data center
deployment.
Environment: Python, Java, Django, AWS, Angular Js, Jenkins, Docker, MySQL, MongoDB, HTML,
JavaScript, JSON, XML, JQuery, Linux WSDL, Restful, CSS, Postgre-SQL, Swagger, Git, JIRA.
Bill Me Later Inc. Lutherville-Timonium, Maryland, United States Feb 2013 to Jan 2014
Python Developer
Description: - Bill Me Later is a leader in the digital payments industry with its popular Bill Me Later®
product and flexible financing programs. The rapidly growing Bill Me Later network enables top-tier
retailers and travel providers to attract high value customers with an effortless payment experience.
Responsibilities:
Environment: Eclipse, Java, JSP, Ajax, Struts, Spring Framework, JNDI, UDDI, WSDL, SOAP, Agile, UML,
XML, HTML, Log 4j, Oracle 10g, Shell Script, CVS, Maven and Windows XP.
Education - Bachelor’s in computer science In Database administration, Lock haven University 2012.