0% found this document useful (0 votes)
8 views3 pages

ArvindKumar for Data Engineer_Chennai_Sagent

Arvind Kumar S is a Data Engineer with 8.3 years of experience, specializing in Python, Spark, and Informatica Power Center. He has hands-on experience with AWS services and has worked on various projects involving data processing, ETL operations, and data transformations. His educational background includes a Master's degree in Power Electronics and Drives from Easwari Engineering College, Chennai.

Uploaded by

KUMARESAN 1810
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views3 pages

ArvindKumar for Data Engineer_Chennai_Sagent

Arvind Kumar S is a Data Engineer with 8.3 years of experience, specializing in Python, Spark, and Informatica Power Center. He has hands-on experience with AWS services and has worked on various projects involving data processing, ETL operations, and data transformations. His educational background includes a Master's degree in Power Electronics and Drives from Easwari Engineering College, Chennai.

Uploaded by

KUMARESAN 1810
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 3

ARVIND KUMAR S Phn.

No: +91-9025365326
Data Engineer Email: [email protected]

Summary

• Enthusiastic persistent and result driven individual with 8.3 years of overall experience in complex
and heterogeneous IT environment. With around 4.5 years of experience in Python and Spark and
3.8 years of experience in Informatica Power center.
• Experience in Apache Spark along with Python.
• Hands on experience on AWS (Lambda, Storage S3 and Glue).
• Used Standard Python Modules like Json, PySpark.
• Good experience in writing Spark applications using Python for big data processing.
• Very Excellent development experience in Python, Spark, SQL Hands-on experience with Pyspark
Data Frame and Pandas Data Frame.
• Good working knowledge in the Agile methodology with hands on experience on JIRA.
• Good working experience on AWS distributions and efficiently used Python and Pyspark for data
ingestions and Data Transformations, respectively.

Skills

Distributions AWS & Snowflakes


IDE PyCharm
Processing Engineer SPARK
NOSQL DB Mongo DB, Dynamo DB & Oracle
Operating Systems Windows
Languages Python and SQL
VCS GIT HUB and Terraform
ETL Tools Informatica Power Center 10.4.1

Experience
Role: Data Engineer
Company: HCL
Duration: 01/2021- Present
Client: Else, Oli

Project Description:

Purpose of the project was to get the behavior of an employees in an organization and we get
the data from S3 and process via Python and Pyspark, using PyCharm and AWS glue

Roles and Responsibilities:

 Using Pyspark and performing ETL operations on raw data from client.
 Worked on Merging of Data frames and filtering out desired results and providing it to ML team
 Write to the S3 Bucket in Parquet Format.
 Worked on transaction data and performing transformations for increasing value of data
received.
 Performed Transformation and providing it back to ML engineers for Analysis
 Query the S3 Data or Dynamo DB data in Athena using Crawler.
 Collect data from multiple APIs and combine and process them using PySpark and Python
 Utilizing GIT and TERRAFORM to implement the accepted POC build in a higher environment.
Role: Data Engineer
Company: Intellecto Global Services
Duration: 11/2019 to 12/20
Client: ELSEVIER

Project Description:

Purpose of the project was to get the data from MongoDB and process via Python - Pandas and
Pyspark using PyCharm.

Roles and Responsibilities:

 Established connectivity to MongoDB using PyCharm and PyMongo for seamless data access
and manipulation.
 Developed Python-Pandas scripts to create and update collections within the MongoDB
database.
 Implemented comparison and merging strategies for collections based on client-specific
conditions to ensure data integrity and consistency.
 Participated actively in code reviews and deployment processes to maintain code quality and
deployment efficiency.
 Coordinated closely with onshore partners to gather and understand project requirements,
facilitating continuous delivery and alignment with client expectations.

Role: Data Engineer


Company: Intellecto Global Services
Duration: 07/2018 to 10/19
Client: Ivo Korchalin

Project Description:

Working as a Part of Developing a Health Care Record of patients and their medical history from
Data records available and feeding it to AI for Auto Prescription. Data retrieved through APIs and
processed through Pandas for Data insertion, manipulation, and filtering

Roles and Responsibilities:

 Data Retrieving through API and code through AWS LAMBDA


 Data Ingestion through Lambda in AWS S3
 Performing POC on client’s local machine through Pycharm and Implementing it on Lambda
 Developing an existing lambda module, testing it separately in Pycharm using VCS git link
 Performed Spark Optimization technique by applying Repartition and Coalesce for betterment
of data processing
 Created AWS IAM when implementing Lambda
 Implementing Batch processing on newly arrived data on existing S3 Bucket through AWS Glue
 Creating Crawlers with respective to the Database and Data location given
 Performed Feature testing on AWS Glue and AWS Lambda before adding the VCS branch to
main Tree on GIT
Role: Informatica Developer
Company: 4i Apps Solution Pvt. Ltd, Chennai
Duration: 10/2016 to 06/18
Client: Viacom

Project Description:

The objective of this project is to handle the business logic between medical representatives and doctors
using Informatica Power center.

Roles and Responsibilities:

 Knowledge in Full Life Cycle development of Data Warehousing.


 Have worked in developing ETL program for supporting Data Extraction, transformations and
loading using Informatica Power Center.
 Experience in SDLC process. Analysis, Design, Coding, Unit Testing, Integration Testing and
Implementation experience.
 Experience with dimensional modeling using star schema and snowflake models.
 Strong with relational database design concepts.
 Involves Performance Tuning for Optimize the Target, Source, Mapping, Transformations and
Sessions...
 Experienced in interacting with Business users in analyzing the Business process requirements and
transforming them into documents, designing, and rolling out the deliverables.
 Extensive experience in ETL process consists of data transformations, sourcing, mapping,
conversion and loading. Familiar with ETL Tool Informatica having fair knowledge on
transformations, extensive experience in creation of ETL Mappings.
 Worked on optimizing the mappings by creating re-usable transformations and Mapplets. Created
debugging and performance tuning of sources, targets, mappings, transformations and sessions.
 Having good experience in Oracle 11g.

Education

M.E (Power Electronics and Drives) from Easwari Engineering College, Chennai in 2012.

You might also like