Deepak Garg
Deepak Garg
SUMMARY:
Big Data (Spark, Scala, Pyspark, Sqoop, Hive) & Cloud technologies (Azure, AWS, GCP) with 7 years of overall
experience.
Expertise in understanding and analyzing the provided business component specifications and developing
solutions by using Big Data technologies.
Good experience in implementing the pipeline in Azure Data factory and Azure Databricks.
Experience in Data Ingestion, Data Filtering, Data Analysis, Data Modeling (Hive), Data storage, data processing
and data analysis.
Conversant with design, development, maintenance and support of Big Data Analytics using Hadoop Ecosystem
components like HDFS, Hive, Sqoop, Spark, Scala and Airflow.
Good experience in preparing technical documents and performing unit tests as per the functional/business
requirements provided by the client.
Good knowledge of the Agile & Scrum process methodology.
Good experience in Extraction, Transformation and Loading (ETL) in SQL Server database.
Good experience of coordinating with different teams spread across various locations.
Proficiency in interacting with clients to ensure delivery of a seamless solution that meets the client requirements.
Dedicated and self-motivated achiever adept at juggling multiple tasks in a high-pressured environment with
exceptional team-building skills.
EDUCATION:
B.Tech in 2013.
HIGHLIGHTS:
DOMAIN/ SUB-DOMAIN CLOUD CORE COMPETENCIES CORE COMPETENCIES
Telecommunication Spark Azure Data Factory
Healthcare Spark Streaming Databricks
BFSI Scala, Python Azure Databricks
Hadoop, Hive, Sqoop, Cassandra CosmosDB
Azure SQL
Azure Key Vault
Azure Data Lake Storage Gen2
AWS S3
Redshift
Lambda
TECHNICAL SKILLS:
Technology Spark, SparkSQL, Hadoop, Hive, Sqoop
Language Scala, Python
Database MySQL, HBase, Cassandra
Tools Hortonworks, Cloudera, Databricks, IntelliJ, JIRA, Putty, WinSCP, Docker
AWS Technology S3, Redshift, Athena, EMR, Lambda
Azure Technology ADF, Azure Data Lake Storage, Azure Blob Storage, Azure Databricks, CosmosDB
Operating Systems Window/XP/7/8/10, Windows 2003 server, Linux
Others Confluence, Jira, GitHub, Big Bucket
Methodologies SAFe, Agile, Scrum
PROFESSIONAL EXPERIENCE:
Duration Organization Designation/ Role Responsibilities
Aug 2021 to Till Date Freelancer Senior Software Engineer Big Data Engineer
Jan 2020 to Nov 2020 Cognizant Senior Software Engineer Senior Data Engineer
Feb 2014 to Mar 2019 Infosys Limited Technology Analyst Big Data Developer
PROJECT DETAILS:
Project Name: Uklyf Simplification
Organization: StackNexus (Freelancing Project)
Duration August 2021 to till date Role Azure Data Engineer
Project Type Enterprise Data Lake Technology Spark, Scala, Azure Data Factory, Azure
Date Lake Storage, Azure Cosmos,
Azure Databricks
Responsibilities Migrated Informatica to Spark
Designed and developed end-to-end Spark module for the application.
Used Connectors to load data from SFTP to Azure Data Lake Storage
Used ADLS to store raw data.
Developed Azure Databricks Notebook and apply business related transformation
to standardize the data.
Moved transformed data to the salesforce.
Scheduled the pipeline using ADF trigger and monitored it.
Project Name: Synchrony Bank | Ingestion Data in Hive, Transformation using Spark & loading it into S3.
Organization: Cognizant Solutions
Duration Jan 2020 – Nov 2020 Role Senior Big Data Developer
Project Type Development, Enhancement Technology Pyspark, Hive, Sqoop, Shell scripting,
Hadoop, S3, Redshift, API
Responsibilities Ingested data into Hive using PySpark.
Applied business related transformation on the data using Spark and Spark SQL & stored
the data into s3.
Performed unit testing and prepared the UTC document.
Prepared Release Notes and deployment documents.
Modified the code and analyzed Spark web UI to optimize query performance.
Responsible for understanding and analyzing the requirements.