Azure Data Engineer
Azure Data Engineer
Job description
Responsibilities :
Design, develop, and implement Azure-based data solutions, including data lakes, data
warehouses, and data processing systems.
Build and optimize data pipelines for extracting, transforming, and loading (ETL) data from
various sources into Azure environments.
Collaborate with data architects, analysts, and other stakeholders to understand data requirements
and translate them into technical solutions.
Ensure data security, integrity, and performance by implementing appropriate Azure data services
and best practices.
Perform data modeling, schema design, and database optimization for efficient data storage and
retrieval.
Develop and maintain documentation for data processes, configurations, and procedures.
Monitor and troubleshoot data solutions to identify and resolve performance issues or bottlenecks.
Stay updated with emerging Azure technologies and trends to recommend enhancements and
improvements.
Description :
Snowflake Data Engineer who can perform the activities listed below.
Knowledge of Snowflake platform
Creating objects (Tables, Views, Procedures (using Javascript, SQL), User-defined Functions)
Advanced SQL Skills.
Good understanding of advanced Snowflake concepts
Tasks, Streams and Dynamic Tables.
Different Types of Integrations
Snowflake Costs
Role-based Access Control(RBAC)
Snowflake Security
Streamlit in Snowflake
Proficiency in at least one modern data stack tool - Matillion, Azure Data Factory, Fivetran+dbt
Intermediate Python Skills.
Responsibilities:
Design, build, and maintain scalable data pipelines using PySpark and Databricks
Optimize data processing and storage for maximum performance and efficiency
Troubleshoot and debug data-related issues, and implement solutions to prevent reoccurrence
Collaborate with data scientists, software engineers, and other stakeholders to ensure that data
solutions are aligned with business goals
Requirements:
Strong experience in Python programming or PySpark, and SparkSQL
Clear understanding of Spark Data structures, RDD, Dataframe, dataset
Expertise in Databricks and ADLS
Expertise handling data type, from dictionaries, lists, tuples, sets, arrays, pandas dataframes, and
spark dataframes
Expertise working with complex data types such as, structs, and JSON strings.
Clear understanding of Spark Broadcast, Repartition, Bloom index filters
Experience with ADLS optimization, partitioning, shuffling and shrinking
Ideal experience with disk caching
Ideal Experience with cost based optimizer
Experience with data modeling, data warehousing, data-lake, delta-lake and ETL/ELT processes
in ADF
Strong analytical and problem-solving skills
Excellent documentation, communication and collaboration skills