0% found this document useful (0 votes)
18 views2 pages

Hrushi de Update

Hrushikesh Langute is an Azure Data Engineer with over 6 years of experience in designing and maintaining scalable data solutions in cloud environments, proficient in SQL, Python, and Azure services. He has worked extensively with data pipeline development, optimization, and data modeling across various industries, including airline and insurance. His recent roles involved automating ETL processes, optimizing data transformations, and ensuring data quality and integrity using tools like Azure Data Factory and Snowflake.

Uploaded by

umeshrathod130
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
18 views2 pages

Hrushi de Update

Hrushikesh Langute is an Azure Data Engineer with over 6 years of experience in designing and maintaining scalable data solutions in cloud environments, proficient in SQL, Python, and Azure services. He has worked extensively with data pipeline development, optimization, and data modeling across various industries, including airline and insurance. His recent roles involved automating ETL processes, optimizing data transformations, and ensuring data quality and integrity using tools like Azure Data Factory and Snowflake.

Uploaded by

umeshrathod130
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 2

Hrushikesh Langute

Azure Data Engineer

Contact Profile Summary


Azure Data Engineer with over 6+ years of experience specializing in designing,
pl.hrushikesh@gmail.com building, and maintaining scalable data solutions in cloud environments.
Proficient in SQL, Python, PySpark, Azure Data Factory and Snowflake.
7821989038 Expertise in data pipeline development, optimization, data modeling, and
India working in diverse domains such as Airline and Insurance (property
and casualty). Adept at transforming raw data into actionable insights
while ensuring data quality, performance and security.

Technical Skillset Work Summary


Languages:- SQL, Python, PySpark Over 6+ years of experience as a Data Engineer.
Azure Services:- Microsoft ADF, ADB, Azure SQL Developed and scheduled scalable ETL workflows using Azure Data
Database, Azure Monitor, Azure Key Vault, ADLS Factory, Databricks, and ADLS to automate data processing.
Data Warehouse:- Snowflake Experienced in big data technologies, particularly Azure Databricks,
Big Data Technologies:- Spark, Databricks utilizing frameworks like Spark for scalable, distributed data processing.
Data Integration Tools:- Azure Data Factory worked with various Spark components such as Spark Core, SparkSQL,
(ADF), SSIS DataFrame API and File Formats like Text, CSV, JSON, Parquet.
Version Control:- Git, GitHub Worked with various file formats such as Text, CSV, JSON, and Parquet.
ETL/ELT Tools:- Azure Data Factory, Snowflake Leveraged a variety of Data engineering tools to streamline and enhance
Streams and Tasks including data extraction, transformation, loading, and analysis.
Visualization:- Power BI Proficient in using Python concepts like basic data structures like String,
CI/CD:- Azure DevOps (Knowledge) List, Set, Tuple, Dictionary, Anonymous function, and basic OOPs
Methodologies:- Agile, Scrum concepts.
IDES:- Oracle pl/sql Developer, PyCharm, Jupyter Implemented data enhancements, schema validation, and techniques such
Notebook, Thonny as normalization and Slowly Changing Dimensions (SCD) to optimize data
processing and storage.
Proficient in SQL concepts like RDBMS, DDL, DML, DRL & experience with
Education databases like Oracle and Snowflake, with expertise in writing, optimizing
Bachelor of Computer Application, MIT, PUNE complex SQL queries for efficient data processing.
Experienced in building data pipelines, and warehouses, having good
knowledge of OLTP, OLAP, Dimension and Fact Tables.
Language Configured Snowflake Storage Integrations for secure connections with
German -Basic Azure Blob Storage and designed Snowpipe for real-time data ingestion.
Leveraged Snowflake's data-sharing features to facilitate collaboration and
improve project outcomes.
Industry Experience Experience in Snowflake for implementing data warehouse solution for
downstream consumption, including efficient data loading using Snow
GQBAY Software Private Limited, Hyderabad Pipe.
December 2023 -- Current Date Experienced in building data pipelines and warehouses, with good
knowledge of OLTP, OLAP, dimension and fact tables.
Allstate Private Limited, Pune Strong technical, communication, analytical, and problem-solving skills.
May 2018 -- November 2023 Strong team player with excellent communication, analytical, presentation,
and interpersonal skills.
Good collaboration with teams to understand and translate business
requirements into data solutions.
Actively participated in daily scrum calls, sprint, retrospectives and
grooming sessions.
Used Azure monitor to get email notifications for pipeline failures in ADF.
Work Experience
______________________________________________________________________
Organization : GQBAY Software Private Limited, Hyderabad
Data Engineer (December 2023 – Present)
Client: Air Greenland
Description: Air Greenland operates helicopter flights to most settlements in Greenland ("settlement flights") on contract with the
government of Greenland, with the destination network subsidized and coordinated by the Ministry of Housing, Infrastructure and
Transport.

Key Contributions:

Responsible for handling data from multiple sources in Azure Data Lake Storage (ADLS) to ensure smooth data processing and
availability.
Enhance the execution speed of PySpark jobs by utilizing Spark optimizations like partitioning, caching, and broadcasting within
the ADB cluster.
Use ADLS features to store data at various stages (raw, silver, refined) ensuring proper segregation and management of datasets
for different processing stages.
Load processed and refined data from ADLS into Snowflake for building a data warehouse, ensuring high data quality and
consistency.
Create relational tables in Snowflake, define relationships between them based on client requirements, and load data from ADLS
into these tables.
Automate the ETL processes using Azure Data Factory (ADF) to create seamless job workflows, streamlining data extraction,
transformation, and loading.
Develop and optimize PySpark scripts to perform complex data transformations on large datasets in the ADB cluster.
Validate and adjust business logic as needed after cross-verifying data between the source and target, ensuring data integrity in
the final output.
Validating the schema of the data and cleaning the data for further transformations.

Organization: Allstate Solutions pvt.ltd, Pune


May 2018 – December 2023
Client: Encompass
Description: The Allstate Corporation is an American insurance company, headquartered in Illinois.operation is focused on the
areas of business intelligence, analytics, testing, and mobility. The company also has personal line insurance, Property
and casualty, Vehicle and Property Insurance, Fire and Casualty Insurance.
Key Contributions:

Extracted and loaded large insurance datasets (policy, claims, underwriting) into the cloud using SQL and PySpark.
Integrated data from legacy systems and third-party services into a unified data lake for analysis.
Executed PySpark code for transformations within Azure Databricks to achieve desired data outcomes and monitored jobs using
Azure Monitor.
Used PySpark for data transformations like cleansing, normalization, and aggregation for analytics.
Implemented aggregations for calculating claims, premiums, and customer lifetime value.
Optimized PySpark jobs using broadcast joins, partitioning, and caching for faster execution.
Designed and implemented custom functions using the def keyword, encapsulating reusable blocks of code to promote modularity
and code reuse.
Maintaining Data Warehouse solutions on Snowflake and implementing incremental load in SCD Type 2.

Client: Esurance
Description: Esurance Insurance Services, Inc. is an American insurance company. It sells auto, home, motorcycle, and renters
insurance direct to consumers online and by phone. Founded in 1999, the company was acquired by Allstate in 2011

Key Contribution:

Worked on SQL queries to fetch data from a database.


Designing and implementing Python utilities for ETL scripts.
Defined data structures, enforced data integrity constraints, and optimized database performance through appropriate indexing
strategies.
Designed and implemented custom functions using the def keyword, encapsulating reusable blocks of code to promote
modularity and code reusability.
Testing the accuracy and reliability of the processed data after transformations

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy