Srilakshmi M Resume
Srilakshmi M Resume
Mannemala
Data Engineer Consultant Data Engineer Hyderabad, IN
Contact#: +0091 7981598455 Sapient (Jan 2022-Apr 2023)
e-mail: Hyderabad, IN
Performing data ingestion from RDBMS to HDFS using Sqoop for
further processing using Hive and Spark SQL.
Designing and building scalable Big Data pipelines using PySpark for
data transformations.
Developing Python scripts as reusable, dynamic & generic utilities
for Spark and ETL jobs.
SUMMARY Developing ETL scripts using PySpark and Hive
7 years of IT Professional with 4.5 years
Maintaining and Scheduling of Spark jobs using Oozie Workflow.
experienced & result- oriented Data and
Designing solutions & codes using the Hadoop Framework to
Spark Engineer possessing a proven track
generate multiple reports and integration with dashboards.
record in software development using
Hadoop Apache Spark, Hive, SQL, Python, Developing SQL queries for data analysis and
AWS. extraction Designing of Shell Scripts for Data
Proficient in processing movement and file processing.
structured/unstructured data, deploying Consuming Data from Microsoft SQL Server and
Apache Spark to analyze huge datasets. Oracle as a data extraction layer.
KEY SKILLS
Data Engineer Hyderabad, IN
SQL
CapGemini (Feb 2018-Dec 2021)
Apache Spark (Python)
Big Data- Hadoop Built scalable and configurable applications on Big
ETL Data/Spark Framework on AWS EMR Cluster.
Python Built Data Marts on AWS cloud for Business Applications and
AWS Glue, Red Shift, S3 Reporting Deployment and Execution of Spark Batches on EMR
Clusters.
TECHNICAL SKILLS
Designed and Built Hive External Tables with S3 buckets as data
Big Data Ecosystem: Hadoop, Hive,
source location.
Sqoop, Apache Spark, Oozie
Extensively used AWS ' services for data engineering tasks.
Spark Framework: Spark RDD's,
DataFrames, Spark SQL Performance Building Data Marts on AWS cloud for Business Applications
Optimization and Tuning & Reporting.
Developing ETL scripts using PySpark.
Maintaining the data pipeline using Apache Airflow.
Working on PySpark framework to build data pipelines for
ETL processes.
Leveraging AWS cloud platform for accomplishing Data
Engineering goals.
Developing SQL queries for data analysis and extraction.
PROFESSIONAL EXPERIENCE Contd..
SQL Developer
Cloud: AWS WIPRO TECHNOLOGIES Hyderabad IN
(Jan 2017-Jan 2018)