Data and Analytics GCP Data Engineer
Data and Analytics GCP Data Engineer
COM 5199950106
PROFESSIONAL SUMMARY
Experienced data Integration Developer skilled in designing and implementing robust data pipelines on Google Cloud
Platform (GCP), utilizing technologies such as Google Cloud Dataflow, Apache Beam, and Apache Kafka to enable
seamless data ingestion and real-time processing. Proficient in SQL for optimizing ETL processes and database
performance, with a proven track record in achieving significant performance improvements during data warehouse
migrations. Strong in Python scripting for automation, reducing manual effort and enhancing data accuracy across
platforms. Adept at data mapping, transformation, and ensuring data quality and compliance in cloud environments.
Experienced and driven SQL Developer and Business Intelligence Specialist with a passion for leveraging data to
drive strategic decision-making. Skilled in crafting SQL scripts and stored procedures for ETL processes, optimizing
database performance, and creating intuitive dashboards using Power BI. Excels in collaborating with HR
stakeholders to understand their needs and translate them into actionable insights.
Expertise in conducting comprehensive risk assessments, developing robust risk management strategies, and
implementing effective controls to optimize organizational resilience. Strong analytical skills combined with a deep
understanding of industry-specific risk factors enable precise risk forecasting and proactive decision-making. Adept
at communicating complex risk issues clearly and concisely to stakeholders at all levels, fostering a culture of risk
awareness and adherence to compliance standards.
PROFESSIONAL EXPERIENCE:
Responsibilities:
Designed and built advanced analytical solutions to address complex financial challenges, leveraging strong
quantitative skills and proficiency in Excel and SQL for financial modeling, data analysis, and forecasting.
Managed cloud data warehouse resources and services, overseeing data storage solutions, query optimization,
and performance tuning for high-volume financial data processing tasks. Implemented strategies to improve data
management efficiency and support complex financial analytics.
Utilized GCP BigQuery and Dataflow to develop efficient data warehousing solutions for high-volume financial
data, designing architectures that supported complex queries, analytics, and reporting for business intelligence.
Ensured efficient data management and analytical capabilities.
Built and maintained data transformation and loading processes for financial datasets using DBT and GCP
BigQuery, ensuring high data quality and efficient ETL workflows. Designed and managed processes for financial
data integration and reporting, supporting advanced analytics and financial performance monitoring.
Developed and maintained batch and real-time data processing frameworks using Spark and GCP Dataflow,
optimizing data pipelines for efficient handling of large financial datasets. Created frameworks for both batch
and streaming data processes to meet diverse financial data processing needs.
Designed and implemented data-driven services for asynchronous environments, creating scalable data pipelines
and ensuring robust data processing workflows. Enabled efficient integration and analysis of financial data for
both real-time and batch processing scenarios.
Engineered real-time data processing applications using Spark Streaming, enabling continuous monitoring of
financial transactions and the detection of anomalies. Developed applications for real-time financial data
analysis, improving business decision-making and compliance.
Led the creation of automated data transformation and reporting processes using DBT, enhancing data pipeline
efficiency and supporting comprehensive financial reporting and business intelligence initiatives. Developed best
practices for automated data transformation and reporting.
Developed and maintained automated ETL workflows with DBT and GCP Dataflow, establishing efficient data
pipelines for transforming financial and operational data. Ensured reliable data movement, performed data
transformations, and created automated processes for business intelligence and comprehensive financial
reporting.
Oversaw meticulous data mapping and transformation processes to facilitate seamless data transfer and ensure
compatibility across diverse systems. Facilitated effective decision-making and reporting for stakeholders through
robust data management practices.
Developed and deployed a cutting-edge real-time data streaming application using Apache Kafka and Google
Cloud Dataflow. Achieved a significant reduction in data latency, enabling timely insights crucial for agile business
decision-making.
Designed and maintained SQL database schemas, indexing strategies, and query optimization techniques.
Ensured efficient data retrieval and storage management within cloud-based environments, optimizing
performance and scalability.
Executed meticulous data cleansing and validation processes during the migration of on-premises data to GCP.
Upheld stringent data quality and integrity standards while adhering to regulatory and compliance requirements,
ensuring data reliability and trustworthiness.
Implemented Apache Airflow to orchestrate automated ETL pipelines, seamlessly transferring data from
MongoDB to Google Cloud Storage. Enhanced data accessibility and operational efficiency through streamlined
and automated workflows.
Collaborated closely with cross-functional teams to diagnose and resolve complex data issues. Optimized data
flow and integration processes, showcasing strong communication skills and a deep understanding of data
architecture and integration principles.
Demonstrated perfect proficiency in SQL, T-SQL, including SQL Server, SSIS, and Azure environments like
Databricks and Power Platform for data manipulation and analysis. Developed complex SQL queries and stored
procedures to extract, transform, and load data from various sources, supporting decision-making processes.
With Power BI, I excelled in creating interactive dashboards that visualized complex data sets from disparate
sources. For instance, I developed a sales performance dashboard that integrated data from CRM systems and
ERP databases, allowing real-time monitoring of sales trends, customer demographics, and pipeline analysis.
These dashboards not only facilitated strategic planning but also empowered sales teams with actionable
insights to optimize their approach.
Utilized Power BI features such as DAX expressions and data modeling to create visually appealing and
informative reports, driving business growth.
Spearheaded the optimization and creation of dynamic PowerBI dashboards, driving informed decision-making
and tracking performance outcomes. Leveraged expertise in DAX and Python to develop intuitive and engaging
reports, empowering leadership with self-service capabilities.
Applied programming skills in SQL, Python, and PowerShell to automate data pipelines and perform data
transformations. Developed custom scripts and utilities to automate routine tasks and improve productivity,
streamlining data management processes.
Effectively communicated complex financial insights and analysis through written reports and presentations,
including reports on financial performance analysis, budget forecasting presentations, and presentations on
investment strategies based on data-driven analysis. These reports and presentations demonstrate strong verbal
and written communication skills in conveying financial information to diverse audiences.
Spearheaded the implementation and customization of CI/CD pipelines for client projects, utilizing Jenkins,
GitLab CI/CD, and Kubernetes to automate software delivery processes. Achieved a 50% reduction in deployment
times and improved reliability by integrating automated testing frameworks such as JUnit and Selenium.
Led workshops and training sessions for clients on adopting CI/CD best practices, leveraging tools like Docker for
containerization and Ansible for configuration management. Facilitated seamless transitions to automated
pipelines, resulting in accelerated time-to-market and enhanced overall software quality and scalability.
Streamlined project management workflows through proficient utilization of Jira, facilitating efficient task
tracking, issue resolution, and collaboration among team members, resulting in improved project delivery
timelines and outcomes.
Data Analyst
BMO Bank | Toronto, ON August 2020 – November 2022
Responsibilities:
Led data analysis initiatives within a financial context, focusing on leveraging data-driven insights to drive
business decisions and enhance customer experience.
Developed and maintained scalable data pipelines using GCP tools and Python, effectively managing the
asynchronous movement of large datasets across multiple data sources and applications. Ensured data
consistency and reliability through rigorous data validation techniques and automated monitoring, enabling
seamless data integration and comprehensive financial analysis.
Developed and maintained automated ETL workflows with DBT and GCP Dataflow, establishing efficient data
pipelines for transforming financial and operational data. Ensured reliable data movement, performed data
transformations, and created automated processes for business intelligence and comprehensive financial
reporting.
Engineered advanced ELT processes leveraging DBT for data transformation, creating robust workflows that
extracted, transformed, and loaded financial and operational data into BigQuery. Optimized these processes for
both analytical and transactional queries, resulting in efficient data management for reporting, forecasting, and
business intelligence.
Built and managed cloud data warehouse environments on BigQuery and Snowflake, overseeing data storage
solutions, query performance optimization, and resource management. Implemented best practices for high-
volume data processing tasks and complex analytics to support financial reporting and operational efficiency.
Engineered data integration solutions across cloud platforms (AWS, GCP, Azure), creating seamless data
workflows that connected disparate data sources and supported cross-platform data analysis and reporting.
Enabled comprehensive financial data integration and analytics across diverse cloud environments.
Designed and implemented sophisticated data visualizations and interactive dashboards using Tableau and
Looker. Developed comprehensive reporting solutions to deliver actionable insights and performance metrics for
stakeholders across various departments, facilitating strategic decision-making and enhancing data-driven
business strategies.
Performed in-depth data analysis using SQL and Python, conducting exploratory data analysis (EDA) to uncover
trends, anomalies, and actionable insights in large datasets. Presented findings through detailed visualizations
and reports, supporting financial decision-making and strategic business planning.
Designed and implemented data transformation workflows using DBT, creating modular, maintainable
transformation scripts for data extraction, cleaning, and aggregation. Developed frameworks that supported
analytics and reporting needs for financial data, improving the efficiency and quality of data transformation
processes.
Created and maintained comprehensive technical documentation for data pipelines, including detailed data flow
diagrams, ETL process descriptions, and transformation logic. Facilitated knowledge sharing, ensured alignment
with best practices, and provided clear documentation for ongoing maintenance and future enhancements.
Utilized Terraform for the allocation and management of cloud resources, automating the provisioning and
configuration of GCP services. Designed infrastructure as code solutions to support scalable, maintainable, and
secure data engineering environments for large-scale financial data projects.
Collaborated with cross-functional engineering teams to enhance existing codebases, utilizing Python and
Typescript to improve code performance, track system metrics, and increase overall application reliability.
Contributed to codebase optimizations and instrumentation to support scalable financial data processing
applications.
Implemented CI/CD practices for data engineering projects using GitHub Actions and Azure DevOps, automating
the build, test, and deployment processes for ETL workflows and data engineering solutions. Ensured reliable
and efficient deployment of updates and improvements to financial data engineering projects.
Developed and optimized Python-based data processing scripts for large-scale financial datasets, performing
complex data manipulations and calculations. Created scripts to support financial forecasting, risk assessment,
and reporting functions, enhancing analytical capabilities for financial decision-making.
Led cross-functional teams in the development and deployment of business intelligence solutions, providing
guidance and support throughout the project lifecycle to achieve project objectives and deliver value to
stakeholders and managed project timelines, resources, and budgets effectively, ensuring project milestones
were met on time and within budget constraints while maintaining high-quality standards.
Power Bi Developer
Bison Transport | Toronto, ON August 2017 – July 2020
Responsibilities:
Leveraged advanced data modeling techniques in Power BI to design and implement robust data models,
ensuring efficient storage and optimized query performance. Utilized relationships between multiple tables and
implemented complex calculations to derive actionable insights from data.
Employed DAX calculations extensively to create sophisticated measures and KPIs, enabling deeper analysis and
visualization of key business metrics. Utilized DAX functions such as CALCULATE, SUMX, and RELATED to perform
complex aggregations and comparisons within reports.
Developed custom visualizations using the Power BI Visual API to enhance data presentation and analysis.
Integrated custom visuals seamlessly into reports to provide unique insights and improve user engagement.
Utilized Power Query Editor for advanced data transformations, including merging queries, creating conditional
columns, and applying custom functions. Conducted thorough data cleansing and manipulation to ensure data
accuracy and reliability in reports.
Implemented parameterized reports in Power BI to enable dynamic filtering and customization of report views.
Empowered users to interact with reports dynamically, adjusting filters and parameters to explore data from
different perspectives.
Leveraged quick measures and analytical expressions in Power BI to create complex calculations and KPIs
without writing DAX code manually. Streamlined report development process and improved efficiency in
deriving insights from data.
Developed hierarchical data visualizations in Power BI to enable drill-down analysis and exploration of
hierarchical data structures. Enhanced user experience by providing intuitive navigation and deeper insights into
hierarchical data relationships.
Integrated Power BI with Azure Machine Learning to incorporate predictive analytics and machine learning
models into reports and dashboards. Leveraged machine learning algorithms to forecast trends, detect
anomalies, and drive data-driven decision-making.
Configured Direct Query and Live Connection modes in Power BI to enable real-time analysis of data from
external data sources. Provided users with up-to-date insights and enabled faster decision-making based on
real-time data.
Developed paginated reports using Power BI Report Builder for operational reporting and printing needs.
Designed reports with pixel-perfect precision, ensuring accurate representation of data for regulatory
compliance and business reporting requirements.
Leveraged Power BI service integration for collaboration, sharing, and distribution of reports across
organizational teams. Enabled seamless sharing of interactive dashboards and facilitated collaboration on data-
driven insights.
Integrated Power BI with Azure Cognitive Services to incorporate advanced AI insights such as sentiment analysis
and image recognition into reports. Enhanced data analysis capabilities with AI-driven insights, providing deeper
understanding of customer behavior and market trends.
Utilized advanced visualization options in Power BI to create visually appealing and insightful reports. Employed
custom formatting, color schemes, and interactive features to enhance data visualization and storytelling.
Established connections to diverse data sources including cloud services, databases, and custom APIs for
comprehensive data analysis. Combined data from multiple sources to provide holistic insights and facilitate
data-driven decision-making across the organization.
Collaborated with data scientists and analysts to define data requirements and develop solutions for financial
modeling and analysis. Translated business needs into technical specifications and delivered data solutions that
supported advanced financial analysis, risk management, and investment strategies.
Utilized advanced SQL techniques to design and execute complex queries for data extraction and reporting,
performing detailed calculations and aggregations. Supported business intelligence efforts and financial analysis
tasks by creating queries that facilitated in-depth data exploration and reporting.
Implemented advanced data analysis techniques using Python and SQL, performing exploratory data analysis and
creating data visualizations to support financial decision-making and reporting. Applied analytical methods to
uncover insights and trends in financial data.
Created detailed data documentation for ETL processes, data models, and transformation workflows, providing
clear and comprehensive resources for ongoing maintenance and future development. Documented data
engineering practices and solutions to support team collaboration and project continuity.
Developed and deployed data quality frameworks and monitoring solutions for ETL processes, utilizing Python
and SQL to create automated validation, error detection, and resolution mechanisms. Ensured the accuracy and
reliability of financial data pipelines and reporting systems.
Collaborated with data scientists and analysts to define data requirements and develop solutions for financial
modeling and analysis. Translated business needs into technical specifications and delivered data solutions that
supported advanced financial analysis, risk management, and investment strategies.
Collaborated with cross-functional teams including data analysts, business stakeholders, and IT professionals to
gather requirements, define project scope, and prioritize features, ensuring that dashboards and reports met the
needs of diverse stakeholders and contributed to data-driven decision-making.
EDUCATION:
Bachelor of Electronics and Communication Engineering from Vardhaman college of Engineering – April 2017
Hyderabad, India
TECHNICAL SKILLS:
Reporting Tools: Tableau, Power BI, Microsoft Excel, SAP BusinessObjects, Cognos, Looker.
Data Modeling: Proficient in Star-Schema Modeling, Snowflake-Schema Modeling, FACT and dimension tables. Skilled in
Erwin for physical and logical data modeling. Experienced in utilizing Pivot Tables in Microsoft Excel for data analysis and
reporting.
Cloud Services: Well-versed in AWS (Amazon Web Services), Azure (Microsoft Azure), and GCP (Google Cloud Platform).
Database Optimization: Experienced in fine-tuning SQL queries and database configurations to optimize performance and
ensure data integrity.
Business Intelligence: Skilled in developing data models, ETL processes, and interactive dashboards using Power BI to
provide valuable insights for HR decision-making.
Languages: Proficient in Python, R, SQL, JAVA, .NET.
Data Analysis Methodologies: Well-versed in CRISP-DM (Cross-Industry Standard Process for Data Mining), Agile Data
Analysis, Lean Data Analysis, Six Sigma Data Analysis, Predictive Analytics, and Descriptive Analytics.
ETL Tools & Others: Proficient in Visual Studio, Informatica Power Center 8.x. Experienced in tools including Source
Analyzer, Target Designer, Mapping Designer, Workflow Manager, MS Office, Rational Requisite Pro, and SharePoint.