0% found this document useful (0 votes)
31 views33 pages

Coding: Development & Advanced Engineering Job Simulation: Sai Krishna Kaushik Paruchuri (1604-21-733-009)

Uploaded by

Anum Rida
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
31 views33 pages

Coding: Development & Advanced Engineering Job Simulation: Sai Krishna Kaushik Paruchuri (1604-21-733-009)

Uploaded by

Anum Rida
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 33

SUMMER INTERNSHIP REPORT

On

Coding: Development & Advanced Engineering Job


Simulation

Submitted By

SAI KRISHNA KAUSHIK PARUCHURI


(1604-21-733-009)

Computer Science and Engineering Department


Muffakham Jah College of Engineering and Technology
(Affiliated to Osmania University)

2024-2025

1
2
CERTIFICATE OF COMPLETION

This is to certify that the “INTERNSHIP REPORT ON CODING: DEVELOPMENT &


ADVANCED ENGINEERING JOB SIMULATION” submitted by ANUM RIDA bearing
Roll number 1604-21-733-009 in partial fulfillment of the requirements for the award of the
DEGREE OF BACHELOR OF ENGINEERING IN COMPUTER SCIENCE by the
OSMANIA UNIVERSITY, Hyderabad. This is a record of bonafide work carried out by him
under the guidance and supervision.

Course coordinator
Mrs. AFREEN SULTANA
Associate Professor
Department of CSE

3
DECLARATION

This is to certify that the work reported in the summer internship entitled “CODING:
DEVELOPMENT & ADVANCED ENGINEERING JOB SIMULATION” is a record
of work done by me in Department of Computer Science and Engineering, Muffakham Jah
College of Engineering and Technology, Osmania University. The report is based on the
project work done entirely by me and not copied from any other source.

By
Sai Krishna Kaushik Paruchuri
(1604-21-733-074)

4
ACKNOWLEDGEMENTS

I would like to express my sincere gratitude and indebtedness to the course


coordinator Mrs. Afreen Sultana for her valuable suggestions and interest throughout the
course of this project.

I am also thankful to the Head of the Department Dr. Syed Shabbeer Ahmad for providing
excellent infrastructure and a nice atmosphere for completing this project successfully.

I convey my heartfelt thanks to the lab staff for allowing us to use the required
equipment whenever needed.

Finally, I would like to take this opportunity to thank our families for their support
through the work. We sincerely acknowledge and thank all those who gave directly or
indirectly their support in completion of this work.

Sai Krishna Kaushik Paruchuri


(1604-21-733-074)

5
INDEX

Title Page No

Certificate 3

Declaration 4

Acknowledgments 5

Abstract 7

Chapters

1. Introduction 8

2. Objectives 9

3. Features 10

4. Technological Overview 12

5. Implementation 16

6. Appendices 21

7. Learning Experience 25

Conclusion 30

6
ABSTRACT

The internship titled "Data Visualisation: Empowering Business with Effective Insights"
provided an intensive, hands-on experience in applying advanced data analytics to address
real-world business challenges. The primary focus was to bridge the gap between theoretical
knowledge and industry practices by leveraging data-driven methodologies to support
strategic decision-making for an online retail store. The program emphasized data
preprocessing, visualization, and predictive analytics to uncover actionable insights and drive
organizational growth.

During the internship, I developed valuable expertise in SQL, Python, Tableau, and Power
BI, while integrating machine learning techniques to forecast revenue trends and analyze
critical performance metrics. Key responsibilities included rigorous data cleaning and
transformation, designing interactive dashboards, and implementing predictive models to
uncover seasonal revenue patterns, identify high-value customers and regions, and assess
product demand globally.

The experience underscored the importance of aligning technical outputs with business
objectives, ensuring that visualizations conveyed insights effectively to senior stakeholders.
Advanced coding techniques were employed to develop machine learning models and
implement complex data transformations, while dashboards were tailored to provide clarity
and actionable insights for strategic planning.

Additionally, the internship highlighted the significance of collaboration, analytical rigor, and
effective communication in professional environments. Regular interactions with mentors and
stakeholders enhanced my ability to present insights concisely and accurately, adhering to
industry preferred practices for data quality and visualization.

This report details the methodologies employed, challenges encountered, and solutions
implemented during the internship. It further demonstrates how this experience has enhanced
my proficiency in data analytics, predictive modeling, and visualization techniques, preparing
me to contribute effectively in dynamic, data-centric environments.

INTRODUCTION

The "Data Visualization and Predictive Analytics for Strategic Decision-Making"


internship was an invaluable opportunity to develop advanced technical skills in data
analysis, visualization, and predictive modeling. Hosted by Tata, this program emphasized
the practical application of analytical tools and techniques to solve real-world business
challenges, while reinforcing foundational concepts in data science and analytics. By
integrating theoretical knowledge with practical, hands-on tasks, the internship helped me

7
gain a comprehensive understanding of the role of data analytics in driving business strategy
and growth.

Throughout the program, the primary focus was on equipping participants with expertise in
using tools such as SQL, Python, Tableau, and Power BI to analyze and visualize complex
datasets. Emphasis was placed on data cleaning and transformation to ensure high-quality,
actionable analysis. In addition to these technical skills, I gained valuable experience in
creating visually compelling dashboards and predictive models, designed to address business
questions like revenue forecasting, customer segmentation, and regional demand analysis.
The use of predictive modeling allowed me to apply machine learning techniques to identify
patterns, forecast future trends, and assist in making data-driven decisions for the business.

During the internship, I worked on projects that involved applying data-driven methods to
derive insights and assist in decision-making processes. A critical aspect of the experience
was learning to present these insights in an easily digestible format for senior management,
ensuring they could make informed decisions based on accurate data.

This report offers a detailed account of the tasks undertaken, challenges faced, and solutions
implemented during the internship. It reflects how this hands-on experience significantly
enhanced my technical proficiency, analytical thinking, and readiness to navigate the ever-
evolving field of data analytics.

OBJECTIVES

2.1 Objectives of the Internship

Primary Objectives:

● To gain comprehensive expertise in industry-standard data analytics tools and


technologies, such as SQL, Python, Tableau, and Power BI, enabling the analysis of
large datasets and the extraction of actionable insights that can drive business
strategies.

● To develop proficiency in predictive modeling techniques, utilizing advanced


machine learning algorithms for forecasting business trends, with a particular
emphasis on accurate revenue prediction, demand forecasting, and trend analysis.

● To acquire a solid foundation in data cleaning, transformation, and preprocessing


techniques, ensuring data integrity and quality for meaningful analysis that drives
informed decision-making.

8
● To design and implement compelling data visualizations, utilizing effective charts,
graphs, and dashboards that simplify complex datasets, ensuring that business
stakeholders can easily interpret insights and make data-driven decisions.

● To apply analytical methodologies to real-world business scenarios, including


customer segmentation and revenue analysis, enabling companies to leverage data for
growth and operational efficiency improvements.

● To enhance communication skills by preparing and delivering clear, concise, and


impactful reports and presentations, providing senior management with actionable
insights that support strategic decision-making and business growth.

Secondary Objectives:

● To develop a deep understanding of how data-driven insights can shape and influence
business strategy, particularly within the retail and e-commerce sectors, driving
growth, operational optimization, and competitive advantage.

● To strengthen problem-solving capabilities by addressing complex analytical


challenges and implementing innovative solutions that improve business performance
and outcomes.

● To foster collaboration and teamwork skills by actively participating in feedback


sessions and presentations with colleagues, mentors, and business stakeholders,
ensuring alignment with project goals and objectives.

● To deepen knowledge of the critical role that data visualization and storytelling play
in effectively communicating insights, ensuring that complex data is presented in a
format that enables stakeholders to make informed decisions.

● To expand knowledge of industry best practices and emerging trends in data analytics,
machine learning, and data visualization, ensuring continuous growth and alignment
with the evolving technological landscape.

9
3. FEATURES

Throughout the “Data Visualisation: Empowering Business with Effective Insights”


internship, I focused on developing several essential features designed to enhance the
operational efficiency and decision-making processes of an e-commerce business. These
features were tailored to address critical business objectives, optimize system performance,
and enable data-driven insights to support strategic decision-making. The following are the
key features I worked on:

● Data Cleaning and Preprocessing using Python, SQL, and Power


BI/Tableau
The initial phase of the project focused on ensuring data integrity and quality.
I utilized Python and SQL to clean and preprocess the dataset, addressing
issues such as negative quantities and erroneous unit prices. Conditional
statements and SQL queries were employed to filter out invalid records,
ensuring that only clean data was used for analysis. After preprocessing,
Power BI/Tableau was used to visualize the cleaned dataset, facilitating a
more efficient and insightful analysis.
● 3.2Time Series Analysis for Revenue Trends

To address the Supervisor’s request for insights into revenue trends, I


conducted a time series analysis of the 2011 revenue data using Python.
Leveraging libraries such as Pandas and Matplotlib, I performed detailed
analysis and visualized the monthly revenue patterns. This analysis was
subsequently presented in Power BI/Tableau, where an interactive dashboard
was created to highlight seasonal trends, fluctuations, and underlying patterns.

10
These insights were critical for revenue forecasting and strategic decision-
making.

● Identification of Top 10 Revenue-Generating regions using SQL and


Power BI/Tableau
A key deliverable was identifying the top 10 regions by revenue. Using SQL, I
aggregated and ranked countries based on their total revenue and quantity
sold. This data was then visualized in Power BI/Tableau, where I developed an
interactive report that displayed both revenue and quantity sold for each of the
top-performing countries. This allowed the Marketing Supervisor to identify
high-performing markets, enabling targeted marketing efforts and strategic
expansion.
● Customer Segmentation by Revenue using Python and Power BI/Tableau
I performed customer segmentation by analyzing revenue data to identify the
top 10 customers by total spend. Using SQL to extract the relevant data and
Python’s Pandas library for aggregation and ranking, I then visualized the
results in Power BI/Tableau. The final visual displayed customers ranked by
revenue, providing the Marketing Supervisor with key insights to focus on
high-value customers and improve retention strategies.
● Demand Mapping by Region using SQL, Power BI/Tableau
To support the Supervisor’s expansion strategy, I conducted an analysis of
product demand across regions. Using SQL, I aggregated demand data by
region and product category, providing a clear picture of regional
performance. The data was then visualized using Power BI/Tableau, where
interactive maps and charts were created to display demand distribution. These
visuals enabled the Supervisor to identify high-demand regions, which were
essential for strategic planning and resource allocation.
● Predictive Modeling for Revenue Forecasting using Python
To provide actionable insights for future revenue generation, I developed a
predictive model using Python. Utilizing historical sales data and machine
learning algorithms such as linear regression, I forecasted revenue based on
key factors such as seasonality, customer behavior, and regional trends. The
model’s output was visualized in Power BI/Tableau, allowing the Supervisor
and Marketing Supervisor to analyze future revenue trends and make proactive
decisions for business growth.

● Advanced Data Visualization for Strategic Decision-Making using Power


BI/Tableau
I developed comprehensive dashboards and visualizations using both Power
BI/Tableau. In Power BI, I created interactive dashboards that tracked key
performance indicators such as revenue by region, customer segmentation, and
product demand. Tableau was used for in-depth trend analysis and to present

11
complex data in an easily interpretable format. These visualizations
empowered the Supervisor and Marketing Supervisor with the insights needed
for data-driven decision-making and effective expansion planning.
● Comprehensive Revenue and Demand Analytics

I conducted detailed analyses of the revenue and demand data across various
regions and product categories using SQL and Python. By aggregating and
analyzing the data, I identified key areas with significant revenue potential and
high product demand. The results were visualized in both Power BI/Tableau,
providing the Supervisor with a holistic view of the company’s performance
and areas ripe for expansion. These insights were crucial in formulating an
informed expansion strategy.

4. TECHNICAL OVERVIEW

The project was built to provide insights into business performance and opportunities,
focusing on key aspects such as revenue analysis, customer trends, and geographical demand.
The goal was to help the supervisors gain a clear understanding of the business landscape,
ensuring that the company is positioned for growth and can make data-driven decisions.
Below is a breakdown of the technical tools used throughout the project and how they were
applied to solve various business challenges. This project involved a combination of
programming languages, databases, and data visualization tools to clean, analyze, and present
actionable insights. The goal was to extract meaningful data from a large dataset, apply
business logic, and use visualization tools to communicate trends effectively. The following
tools were integral to the success of the project: Python, SQL, Power BI, and Tableau.

Python (for Data Analysis and Transformation)


12
Python was the programming language used for data transformation, analysis, and cleaning.
Its flexibility and rich ecosystem of libraries made it the perfect choice for handling data and
performing analysis.

● Data Cleaning and Transformation:


○ Pandas: Pandas was essential for manipulating and transforming the data into
the format required for analysis. Using Pandas, data was cleaned by removing
records with negative quantities and prices below zero, ensuring only valid
information was used in the analysis.
○ NumPy: NumPy was used to perform efficient numerical calculations and
handle large arrays of data. It supported data manipulation by offering high-
performance mathematical operations, such as calculating totals and averages.
● Feature Engineering:
○ Python was used to create new features that helped make sense of the data. For
example, total revenue was calculated by multiplying the quantity sold by the
unit price. This new feature allowed for more meaningful aggregation and
analysis in later stages.

● Data Aggregation:
○ Python was also used to group data by various dimensions, such as region or
customer, to calculate aggregated metrics like total revenue and units sold.
This provided a clearer picture of the data, helping the analysis become more
structured and easier to interpret.
● Data Filtering:
○ Conditional logic was applied to filter out invalid data, such as sales with
negative quantities. This process ensured that only relevant records were
included in the analysis, which directly improved the quality of the final
visualizations.

2. SQL (Structured Query Language)

SQL played a pivotal role in querying the data from relational databases. The ability to filter,
sort, and aggregate data using SQL allowed for quick access to the relevant parts of the
dataset.

● Data Extraction:
○ SQL was used extensively for querying large datasets. Custom queries were
written to filter out irrelevant data (e.g., removing records with null values or
invalid IDs) and to pull only the necessary fields for analysis, such as
customer IDs, quantities sold, and unit prices.
● Data Aggregation:
○ Using SQL functions like SUM() and GROUP BY, the total revenue and units
sold were calculated for each region, product, or customer. This allowed for
efficient aggregation of data and provided a solid foundation for further
analysis in Power BI and Tableau.
● Optimizing Performance:

13
○ SQL was key in optimizing data retrieval by using indexes and joins. This
ensured that the large dataset was processed quickly and that the extracted data
was clean, structured, and ready for analysis.

3. Power BI (for Interactive Dashboards)

Power BI was employed to build interactive, user-friendly dashboards that presented key
business metrics in a digestible format. It provided real-time insights, allowing for quick
decision-making based on the latest data.

● Data Integration:
○ Power BI allowed seamless integration with databases and Excel sheets, where
the cleaned data was imported directly. This streamlined the data import
process and made it easier to update dashboards as new data became available.
● Interactive Visuals:
○ A variety of visuals, such as bar charts, pie charts, and tables, were created to
display metrics like total revenue, top-selling products, and performance by
region. These visuals helped communicate trends effectively to stakeholders.
● Filtering and Slicing:
○ Power BI’s filtering and slicer functions allowed the CMO and CEO to drill
down into specific regions or customers. By adding filters, they could focus on
the most relevant information, such as the top 10 revenue-generating
customers, and make data-driven decisions about future strategies.
● Real-time Insights:
○ The interactive nature of Power BI dashboards made it easy to view and
explore the data in real-time. Stakeholders could adjust filters, view trends,
and generate reports dynamically, which proved useful during decision-
making meetings.

4. Tableau (for Advanced Data Visualization)

Tableau was used to create more detailed and visually sophisticated reports. Its advanced data
visualization capabilities helped convey insights through stunning, easy-to-understand
visuals.

● Advanced Data Visualizations:

14
○ In Tableau, we designed highly detailed visuals such as bar charts, heatmaps,
and geographic maps. These visuals were crucial for understanding
performance across different countries, regions, and customers.

● Geographical Mapping:
○ One of Tableau's standout features is its ability to create geographic maps. By
plotting sales data on a map, we could visually identify high-demand regions
and understand global performance trends. This was particularly helpful when
evaluating expansion opportunities across different regions.
● Storytelling with Data:
○ Tableau's ability to create dashboards that tell a "story" allowed for a more
compelling presentation of the findings. By combining multiple charts, filters,
and interactive elements, we could show how the data evolves over time,
helping to highlight key trends and opportunities.
● User Interactivity:
○ Tableau’s interactive dashboards allowed stakeholders to filter data by region,
customer, or time period, enabling a deeper exploration of the data. This
helped uncover deeper insights, such as which products were driving the most
revenue in different markets.

15
5. IMPLEMENTATION

Task 1: Top 10 Regions by Revenue Analysis

The first task involved analyzing the top 10 countries by revenue for the business. The
primary objective was to identify the regions where the company is performing the best, so
we could focus future marketing and operational efforts on these high-revenue countries.

Data Extraction & Preprocessing

The first step was extracting the necessary data from the database. Using SQL, we filtered
out certain regions from the list of countries, as it was excluded from the analysis. The query
selected country names and their respective total revenue, grouping the results by country and
ordering them in descending order of total revenue:

SELECT country, SUM(revenue) AS total_revenue


FROM sales_data
WHERE country != ‘regionname'
GROUP BY country
ORDER BY total_revenue DESC
LIMIT 10;

This query returned the top 10 countries by revenue, allowing us to identify the most
lucrative markets for the business.

Data Transformation & Validation in Python

Once the data was extracted, we used Python and the Pandas library to clean and validate it.
This step involved checking for missing or incorrect entries and ensuring that the revenue
values were formatted correctly as floating-point numbers:

import pandas as pd
df = pd.read_sql(query, connection)
df = df.dropna(subset=['total_revenue'])
# Ensure that revenue data is in the correct format (float)
df['total_revenue'] = df['total_revenue'].astype(float)

We also checked that the data types for each column were correct and ready for analysis.

16
Visualization in Tableau

With the cleaned data, we moved to Tableau for visualization. The data was imported into
Tableau, and a bar chart was created to represent the top 10 countries by total revenue. This
visualization provided a clear and impactful view of where the company was generating the
most income. By visualizing the data, it was easy to identify patterns, such as whether certain
regions consistently outperform others.

Task 2: Customer Segmentation: Top 10 Customers by Revenue

The second task focused on segmenting the customer base by revenue to identify the highest-
value customers. The objective was to tailor marketing strategies and special offers toward
these clients to increase loyalty and revenue.

Data Extraction & Preprocessing

We began by querying the database for customer names and their respective total revenue.
Similar to Task 1, we used SQL to filter out any missing or invalid customer records:

SELECT customer_name, SUM(revenue) AS total_revenue


FROM sales_data
WHERE customer_name IS NOT NULL
GROUP BY customer_name
ORDER BY total_revenue DESC
LIMIT 10;

This query helped us identify the top 10 customers by revenue, providing the raw data for
analysis.

Data Transformation & Cleaning in Python

In Python, we used Pandas to load the query results and performed additional cleaning to
handle any missing values in the customer or revenue fields:

df = pd.read_sql(query, connection)
df = df.dropna(subset=['customer_name', 'total_revenue'])
# Ensure the revenue data is correctly formatted
df['total_revenue'] = df['total_revenue'].astype(float)

17
By cleaning the data, we ensured that we were working with accurate and complete records.

Data Visualization in Tableau

In Tableau, we created a sorted column chart to display the top 10 customers based on total
revenue. The visualization made it easy to spot the highest revenue-generating customers at a
glance, which was helpful for sales and marketing teams to target these customers with
personalized promotions and offers.

Task 3: Regional Demand Analysis

The third task involved analyzing demand across different regions, excluding a few. By
identifying high-demand regions, we could better allocate resources and plan expansions into
areas with significant growth potential.

Data Extraction & Preprocessing in SQL

Using SQL, we extracted sales data for different regions, excluding the United Kingdom. We
aggregated the data by region and summed the total sales for each region:

SELECT region, SUM(sales) AS total_sales


FROM sales_data
WHERE region != 'United Kingdom'
GROUP BY region
ORDER BY total_sales DESC;

This query helped us extract the relevant data for the regional analysis.

Data Transformation & Cleaning in Python

After extracting the data, we used Python (Pandas) to clean the data, removing any rows with
missing or invalid values. We also ensured the sales data was correctly formatted for further
analysis:

df = pd.read_sql(query, connection)
df = df.dropna(subset=['region', 'total_sales'])
df['total_sales'] = df['total_sales'].astype(float)

18
The cleaned data was now ready for visualization.

Geospatial Visualization in Tableau

Using Tableau, we visualized the regional demand on a map. The regions were color-coded
based on total sales, highlighting areas with the highest demand. This visual representation
provided a quick and intuitive way to understand regional variations in demand, guiding
strategic decisions on resource allocation and market expansion.

Task 4: Revenue Trends Analysis for 2011

The fourth task focused on identifying revenue trends for the year 2011 by analyzing monthly
revenue patterns. Understanding these trends helped the business forecast revenue and plan
for seasonal fluctuations.

Data Extraction & Preprocessing in SQL

To gather the necessary data, we extracted monthly revenue data for 2011 using SQL:

SELECT MONTH(date) AS month, SUM(revenue) AS total_revenue


FROM sales_data
WHERE YEAR(date) = 2011
GROUP BY MONTH(date)
ORDER BY month;

This query gave us the total revenue per month in 2011, providing the foundation for trend
analysis.

Data Transformation & Cleaning in Python

After the data extraction, we used Python to clean and format the data, ensuring there were
no missing values and the revenue data was correctly structured:

df = pd.read_sql(query, connection)

19
df['total_revenue'] = df['total_revenue'].astype(float)

With the cleaned data, we were ready to perform trend analysis.

Visualization in Tableau

Using Tableau, we created a line chart that plotted total revenue for each month of 2011.
This visualization allowed us to identify trends, such as peak months (high season) and
months with lower revenue. This insight was invaluable for future sales forecasting and
planning.

Task 5: Predictive Modeling for Sales Forecasting (Optional)

While not explicitly required, we explored predictive modeling using a linear regression
model to forecast future sales trends based on historical data.

Data Preparation & Model Building in Python

We used scikit-learn to build a predictive model that could forecast future sales based on
past trends. First, we formatted the data into a time-series format. Then, we split the data into
training and testing sets and built a linear regression model:

from sklearn.linear_model import LinearRegression


from sklearn.model_selection import train_test_split

# Prepare data for modeling


X = df[['month']] # Features: months of the year
y = df['total_revenue'] # Target: total revenue

X_train, X_test, y_train, y_test = train_test_split(X, y,


test_size=0.2, random_state=42)

# Train the model


model = LinearRegression()

20
model.fit(X_train, y_train)

# Predict future sales


predictions = model.predict(X_test)

This model provided predictions for future revenue, allowing us to gauge potential sales
trends.

Visualization in Tableau

To compare the predicted sales with actual sales, we plotted both the real and predicted
values in Tableau. This side-by-side visualization allowed us to evaluate the accuracy of the
model and see how well it captured actual sales patterns.

Conclusion:

The analysis process involved multiple steps using SQL for data extraction, Python for data
cleaning and predictive modeling, and Tableau for creating impactful visualizations. These
tasks provided valuable insights into the business’s revenue performance, customer
segmentation, regional demand, and revenue trends. By integrating these tools effectively, we
ensured the data was clean, well-analyzed, and easily accessible for making data-driven
business decisions. The optional predictive modeling task further added a forward-looking
perspective, helping the company prepare for future trends.

21
6.APPENDICES

This section contains supplementary materials that provide deeper insights into the tools,
techniques, and processes used throughout the internship project. It includes code snippets,
data visualizations, and other resources that were critical in implementing the various tasks
discussed in the main report. The appendices are intended to offer additional context,
facilitate a better understanding of the methodologies employed, and provide transparent
access to the supporting materials that contributed to the success of the project.

6.1. Code Snippets

Below are some of the key code snippets from the project, showcasing the implementation of
critical functionality:

6.1.1. Segmentation based on Revenue:

This snippet calculates and ranks the top 10 countries by total sales revenue, and top 10
customers based on the same.

22
SELECT country,
SUM(revenue) AS total_revenue FROM sales_data WHERE country != 'United
Kingdom'
GROUP BY country
ORDER BY total_revenue
DESC LIMIT 10;

SELECT customer_name,
SUM(revenue) AS total_revenue FROM sales_data WHERE customer_name IS NOT
NULL
GROUP BY customer_name
ORDER BY total_revenue
DESC LIMIT 10;

6.1.2. Data Transformation using Python

This Python code snippet loads the SQL query results into a Pandas DataFrame, handles
missing data, and ensures the 'total_revenue' column is of the correct data type.
import pandas as pd

# Load the dataset


data = pd.read_csv('sales_data.csv')

# Clean and transform data


data['units_sold'] = data['units_sold'].fillna(0) # Filling missing values
with 0
data = data[data['country'] != 'United Kingdom'] # Exclude United Kingdom

# Standardize column names


data.columns = [col.strip().lower().replace(' ', '_') for col in
data.columns]

# Filter data for analysis


cleaned_data = data[['country', 'units_sold', 'revenue']]

23
6.2. Data Visualisations

Data visualizations played a crucial role in translating processed data into actionable insights.
Using Tableau and Power BI, a variety of interactive charts and dashboards were designed to
analyze trends, identify patterns, and highlight key metrics. These visualizations focused on
revenue distribution, regional performance comparisons, and other critical indicators,
enabling a clearer understanding of the data and supporting informed decision-making.

6.2.1. Region based Revenue Analysis

This visualization identifies the top 10 countries generating the highest revenue, excluding
the United Kingdom. It also includes data on the quantity of products sold alongside revenue
for each country, giving the supervisor a comprehensive view of market performance. This
information helps prioritize efforts in countries contributing significantly to overall revenue.

24
6.2.2. Customer based Revenue Analysis

This visual highlights the top 10 customers ranked by revenue contribution, starting with the
highest. The declining pattern ensures clear visibility of revenue distribution among these key
customers. The insights support the supervisor’s strategy to focus on retaining high-value
customers and boosting customer satisfaction.

25
6.2.3. Monthly revenue trends

This analysis focuses on revenue data for one year, broken down by month, to help the
supervisor identify seasonal trends and patterns. The visualization provides a granular view
of monthly revenue, enabling a deeper understanding of fluctuations and potential drivers of
these trends. The insights will aid in more accurate forecasting for the upcoming year.

26
6.2.4. Product Demand by Region(Demand Analysis)
This graph provides a snapshot of product demand across all regions in a single view. It
enables the supervisor to pinpoint regions with the highest demand, helping the company
identify potential expansion opportunities. The visual ensures that key demand areas are
clearly visible without requiring additional interaction or navigation.

27
6.3. Additional Notes

● The code snippets provided are designed to showcase key parts of the implementation
but may not represent the full scope of the project. The complete codebase includes
additional error handling, validations, and integrations.
● The data visuals serve as high-level representations of the business and are meant to
aid in understanding the project.

7.Learning experience

28
The internship at Accenture has been an invaluable experience in both developing my
technical skills and understanding the real-world applications of modern software
development practices. Below is a reflection on the key areas I learned during this period:

7.1. Reflection on the Skills Learned

Throughout my internship, I gained proficiency in several key technologies and practices that
are essential in the software development field. Among the skills learned, the most notable
include:

● Java: I deepened my understanding of Java, particularly in the context of backend


development using Spring Boot. Writing efficient and clean Java code, implementing
object-oriented principles, and applying best practices in designing scalable systems
was a key part of my daily tasks.
● Spring Boot: I became proficient in using Spring Boot for building Java-based
microservices. The framework’s simplicity, scalability, and ease of integration with
various components (such as databases and external services) made it a great tool to
work with.
● Agile Methodology: I was able to apply the principles of Agile methodology in real-
world development, participating in sprint planning, daily stand-ups, and retrospective
meetings. This helped me appreciate the iterative nature of Agile and its role in
maintaining a focus on customer needs while managing risks effectively.
● DevOps & Continuous Integration: Through the use of Jenkins for automating
builds, tests, and deployments, I learned the importance of continuous integration (CI)
in streamlining development workflows, ensuring faster release cycles, and
maintaining the quality of the application.
● Object-Oriented Programming (OOP): I had the opportunity to implement and
optimize OOP principles in various parts of the codebase, from the design of service
classes to the organization of business logic.
● Code Refactoring: I gained hands-on experience in improving legacy code through
refactoring techniques, which improved the readability, maintainability, and
efficiency of the application.

These skills have been instrumental in shaping my understanding of modern software


engineering practices.

7.2. Application of Agile Methodology in Real-world Projects

29
Agile methodology is widely used in the software industry, and this internship gave me
firsthand experience of how Agile works in practice. Through participating in weekly sprint
cycles, I was able to see how the iterative approach helped the development team respond
quickly to changes in requirements. Some key takeaways include:

● Collaboration: Agile encouraged constant collaboration with the team, product


owners, and stakeholders. This ensured that there was always a clear understanding of
project goals and priorities.
● Flexibility: Agile provided the flexibility to adapt to evolving requirements. Whether
it was tweaking the search functionality or refactoring legacy code, Agile allowed the
team to make iterative changes without major disruptions to the overall project.
● Time Management: The use of sprint cycles and regular retrospectives helped the
team manage time effectively, allowing for continuous progress while addressing any
blockers or challenges that arose.

The Agile approach significantly improved the productivity of the team and helped maintain
a focus on delivering valuable features to the client.

7.3. The Role of DevOps and Continuous Integration in Development

DevOps practices and continuous integration (CI) played a critical role in ensuring smooth
development and deployment processes. Using Jenkins for CI, I observed the following
benefits:

● Automated Testing: Jenkins was configured to run unit and integration tests
automatically every time new code was committed. This ensured that errors were
caught early, and quality was maintained throughout the development process.
● Faster Development Cycles: By automating the build, test, and deployment
processes, Jenkins enabled the team to release features more frequently and with
confidence. This was especially important in a fast-paced environment, where quick
iteration and deployment were essential.
● Collaboration Between Teams: DevOps practices helped foster better collaboration
between the development and operations teams. This led to more efficient workflows
and quicker problem resolution, ultimately contributing to the overall success of the
project.

DevOps and CI practices helped reduce friction in the development process, ensuring that the
team could maintain high productivity and deliver features rapidly.

30
7.4. Importance of Object-Oriented Programming (OOP) in Software Design

Object-Oriented Programming (OOP) was crucial in the development of the application. By


using concepts such as inheritance, polymorphism, encapsulation, and abstraction, I was able
to:

● Organize Code Efficiently: OOP helped in organizing the application’s logic into
smaller, reusable components. For example, the search functionality was abstracted
into a service class, which could be used by both the user-facing search feature and
the report generation function.
● Promote Code Reusability: By following OOP principles, I created modular classes
that could be reused in different parts of the application. This helped reduce code
duplication and made the codebase more maintainable.
● Maintain Code Quality: OOP’s focus on clean, well-structured code ensured that the
application was easier to maintain, update, and scale. The use of classes and objects
made the codebase intuitive to understand for new team members or external
collaborators.

The adoption of OOP allowed for a clean, modular design that was flexible and easy to
maintain over time.

7.5. Practical Knowledge of Software Development Lifecycle (SDLC)

Understanding the Software Development Lifecycle (SDLC) was crucial in gaining a


complete picture of how software is developed from start to finish. I had the opportunity to
work across various stages of the SDLC, including:

● Planning: During sprint planning meetings, we defined requirements, identified


potential risks, and broke down tasks for the upcoming sprint. This helped us stay
organized and prioritize work effectively.
● Development: I contributed to the development process by writing code for new
features, refactoring old code, and resolving issues that arose during development.
● Testing: Testing was integrated early in the process through the use of automated
tests. This ensured that the code was reliable and that new features didn’t break
existing functionality.
● Deployment: Continuous integration and deployment processes ensured that the
application could be deployed efficiently and with minimal downtime.

The SDLC provided a structured approach to developing software that helped manage time,
resources, and deliverables effectively.

31
7.6. Significance of Spring Boot for Java Services

Spring Boot played an essential role in simplifying the development of the backend services
for the e-commerce application. Some of the reasons why Spring Boot was invaluable
include:

● Simplified Configuration: Spring Boot’s auto-configuration capabilities significantly


reduced the setup time. I didn’t need to manually configure the entire application;
Spring Boot handled most of it automatically.
● Microservices Architecture: Spring Boot is well-suited for building microservices,
which was key for the e-commerce application. Each part of the application could be
developed as a separate service, making it easier to scale and maintain.
● Integration with Databases: Spring Boot made it easy to integrate with relational
databases using Spring Data JPA, enabling efficient queries to retrieve products for
search functionality.
● Security: Spring Boot provided tools for securing the application, including handling
authentication and authorization, which was important for building a secure e-
commerce platform.

Spring Boot streamlined development by providing tools and conventions that made backend
service creation faster, more consistent, and easier to manage.

32
CONCLUSION

The internship at Accenture proved to be an enriching and valuable experience, offering the
opportunity to apply academic knowledge in real-world projects. During this period,
involvement in addressing challenges faced by a client in the e-commerce sector provided
insights into the importance of clean, maintainable code, effective collaboration in an Agile
environment, and continuous improvement in development processes.

Key tasks such as implementing search functionality, refactoring legacy code, and optimizing
services for efficiency allowed for a deeper understanding of the software development
lifecycle, Agile methodology, and the significance of continuous integration in maintaining
code quality.

The internship not only helped develop technical skills but also improved problem-solving,
teamwork, and communication abilities. The guidance and support from the Accenture team
played a crucial role in ensuring the success of the internship.

Overall, this experience has equipped the intern with practical knowledge and skills that will
be valuable in a future career in software engineering. The lessons learned during this period
will undoubtedly contribute to continued growth and success in the industry.

33

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy