INRTERNSHIP
INRTERNSHIP
(Project Work)
On
Submitted By
VUSA MAHENDRA - (21691A3252)
1
MADANAPALLE INSTITUTE OF TECHNOLOGY & SCIENCE
(UGC-AUTONOMOUS INSTITUTION)
Affiliated to JNTUA, Ananthapuramu & Approved by AICTE, New Delhi
NAAC Accredited with A+ Grade, NIRF India Rankings 2022 - Band: 251-300 (Engg.)
NBA Accredited - B.Tech. (CIVIL, CSE, CST, ECE, EEE, MECH), MBA & MCA
www.mits.ac.in
BONAFIDE CERTIFICATE
This is to certify that the 20CSD702-Project Work & Internship entitled “TRAINING
PROGRAM RECOMMENDATION SYSTEM ” is a bonafide work carried out by
Submitted in partial fulfillment of the requirements for the award of degree Bachelor of
Technology in the stream of Computer Science & Engineering (Data Science) in
Madanapalle Institute of Technology & Science, Madanapalle, affiliated to Jawaharlal
Nehru Technological University Anantapur, Ananthapuramu during the academic year
2024-2025.
ii
CERTIFICATE
iii
DECLARATION
Date :
Place : Madanapalle
PROJECT MEMBER
VUSA MAHENDRA - (21691A3252)
I certify that above statement made by the student is correct to the best of my
knowledge.
Date : Signature
iv
TABLE OF CONTENTS
2. SYSTEM ANALYSIS 4
2.1 Introduction 5
2.2 Existing System 5
2.3 Disadvantages of Existing System 6
2.4 Proposed System 6
2.5 Advantages over Existing System 7
3. SYSTEM SPECIFICATION 9
3.1 Hardware Requirement Specification 10
3.2 Software Requirement Specification 10
4. SYSTEM DESIGN 12
4.1 System Architecture 13
4.2 Modules Flow Diagrams 14
v
5.3 Method of Implementation (CODING) 19
5.4 Output Screens and Result Analysis 20
7. CONCLUSION 27
7.1 Conclusion 28
REFERENCES 29
vi
ABSTRACT
The technical framework incorporates data from faculty and training program
datasets, with preprocessing steps to ensure data consistency. The recommendation
model combines memory-based and model-based collaborative filtering approaches,
utilizing K-Nearest Neighbors (KNN) and Singular Value Decomposition (SVD)
algorithms to deliver accurate and scalable recommendations. Performance is
assessed using precision, recall, MAE, and RMSE metrics. This system ultimately
aims to foster a continuous, targeted professional development culture within
NITTTR, supporting faculty expertise and enriching the learning environment.
vii
List of Figures
viii
LIST OF ABBREVIATIONS
ix
CHAPTER 1
INTRODUCTION
1
1. INTRODUCTION
2
Through collaborative filtering techniques, the system delivers peer-informed
recommendations that promote a collaborative learning environment within
departments. By analyzing training trends among colleagues in similar fields, it
encourages faculty members to pursue programs that align with both personal and
departmental objectives. Ultimately, the system aims to improve faculty expertise,
thereby enhancing their teaching and research capabilities while contributing to
NITTTR’s overarching goal of advancing educational excellence.
1.4Limitations of Project:
3
CHAPTER 2
SYSTEM ANALYSIS
4
2.1 Introduction
5
Furthermore, there is little to no tracking of faculty development goals in the current
system, which means that choices may not consistently align with the institution’s
vision for targeted faculty growth and inter-departmental knowledge enhancement.
6
aligning training recommendations with faculty members' professional goals, thereby
enriching the overall learning and teaching environment within NITTTR.
7
and career development needs.
Collaborative Learning Opportunities: The system promotes peer-based
recommendations, suggesting programs that other faculty members in the same
branch have completed. This encourages shared learning experiences, enabling
faculty to benefit from similar courses and to foster professional growth
through mutual knowledge sharing.
Streamlined Faculty Development: With a data-driven recommendation
system, NITTTR can better align faculty training with departmental and
institutional objectives. By promoting training programs that support
departmental growth, the system ensures that professional development aligns
with institutional goals, promoting an integrated learning environment.
Improved Course Discovery: The automated approach helps faculty discover
new and relevant courses that they might not have considered otherwise. By
recommending courses based on collaborative filtering, the system opens up
additional pathways for faculty to explore and expand their skill sets.
Data-driven decision-making: The proposed system leverages insights from
faculty data to offer strategic recommendations, contributing to more informed
decision-making around professional development. This ensures that faculty
training choices are not only aligned with personal growth but also support
broader institutional strategies.
8
CHAPTER 3
SYSTEM SPECIFICATION
9
3.1 Hardware Requirement Specification
10
Figure 3.2.1 Required libraries
o Data Processing Libraries:
Pandas: For data manipulation and handling of structured
datasets.
NumPy: For numerical operations and matrix manipulations
required in data preprocessing and similarity calculations.
o Machine Learning Libraries:
Scikit-Learn: For data splitting, normalization, and machine
learning utilities.
Surprise Library: For building collaborative filtering
recommendation algorithms like K-Nearest Neighbors (KNN) and
Singular Value Decomposition (SVD).
o Matplotlib / Seaborn: For data visualization, allowing visual insights
into data distributions and model performance.
Database:
o Excel or SQLite or MySQL: For storing user and training program data.
11
CHAPTER 4
SYSTEM DESIGN
12
4.1 System Architecture
13
Collaborative Filtering Models:
o This is the core layer where recommendation models are applied, using
both Memory-Based and Model-Based Collaborative Filtering
approaches:
Memory-Based Filtering: Uses similarity calculations between
users or items to recommend relevant courses (User-based and
Item-based).
Model-Based Filtering: Applies Singular Value Decomposition
(SVD) for dimensionality reduction, ensuring efficient
recommendations for larger datasets.
Evaluation Layer:
o The output recommendations from the models are evaluated using
metrics like Precision, Recall, Mean Absolute Error (MAE), and Root
Mean Squared Error (RMSE). This ensures the system’s performance
meets expected standards.
Output Layer:
o The final recommendations are provided to faculty members, offering
personalized course suggestions that align with their individual learning
preferences and professional development goals.
The module flow diagrams represent the sequential steps within each layer of
the system, providing a visual representation of the entire recommendation process.
Flow Diagram Descriptions:
Data Collection Module:
o Inputs: Faculty data and training program data.
o Process: Collects, organizes, and stores user and program details.
o Output: Structured datasets ready for preprocessing.
Data Preprocessing Module:
o Inputs: Raw data from the collection layer.
o Process: Applies one-hot encoding, normalization, and standardization
to prepare data for model training.
o Output: Clean, structured data ready for model input.
14
Data Splitting Module:
o Inputs: Preprocessed data.
o Process: Divides data into training and testing sets for model validation,
ensuring data is representative of real-world scenarios.
o Output: Training and testing datasets for collaborative filtering models.
Collaborative Filtering Models Module:
15
o Process: Assesses recommendation accuracy using metrics like
precision, recall, MAE, and RMSE.
o Output: Performance scores for each model.
Recommendation Output Module:
o Inputs: Evaluated recommendations.
o Process: Presents final recommendations to faculty in a clear and user-
friendly manner.
o Output: Personalized course recommendations for faculty professional
development.
16
CHAPTER 5
17
5.1 Introduction
18
5.3 Method of Implementation (Coding)
19
Below is a sample of the coding approach used to implement the core functions:
20
5.4 Output Screens and Result Analysis
22
CHAPTER 6
6.1 Introduction
The Testing and Validation phase ensures that the system meets its functional
23
requirements, providing personalized training program recommendations that align
with user preferences and needs. This section covers the process of designing test
cases, evaluating model performance, and validating recommendations using
appropriate metrics. Effective testing and validation contribute to the reliability and
usability of the recommendation system in real-world scenarios.
Test cases are designed to verify that the system performs accurately across
various functions and scenarios, including data processing, model performance, and
recommendation accuracy. Here are key test cases for the recommendation system:
1. Data Integrity Test:
o Objective: Ensure that data from both participants_df and dept_df is
correctly loaded, cleaned, and preprocessed.
o Test Steps: Verify that categorical data is encoded, numerical data is
normalized, and data types match across merged columns.
o Expected Outcome: Data integrity is maintained with no missing or
mismatched values.
2. Merge Validation Test:
o Objective: Test the successful merging of participants_df and dept_df
based on course_id, department, and courses_attended.
o Test Steps: Confirm that merged DataFrame includes only relevant
columns and matched rows from both datasets.
o Expected Outcome: Accurate merging without empty rows or incorrect
matches.
3. User-Based Collaborative Filtering Test:
o Objective: Verify that the user-based collaborative filtering model
recommends relevant courses based on user similarities.
o Test Steps: Select a test user, compare their actual course history to the
model’s recommendations, and verify that similar users’ choices
influence suggestions.
o Expected Outcome: Recommended courses align with the user’s
interests and prior course completion patterns.
4. Item-Based Collaborative Filtering Test:
o Objective: Ensure that item-based collaborative filtering accurately
recommends similar courses.
24
o Test Steps: Check that the model identifies similar courses based on
other users’ ratings and course characteristics.
o Expected Outcome: Recommended courses have content or
characteristics similar to the user’s previously attended courses.
5. Model Evaluation Metrics Test:
o Objective: Confirm that model performance metrics (e.g., Precision,
Recall, MAE, RMSE) are correctly calculated.
o Test Steps: Run the model evaluation functions and check metric
outputs for each recommendation model.
o Expected Outcome: Evaluation metrics accurately reflect model
performance, with meaningful values for each metric.
6. Cold Start Test:
o Objective: Test system performance when recommending courses for
new users or users with limited data.
o Test Steps: Use a new faculty profile with minimal data and assess if the
system recommends general or popular courses from the same
department.
o Expected Outcome: The system provides relevant courses based on
general trends in the user’s department.
6.3 Validation
25
recommended items.
o Formula:
2. Recall:
o Measures the proportion of relevant recommended items among all
relevant items in the test set.
o Formula:
26
4. Root Mean Squared Error (RMSE):
o Assesses the model’s accuracy by measuring the square root of the
average squared differences between predicted and actual ratings.
o Formula:
27
5. Coverage:
o Measures the diversity of recommendations, indicating the proportion of
items in the dataset that are recommended.
o Formula:
28
CHAPTER 7
CONCLUSION
29
7.1 Conclusion
30
REFERENCES
Yehuda Koren, Robert Bell, and Chris Volinsky, "Matrix Factorization Techniques
for Recommender Systems," IEEE Computer Society, 2009.
31
Michael D. Ekstrand, John T. Riedl, and Joseph A. Konstan, "Collaborative Filtering
Recommender Systems," Foundations and Trends in Human-Computer Interaction,
2011.
32