
2 RELATED WORKS
Gupta et al. devise a web system that manages
placements using Random Forest Regressor to
estimate the probability of placement based on
academic and skill-based variables. This function
generates recommendations and newborn recruitment
filters, and will be developed in the future by
increasing the diversity of the dataset and implement
a tracker in real time.
Kumar et al. Build a Placement Predictor
Algorithm using Machine Learning Techniques for
Predicting Placement Probability by analyzing
Student Data Sets. Students can spot skill gaps using
the model, and future updates are set to focus on real-
time analytics and better precision.
Jeganathan et al. in which a Fuzzy Inference
System is used to classify students into various
placement categories to manage their training records
efficiently. Future work to train hybrid models may
lead to further improvements.
Shahane et al. analyze past placement data with
machine learning models, achieving 95.34%
accuracy using Logistic Regression. Future work
explores deep learning techniques and cross-
validation strategies for better reliability.
Thangav et al. propose a rule-based placement
predictor for B.Tech students, achieving 71.66%
accuracy. Future enhancements include refining
classification techniques and integrating deep
learning for improved prediction.
Manoj et al. use XGBoost to predict placements
and classify students based on academic and technical
skills. The system aids targeted training, with future
work addressing bias, university ranking impact, and
FAANG job predictions.
Ramaswamy et al. present a brain tumor detection
model using a modified Link-Net with SE-
ResNet152, achieving 99.2% accuracy. Future work
focuses on improving feature fusion and integrating
additional pre-trained models.
Ramaswamy et al. also propose an Optimized
Gradient Boosting model for Type-2 Diabetes
Mellitus detection, achieving 94.5% accuracy. Future
improvements include additional clinical features and
advanced ensemble techniques.
Eswara et al. develop a placement prediction
system using XGBoost on synthetic datasets,
outperforming standard classifiers. Future work
explores industry trends, alumni feedback, and wider
institutional testing.
Saritha et al. compare Naïve Bayes, Random
Forest, and Decision Trees for placement prediction,
demonstrating effectiveness with varying accuracy.
Future enhancements include deep learning
integration and additional predictive features.
Jayashre et al. optimize campus placement and
salary prediction using multiple ML models, with
Logistic Regression achieving 84% accuracy. Future
work involves expanding datasets and refining
predictive algorithms.
Kadu et al. propose a Student Placement
Prediction and Skill Recommendation System using
Random Forest and cosine similarity for personalized
recommendations. Future work focuses on dataset
expansion and deep learning integration.
3 METHODOLOGY
The methodology for the placement prediction and
optimization system involves a structured, multi-
stage approach integrating machine learning, data
analytics, and cloud-based deployment. The
framework is designed to provide accurate placement
predictions, enhance recruitment efficiency, and
streamline job matching processes. The key phases of
the methodology are detailed below.
3.1 Data Collection and Preprocessing
The system utilizes historical placement records,
including academic performance (CGPA), number of
internships, backlogs, skill sets, and prior job
application outcomes. The dataset undergoes rigorous
preprocessing steps, including:
• Data Cleaning: Handling missing values,
removing inconsistencies, and normalizing
numerical attributes.
• Feature Engineering: Extracting key features
such as skill relevance scores and internship
impact metrics.
• Data Splitting: Dividing the dataset into
training 80% and testing 20% subsets to
evaluate model performance.
3.2 Placement Prediction Model
A supervised machine learning model is trained to
assess the probability of student placement based on
historical data. The model development follows these
stages:
• Algorithm Selection: Multiple models,
including Random Forest, XGBoost, and
Neural Networks, were evaluated, with the
final model achieving an accuracy of 87.4%.
ICRDICCT‘25 2025 - INTERNATIONAL CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION,
COMMUNICATION, AND COMPUTING TECHNOLOGIES
84