Apoorva CSITSS
Apoorva CSITSS
The algorithm of the proposed model, along with its Year : Year that student completed education. Data
computational processes for predicting placement chance is collected were from 2006-2015.
outlined below: Reg-no : Register number of the student. It takes any
Step 1: Data collection. integer values.
The goal is to find the proficient students in the college under Branch : represents the name of the Branch. It can take only
consideration viz., XX for the year 2016. In this college there text values ranging from A-Z
were 1,434 students. These students hailed from various Percent : various marks scored by student in subjects. It can
courses that were operative in the college. The courses are take only the numeric values from 0 to 100.
MBA, MCA, BCA, B.Com, and BBA. Skills : it shows the overall Skills of the student.. It can take
Step 2: Predict Placement chance. only the numeric values from 0 to 10.
This step predicts placement chance of the student and also Effective-score: it shows the overall performance of the
percentage placement of institution using Proposed student Formula to calculate Effective-score is as follows
Classification algorithm. Effective score = percent + skills * 10 It can take only the
Step 3: Evaluate the result numeric values from 0 to 200.
Placed : Placed based on student performance. Value is
taken in the form of Yes\No,
IF Yes Table II represents output of the clustering algorithm which is
Student placed, used as the input to the proposed algorithm with the attributes
ELSE as shown above.
Student has not been placed.
For each student in the selected clusters the
VI. PROPOSED ALGORITHM.
following operations are performed. Store the
An algorithm is proposed to achieve the objective of study. student in a variable ‘S'.
The algorithm is as follows.
Input : current Student, oldStudentList 1 if x = y
Output : Placement chance f(x,y) =
0 otherwise
1. Read Student
2. Read oldStudentList
3. Select count all students in oldStudentList having n
score = score of current Student. C = ∑ f(x,y)
Store it as count Selected x=1
4. Select count all placed students in oldStudentList Where x is effective score of each historic student data and y is
having score = score of current Student. effective score of current student and C is count of Selected
Store it as count Placed And Selected Historic Data
5. If countSelected == 0 then Count all the effective scores of historic data same as effective
6. probability =0.5 score ‘S'.
7. else Illustration:
8. Calculate chance = count Placed And Selected/count If the effective score of current student is 104 then this value
Selected will be searched in the historic data and returns the count of
[End If at Step 5] such data.
9. If chance >=0.4 ( i.e Excellent/Good/Average) ex: c =
10. set placement chance Good f(72,104)+f(121,104)+f(146,104)+f(106,104)+f(104,104)+f(10
else 0,104)+f(165,104)+f(83,104)+f(129,104)+f(110,104)
11. set placement chance Bad =0+0+0+0+1+0+0+0+0+0
12. [End If at Step 9] =1
13. Write placement chance Step 4 of the algorithm.
1 if x = y &x(placed) = true
f(x,y) =
VII. EXPERIMENTAL EVALUATION 0 otherwise
Step : 9, 10 and 11
TABLE IV : Placement Chance for input data Fig 2: screen shot of placement chance prediction in
reg_ branch effective_s centroids of Placement Chance
percentage.
no core cluster
1 MCA 53 55.33 Bad (Not Selected) The above screenshot represents percentage placements of the
institution considered.
2 MCA 72 55.33 Bad (Not Selected)
3 MCA 110 105.0 Bad
4 MCA 41 55.33 Bad (Not Selected)
5 MCA 129 146.66 Good VIII. RESULTS
6 MCA 146 146.66 Good
7 MCA 106 105.0 Bad CONFUSION MATRIX
8 MCA 100 105.0 Good
9 MCA 104 105.0 Good Data mining algorithms like decision tree, Naïve
10 MCA 165 146.66 Good bayes, neural network and proposed algorithm were
applied on the same dataset and the tests were
let us explain the first instance of the table IV.Since conducted separately. Results obtained after the
the centroid of the reg_no1 falls in the eliminated tests for each algorithm were modeled as confusion
cluster in the module1, this student will not be matrix.
placed. similarly the student with reg_no 2 won't be
selected. consider student with reg_no 3.The
TABLE V : Comparison of Proposed algorithm with with unique characteristics that require different
other algorithms approaches for solving the problem. In this study, A
model was proposed along with a algorithm. This
Algorithm Accuracy TPR Precision was compared with three other classification
algorithms such as decision tree, naïve bayes, and
Decision Tree 0.84 0.95 0.74
neural network in terms of accuracy, precision, true
Naïve Bayes 0.87 0.93 0.78
positive rate(recall).The proposed model, proved
to be the best predicting model for solving
Neural Networks 0.83 0.88 0.69 placement chance prediction problems compared to
all other algorithms. Hence, having the information
Proposed 0.92 0.96 0.87 generated through our study, institution would be
Algorithm
able to design strategies to overcome lacunae and
improve placements with best chances of getting
The above table V gives accuracy, true positive rate, placed. Thus admission can be increased.
false positive rate, true negative rate, false negative
rate and precision of different algorithms compared REFERENCES
with the proposed algorithm. Precision and
accuracy of the proposed algorithm is high [1] Jae H. Mina, Young-Chan Leeb, “Bankruptcy prediction using support
vector machine with optimal choice of kernel function parameters”,
compared with other classifying algorithms. The Volume 28, Issue 4, May 2005, Pages 603–614.
false negative rate of the proposed algorithm is low [2] J.A.K. Suykens and J. Vandewalle, “Least Squares Support Vector
against all other algorithms. Machine Classifiers:” Volume 308, Issue 2, 27 April 2001, Pages 397–
407.
[3] Tung-Kuang Wu, Shian-Chang Huang:”Evaluation of ANN and SVM
1 classifiers as predictors to the diagnosis of students with learning
Decision tree disabilities”, Volume 34, Issue 3, April 2008, Pages 1846–1856.
0.8
[4] Guha, S.; Rastogi, R.; Kyuseok Shim “ROCK: a robust clustering
0.6 algorithm for categorical attributes”, Pages 512 – 521.
0.4 Naïve Bayes [5] KakotiMahanta, Arun K. Pujari,” QROCK: A quick version of the
ROCK algorithm for clustering of categorical data”,Volume 26, Issue
0.2 15, November 2005, Pages 2364–2373.
0 Neural [6] Agnieszka Prusiewicz, “MaciejZiębaServices Recommendation in
Network Systems Based on Service Oriented Architecture by Applying Modified
ROCK Algorithm” Volume 88, 2010, pp 226-238.
Proposed [7] Christian Borgelt, “An implementation of the FP-growth algorithm”,
Algorithm Pages 1 – 5, 2000.
[8] BalazsRacz, D: An FP-Growth Variation without Rebuilding the FP-
Tree”.
Fig 3:Comaparison of algorithms with Proposed algorithm [9] Ke Wang, Liu Tang, Jiawei Han, “Junqiang Liu “Top down FP-Growth
for Association Rule Mining”, Volume 2336, 2002, pp 334-340.
[10] SudheepElayidom, Suman Mary Idikkula& Joseph Alexander “A
The above graph represents accuracy, recall, Generalized Data mining Framework for Placement Chance Prediction
precision of various classification algorithms. Problems” International Journal of Computer Application (0975-8887)
Volume 31- No.3, October 2011.
Proposed algorithm has the highest precision, [11] Ajay Kumar Pal, Saurabh Pal “Classification Model of Prediction for
accuracy and recall. Decision tree, Naïve bayes, Placement of students” I.J.Modren Education and Computer Science,
2013, 11, 49-56.
Neural network and proposed algorithm is [12] K. Pal, and S. Pal, “Analysis and Mining of Educational Data for
represented by blue, red, green and purple Predicting the Performance of Students”, (IJECCE) International Journal
of Electronics Communication and Computer Engineering, Vol. 4, Issue
respectively. 5, pp. 1560-1565, ISSN: 2278-4209, 2013.
[13] B.K. Bharadwaj and S. Pal. “Mining Educational Data to Analyze
IX. CONCLUSION Students' Performance”, International Journal of Advance Computer
Science and Applications (IJACSA), Vol. 2, No. 6, pp. 63-69, 2011.
Data mining techniques applied on educational data [14] S. K. Yadav, B.K. Bharadwaj and S. Pal, “Data Mining Applications: A
comparative study for Predicting Student's Performance”, International
in concerned with developing methods for exploring Journal of Innovative Technology and Creative Engineering (IJITCE),
the unique types of data; in educational domain Vol. 1, No. 12, pp. 13-19, 2011.
each educational problem has specific objectives