1. Trang chủ
  2. » Tất cả

Applying bayesian neural network to evaluate the influence o f specialized mini projects on final perform ance o f engineering students a case study

6 6 0

Đang tải... (xem toàn văn)

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Tiêu đề Applying Bayesian neural network to evaluate the influence of specialized mini projects on final performance of engineering students: A case study
Tác giả Minh Truong Nguyen, Viet-Hung Dang, Truong-Thang Nguyen
Trường học University of Sciences, Vietnam National University, Hanoi
Chuyên ngành Computational Science
Thể loại Research article
Năm xuất bản 2022
Thành phố Hanoi
Định dạng
Số trang 6
Dung lượng 2,46 MB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

■ MATHEMATICS AND COMPUTER SCIENCEI COMPUTATIONAL SCIENCE, PHYSICAL SCIENCES I ENGINEERING DOI: 10.31276/VJSTE.644.10-15Applying Bayesian neural network to evaluate the influence o f

Trang 1

MATHEMATICS AND COMPUTER SCIENCEI COMPUTATIONAL SCIENCE, PHYSICAL SCIENCES I ENGINEERING DOI: 10.31276/VJSTE.64(4).10-15

Applying Bayesian neural network to evaluate

the influence o f specialized mini projects on final

perform ance o f engineering students: A case study

‘University o f Sciences, Vietnam National University, Hanoi 2Faculty o f Building and Industrial Construction, Hanoi University o f Civil Engineering

Received 10 June 2022; accepted 8 September 2022

A bstract:

In this article, deep learning probabilistic models are applied to a case study on evaluating the influence of specialized mini projects (SMPs) on the performance o f engineering students on their final year project (FYP) and cumulative grade point average (CGPA) This approach also creates a basis to predict the final performance

of undergraduate students based on their SMP scores, which is a vital characteristic o f engineering training The study is conducted in two steps: (i) establishing a database by collecting 2890 SMP and FYP scores and the associated CGPA o f a group o f engineering students that graduated in 2022 in Hanoi; and (ii) engineering two deep learning probabilistic models based on Bayesian neural networks (BNNs) with the corresponding architectures o f 8/16/16/1 and 9/16/16/1 for FYP and CGPA, respectively The significance o f this study is that the proposed probabilistic models are capable o f (i) providing reasonable analysis results such as the feature importance score o f individual SMPs as well as an estimated FYP and CGPA; and (ii) predicting relatively close estimations with mean relative errors from 6.8% to 12.1% Based on the obtained results, academic activities to support student progress can be proposed for engineering universities.

Introduction

N ow adays, universities are capable o f collecting data w ith

reference to their students in electronic format A s a result,

there is an urgent need to effectively transform large volumes

o f data into know ledge to im prove the quality o f m anagerial

decisions and to predict academic perform ance o f students

at an early stage As a part o f artificial intelligence (AI)

techniques recently adopted in a w ide variety o f hum an life

applications [1,2], various m achine learning (M L) approaches

have been increasingly applied to analyse educational data,

such as student scores, to concentrate academ ic assistance

on students as w ell as to im prove the university training

program s ML is an especially appealing alternative in the

field o f engineer training and education as it is difficult or

unfeasible to develop conventional algorithm s to perform

required tasks [3 ,4 ],

S.S Abu-Naser, et al (2015) [4] developed an artifical

neural netw ork (ANN) m odel for predicting student

perform ance at the Faculty o f Engineering and Inform ation

Technology, A l-A zhar University, based on the registration records o f 1407 students using a feed forw ard back propagation algorithm for training The m odel w as tested

w ith an overall result o f 84.6% E.Y Obsie, et al (2018) [5] developed a neural netw ork m odel for predicting student cum ulative grade point averages for the 8th sem ester (CGPA8) and designed an application based on the predictive models The real dataset em ployed in the study was gathered from

134 students at the H aw assa U niversity School o f Com puter Science that graduated in 2015, 2016, and 2017 It is shown that the student progress perform ance, w hich is m easured

by CGPA8, can be predicted using scores from their first-, second-, and third-year courses Z Iqbal, et al (2017) [6] utilized collaborative filtering (CF), m atrix factorization (M F), and restricted Boltzm ann m achine (RBM ) techniques

to system atically analyse real-w orld data collected from 225 undergraduate students enrolled in the Electrical Engineering program at the Inform ation Technology U niversity (ITU) from w hich the academ ic perform ance o f the ITU students was evaluated It was shown that the RBM technique was

'Corresponding author: Email: thangnt2@huce.edu.vn

I 1 ■.!rral ntScitnct.

" .> jt: t ncintt-rin"

D E C E M B E R 2 0 2 2 • V O L U M E 6 4 N U M B E R 4

Trang 2

MATHEMATICS AND COMPUTER SCIENCE | COMPUTATIONAL SCIENCE, PHYSICAL SCIENCES I ENGINEERING m

better than the other techniques in predicting student

perform ance in the particular course S.D.A Bujang, et

al (2021) [7] introduced a com prehensive analysis o f

m achine learning techniques to predict final student grades

in first sem ester courses by im proving the perform ance o f

predictive accuracy The perform ance accuracy o f six well-

know n m achine learning techniques, namely, decision tree

(J48), support vector m achine (SVM ), naive bayes (NB),

K-nearest neighbour (K-NN), logistic regression (LR), and

random forest (RF) using 1282 student course grades were

presented and follow ed by a m ulticlass prediction m odel to

reduce the over-fitting and m isclassification results caused by

im balanced m ulti-classification using the Synthetic M inority

O versam pling Technique (SM OTE) w ith a tw o feature

selection m ethod It w as shown that the proposed model

integrated w ith RF had significant im provem ent w ith the

highest f-m easure o f 99.5% [7],

It is w orth m entioning that m ost o f the aforem entioned

M L approaches w ere conducted in a determ inistic manner

Hence, there is a need to develop a probabilistic m odel that

is capable o f providing w ell-predicted results as well as

estim ating the confidence o f the results through associated

intervals Such a result is m ore relevant for experim ental data

on student exam scores rather than a single point estim ation

because, even w ith the same student, scattered results can be

obtained from different series o f experiments

In training programs at engineering universities,

specialized mini projects (SM Ps) play an im portant role as

they progressively provide knowledge as well as accumulate

conceiving, designing, implementing, and operating skills

necessary for their FYP, w hich is an integrated topic to solve

a practical problem o f the particular field o f engineer training

This article applies a m achine learning approach to predict FYP

and final CGPA results from those SMPs based on w hich the

influence o f SMPs on the FYP and CGPA can be evaluated in

a data-driven manner A case study is conducted by collecting

2890 datapoints in the form o f score results from eight SMPs,

one FYP, and the CGPA o f a group o f 289 engineering students

that graduated in 2022 in Hanoi Then, two deep learning

probabilistic models based on BNNs are established for FYP

and CGPA predictions It is shown from the obtained results

that the proposed approach is a practical tool providing quick

and reasonable analysis results such as the feature importance

score o f an individual SMP and the estim ated FYP and CGPA

results Furthermore, a relatively close estim ation can be

captured from the BNN model for CGPA, providing useful

information for academic management

Stochastic model using BNNs

As pointed out by various authors, a m ajor obstacle to the

data-driven m ethod is the scarcity o f relevant data, and this

problem becom es accentuated w hen studying the obtained

scores o f various individual students [8-12], Even w ith data

in hand, there exist unavoidable deviations betw een them [13-14], Thus, this study proposes to engineer a probabilistic

m achine learning m odel on the basis o f BNNs rather than determ inistic ones as done in the review ed works The advantage o f such a probabilistic m odel is that it is capable

o f predicting quantities o f interest such as the FYP score or CGPA, as well as estim ating the am ount o f uncertainty that

is associated w ith the prediction values It is evident that the

m ore data available, the m ore accurate the m odel, and vice versa In summary, the key contribution o f this article is to propose a probabilistic M L m odel to predict the FYP score and CGPA results from the given scores o f SMPs so that the effect o f an individual SMP as w ell as FYP on the final perform ance o f students can be evaluated

We begin by briefly review ing the A N N to set up

m athem atical sym bols and terminology G iven a dataset

w ith n being the num ber o f features Herein, each feature is

an input related to SM P and FYP results It is desirable to develop a non-linear m apping from X to Y, i.e., Y=fiX.) A standard architecture o f A N N consists o f an input layer, an output layer, and one or m any hidden layers w ith the total num ber o f layers o f the A N N being L. Each layer consists

o f various neurons that are fully connected w ith all neurons

in neighboring layers M athem atically, a neuron j at layer / could be described by a linear transform ation plus a non­ linear activation function, as follows:

where x} is output o f neuron j at layer /; is output o f neuron k at the previous layer /-1; w ! k is a w eight assigned

to the connection betw een the form er and the latter; Nj j is the total num ber o f neurons o f layer /- l; fa/-1 is a real value

to be determ ined, also know n as bias; and h is an activation function Here, the sigm oid activation function is used to squeeze values into the range (0,1) M oreover, the function is continuous and differentiable everywhere, thus rendering the training process o f the neural netw ork via a gradient descent- based algorithm faster and m ore effective

By setting W as the m atrix o f weights corresponding to layer /, w ith /= 1 , ,L , the A N N could be described by the following equation derived from the description o f neural netw orks in [15]:

w here ?t is a prediction o f F, and f t w ith l= \, ,L denotes transform ation operations at layer / in the ANN The netw ork will be iteratively trained to determ ine the optim al values o f

Wl that m inim ize the discrepancy betw een V) and Y

BNN is a probabilistic deep learning m odel that combines the high prediction perform ance o f A N N w ith the ability to

D E C E M B E R 2 0 2 2 • V O L U M E 6 4 N U M B E R 4 Vietnam Journal of Science.

Technology and Engineering 11

Trang 3

MATHEMATICS AND COMPUTER SCIENCE COMPUTATIONAL SCIENCE, PHYSICAL SCIENCES ENGINEERING

estimate uncertainty o f the Bayes theory [16] In the authors’

opinion, the model is especially suitable for working with

not-so-abundant collected data owing to two reasons: (i) in

practice, similar series o f experiments w ith identical input

parameters still provide different results due to unavoidable

uncertainty; and (ii) fitting an ANN with many parameters to a

limited database m ay cause the over-fitting problem, i.e., ANN

is likely to yield low-accuracy results on new data despite being

well trained In other words, it is necessary to not only perform

prediction o f FYP and CGPA results but also to estimate how

much confidence we have about the prediction results

For this purpose, rather than assigning the deterministic

values for weight W o f the neural network, BNN uses a

Gaussian probability distribution for W as below:

W = p + cr x e withe~N{0,l) ( 3 )

where // and a denote the m ean and standard deviation

m atrices o f W and e is the noise drawn from a zero-m ean

unity-variance norm al distribution Then, p, a are param eters

to determ ine through the learning process

N ote that the output o f BNN is a probability distribution,

thus a specialized loss function L is required to measure the

m odel’s perform ance The adopted m etric is the Kullback-

Leibler divergence (KL) w hose form ula is:

Next, the optimal values o f p" and a" are the solutions o f

the following m inim ization problem:

p ' o * = a rg m in K L ( q ( W \p , a ) \\p ( W \D ') ) ( 5 )

n,0

Via the B ayes’ rule, p(W\D) can be calculated as below:

p {W \D ) = p (D \W )p (W )

Substituting Eq (6) into Eq (4), the loss function L is

rewritten as follows:

L loSP(D) ~ 'ogPW - 'ogpfDllV) (7)

This loss function can be approxim ated from observed

discrete data as follows:

L = — ^ [ l o g q ( W ild er) - lo g p ( W 0 - logp(D |W D ] ( 8 )

s i= l

where A is the total num ber o f samples

N ext, the gradients o f the loss function w ith respect to p

and a are derived by:

d L (W \n ,a ) t d L (W \n ,a )

d L (W \p ,a ) dL (W \p , a )

A a = — d W x e + - 3 - do

Finally, p and a are updated using a small learning rate a

as follows:

a <- a — a x Act; p <- p — a x Ap. (10)

Case study - Database on score results of SMPs, FYP and CGPA

In the postgraduate training program o f civil engineers,

a student is required to pass all the subjects including eight specialized mini projects before being qualified to conduct his/her final year project The SM Ps consist of: Design

o f Architecture (DoA), D esign o f Foundation (DoF),

M echanism o f Reinforced Concrete Structures (M oRCS); Design o f Reinforced Concrete Buildings (DoRCB); Design o f Structural Steel Buildings (DoSSB); Construction Technology 1 (C T l); Construction Technology 2 (CT2); and Construction M anagem ent (CM) that will be num bered from

SM P.l to SMP.8, respectively Each individual SMP provides students w ith the corresponding know ledge and professional skill that w ill be integrated in his/her FYP to design and build

a civil/industrial building in real situation It is notew orthy that the SM Ps’ and F Y P ’s exams are all conducted in the form o f oral defence As a result, all the aforem entioned SMPs significantly influence the FYP, which together with all theoretical subjects and SM Ps contributes to the CGPA (Fig 1)

Fig 1 Projects in training program o f civil engineers in HUCE.

In this research, a dataset o f 2890 scores o f eight SMPs, one FYP, and the CGPA is collected from 289 civil engineers

w ho graduated from Hanoi U niversity o f Civil Engineering (HUCE) in 2022 Figs 2-5 display the histogram s o f all the score results o f their SMPs, FYP, and CGPA on the 4-point scale, showing clear visualization o f the range o f values as well as their distributions

140 120

100

'C 80

a a g* 60 ill

40

20 0 1.0 1.5 2.0 2.5 3.0 3.5 4.0

S c o re r e s u lts S c o re r e s u lts

Fig 2 Score histogram s of S M P.1-D oA and SWIP.2-DoF.

Vietnam Journal o f Science, DECEMBER 2022 VOLUME 64 NUMBER 4

Trang 4

MATHEMATICS AND COMPUTER SCIENCE COMPUTATIONAL SCIENCE, PHYSICAL SCIENCES ENGINEERING U

Fig 3 Score histogram s of SMP.3-MoRCS, SMP 4 -D 0 RCB and

SMP 5 -D 0 SSB.

140

120

100

S '80

V 60

to.

40

20

Fig 4 Score histogram s of SNIP.6-CT1, SMP.7-CT2 and SMP.8-CM.

Score results

Fig 5 Histogram s of FYP and CGPA results.

It can be seen in Fig 2 that the score results o f SMP.I-DoA

o f the investigated 289 students were quite high, among which

74 students obtained 4.0, whereas the biggest group was 3.0

with 111 students There were 67 students that earned a 3.5 and

the number o f students receiving scores o f 2.5, 2.0, and 1.5

were 11,21, and 5, respectively It is noted that although this is

the first SMP o f engineering students in the university program,

there is not much calculation in this task of architectural design

The remaining seven SMPs are all critical to students as they

are curtailed to train them on the design as well as construction

o f buildings It can be observed in the distributions o f these

SMP scores that the group o f 3.0 is almost dominant, except the

cases o f SMP.2-DoF and SMP.5-DoSSB, o f which the dominant

group was 2.0 (Figs 2 and 3) It is shown in Figs 3 and 4 that

the distribution o f the main SMPs in the program were similar

to each other Meanwhile, it can be observed in Fig 5 that the

distribution o f FYP results was quite standard, whereas there

is a regression trend o f the number o f students with the higher

scores in the CGPA results

Analysis results on FYP and CGPA using BNN model

In this study, the adopted architecture o f the BNN for FYP

scores is 8/16/16/1 The model consists o f an input layer with 8

neurons, 2 hidden layers with 16 neurons, and an output layer

with one neuron corresponding to the FYP result, as graphically illustrated in Fig 6A The neurons in the input layer correspond

to the score results o f SMP.l to SMP.8 Since the data size is moderate, it is reasonable to avoid using too deep architectures

o f many hidden layers as well as wide layers with a significant number o f neurons, which may lead to a pronounced increment

o f parameters to determine Besides, the number o f neurons for the hidden layer is set to 16 since it should be a power o f 2 to be convenient for the memory o f the computer For the proposed BNN model, each neuron has two parameters to be determined, characterizing the probability distribution o f its weight At the beginning o f learning, they are initialized as a normal distribution with zero mean and unity standard deviation For prediction o f CGPA, the corresponding BNN architecture is 9/16/16/1 in which the FYP score in the dataset is combined with the SMPs as the 9th element o f the input layer (Fig 6B)

(A) BNN model for FYP score (B) BNN model for CGPA (C) BNN protocol

Fig 6 Graphical representation of BNN whose weights are characterized by probability distributions.

U pdating the m o d el’s w eights as described in Eq.(10) is based on the A dam O ptim ization A lgorithm belonging to the first-order gradient descent optim ization family, w hich gradually adapts the m o d el’s w eights by a sm all am ount after each iteration to reduce the loss function The num ber

o f updates is controlled through a hyper-param eter o f 0.001 This value can also be referred to as the learning rate that

w as determ ined via a prelim inary test to ensure the learning process is convergent w ithin a reasonable learning time It is noted that a sm all learning rate will unnecessarily increase learning tim e, w hereas a large value could lead to prem ature results On the other hand, pre-processing standardization

is adopted to obviate the scale difference issue betw een different features w ith different physical m eanings B y utilizing the aid o f the deep learning library Pytorch for building the overall fram ew ork, the deep probabilistic library Pyro [17] for establishing the B N N -based data- driven m odel, Pandas for data m anagem ent, scikit-leam library [18] for data standardization, and M atplotlib for data visualization, the im plem entation o f the proposed data- driven fram ew ork can be realized in this study

In this study, the investigated database is split into three non-overlapping datasets, nam ely, the training, validation, and testing sub-sets w ith a ratio o f 60:20:20, corresponding

to data from 173, 58, and 58 students

DECEMBER 2022 VOLUME 64 NUMBER 4 Vietnam Journal o f Science,

Technology and Engineering 13

Trang 5

MATHEMATICS AND COMPUTER SCIENCE COMPUTATIONAL SCIENCE, PHYSICAL SCIENCES ENGINEERING

A fter being built, the probabilistic m odels w ere trained

w ith the training database m entioned in the previous section

Figs 7A and 7B depict the learning curves o f the B N N

m odels o f FY P and CGPA results, respectively, show ing

how loss functions evolve versus the num ber o f training

iterations (epochs) on both training and validation datasets

In the follow ing paragraphs, the corresponding results o f

the BNN m odel for CGPA will be given in parentheses

The KL loss function o f FYP (CGPA) quickly dropped for

epochs from 0 to 500, before gradually decreasing to values

around 0.035 and 0.02 (0.03 and 0.025) on the training and

validation datasets, respectively A fter that, a steady trend

is observed, i.e., no clear im provem ent is obtained, until

the num ber o f epochs reached 2000 For epochs 1300 to

1400, there is less fluctuation in loss function than the other

intervals H ence, an iteration value o f 1300 is selected as the

final configuration o f the B N N m odel

0.175 0.150 0.125

8 o.ioo

^ 0.075 0.050 0.025

N um ber of iterations Num ber of iterations

(A) BNN model for FYP (B) BNN model for CGPA

Fig 7 Evolution of the KL loss function against num ber of

iterations on training and validation datasets fo r (A) FYP and (B)

CGPA.

The effect o f each SM P on FY P and CGPA results can

be investigated via a feature im portance study Im portance

scores will be assigned for all SM Ps, a high score m eans that

the corresponding SM P has a significant im pact on the FY P/

CGPA and a low score m eans there is less impact The SM Ps

are ranked based on their im portance score Such results

provide understanding about the correlation betw een the

projects and helps students optim ize their learning strategy

to achieve desired final scores The perm utation feature

im portance m ethod is used along w ith the proposed data-

driven m odel H erein, a feature refers to an SM P score The

sim ple, yet effective, core idea o f this m ethod is to perm ute

the values o f features and evaluate the change in prediction

errors A perm uted feature m eans that original values o f this

feature are shuffled am ong data sam ples w hile other features

rem ain unchanged I f a perm uted feature incurs large errors,

this feature is im portant and contributes significantly to

the prediction results H ence, all features will be perm uted

one by one, and the respective error will be calculated

for each case N ext, these errors are sorted in descending

order, and im portance scores are derived Since the BNN

is a probabilistic m odel, the feature im portance results

obtained w ith BNN are random variables Thus, one repeats

the feature im portance calculations 100 tim es and then

derives their statistical characteristics such as m ean, min, and m ax values The influence o f SM Ps on a student’s final perform ance on their FY P and CGPA in term s o f im portance score are show n in Fig 8A and Fig 8B, respectively

Feature im portance on FYP Feature importance on CGPA

0.05 0.10 0.15 0.20 0.25 0.30

Im portance score (A) BNN model for FYP

0.2 0.3 0.4 0.5 0.6

Im portance score (B) BNN model for CGPA

Fig 8 Feature im portance graph of S M P s’ influence on (A) FYP and (B) CGPA.

It can be observed from Fig 8A that, com pared to the rem aining SM Ps, the m ini projects o f DoSSB and M oRCS have m ore pronounced influence on FYP results It is noted that in the training program the SM P and FY P have 1 and

10 credits, respectively (each credit equals to 15 hours o f lecturing) It can be seen in Fig 8B that the influence o f the FYP on CGPA, w hich is counted from a total o f 168 credits,

is m ost significant w hile M oR C S holds the second position The large gap betw een the feature im portance scores o f FYP and M oR C S is reasonable due to their credit relative ratio o f 10:1 Fig 8 also proves that reinforced concrete and steel structures are the leading applications w ith significant distributions currently in the field o f construction

N ext, the final perform ance o f the trained m odel is evaluated on the test dataset and the prediction results o f

FY P and CGPA are dem onstrated in Fig 9

(A ) BNN m odel fo r F Y P (B) BNN m odel f o r C G P A

Fig 9 Prediction results o f (A) FYP scores and (B) CGPA.

It is show n in Fig 9 th a t fo r each data p o in t (sh o w n

in the d o t sy m b o l), its X -c o o rd in a te d e n o tes tru e F Y P sco res fro m the d atab ase w h ile the Y -co o rd in ate is a

v a lu e p re d ic te d b y the m o d el Ideally, a p e rfe c t m odel

w ill p ro v id e the sam e resu lts as th o se fro m the d atab ase,

as h ig h lig h te d by the re d 4 5 -d eg ree line W h ile the B N N

m o d el fo r F Y P is n o t as clo se as e x p e c te d in F ig 9A , the

Vietnam Journal of Science,

Tethnologt and Engineering

DECEMBER 2022 VOLUME 64 NUMBER 4

Trang 6

MATHEMATICS AND COMPUTER SCIENCE I COMPUTATIONAL SCIENCE, PHYSICAL SCIENCES I ENGINEERING

p re d ic te d p o in ts o f C G P A lie re la tiv e ly clo se to the ideal

lin e in Fig 9B S pecifically, th e m e a n re la tiv e erro rs o f

p re d ic te d F Y P an d C G P A re su lts are 12.1% a n d 6.8% ,

resp ectiv ely T h ese re su lts q u a lita tiv e ly co n firm the

v ia b ility o f the p ro p o se d B N N m o d e l fo r C G PA

Conclusions

T h is stu d y a p p lie d a d a ta -d riv e n m e th o d fo r a ssessin g

th e F Y P sco re a n d C G P A b a se d o n th e re su lts o f eig h t

S M P s, w h ic h are c o n sid e re d as a c h a ra c teristic o f th e

tra in in g p ro g ra m fo r e n g in e e rin g stu d en ts P ro b a b ilistic

m a c h in e le a rn in g m o d e ls b a se d o n th e B N N w e re

in tro d u c e d an d th e th e o re tic a l fo u n d a tio n o f th e m o d e l

a n d k e y p a ra m e te rs o f th e p ro p o se d a p p ro a c h w ere

d escrib ed , fo llo w e d b y a case stu d y u sin g a d atab ase

o f 28 9 0 sco re re su lts o f S M P s, FY P, a n d C G P A fro m a

g ro u p o f civ il e n g in eers th a t g ra d u a te d in 20 2 2 in H an o i

O n e o f th e m a in re su lts o f th e p ro p o se d m o d e l is th a t it

ca n be u tiliz e d to ev a lu a te th e in flu en ce o f an in d iv id u al

S M P o n a s tu d e n t’s final p e rfo rm a n c e in te rm s o f F Y P

a n d C G PA It w a s sh o w n fro m th e re su lts o f th e case

stu d y th a t th e su b je c ts c o m m o n ly a p p lie d in p ra c tic e also

c o n trib u te a m o re sig n ifican t influence In ad d itio n , the

B N N m o d e l fo r C G PA is c a p a b le o f p ro v id in g re la tiv e ly

clo se p re d ic tio n s w ith a m e a n re la tiv e e rro r o f 6.8%

F u rth e rm o re , th e ap p lic a tio n o f th e d a ta -d riv e n m o d el

is stra ig h tfo rw a rd as it is b u ilt b a s e d o n o p e n so u rce

lib ra rie s a n d th e u se r-frie n d ly p ro g ra m m in g lan g u ag e,

P y th o n , w ith o u t re q u irin g a n y sp e c ia liz e d so ftw are

F o r th e n e x t ste p o f th e study, o th e r m o d e ls su ch as

stra ig h t artificial n e u ra l n e tw o rk a n d d ro p -o u t n e u ral

n e tw o rk c a n also b e in c o rp o ra te d fo r c o m p a riso n p u rp o ses

F u rth e rm o re , o ne can co m p le m e n t th e d ata b a se w ith the

sco re re su lts o f all stu d en ts th a t g ra d u a te d b efo re, d u rin g ,

o r a fter th e C O V ID -1 9 p a n d e m ic to g a in an o v erall

p ic tu re to p ro p o se ap p ro p ria te so lu tio n s to im p ro v e

aca d e m ic activ itie s o f e n g in e e rin g u n iv e rsitie s, w h e re

th e p ro je c ts p la y v e ry im p o rta n t ro le fo r u n d e rg ra d u a te

students

COMPETING INTERESTS

The authors declare that there is no conflict o f interest

regarding the publication o f this article

REFERENCES

[1] T Mitchell (1997), Machine Learning , McGraw-Hill

Education, 432pp

[2] J Hu, H Niu, J Carrasco, B Lennox, F Arvin (2020),

“Voronoi-based Multi-robot autonomous exploration in unknown

environments via deep reinforcement learning”, IEEE Transactions

on Vehicular Technology, 69(12), DOI: 10.1109/TVT.2020.3034800.

[3] Y Bengio, Y LeCun, G Hinton (2015), "Deep learning",

Nature, 521(7553), pp.436-444.

[4] S.S Abu-Naser, I.S Zaqout, M Abu-Ghosh, R.R Atallah,

E Alajrami (2015), “Predicting student performance using artificial neural network: In the Faculty of Engineering and Information

Technology”, International Journal o f Hybrid Information

Technology, 8(2), pp.221-228.

[5] E.Y Obsie, S.A Adem (2018), “Prediction of student academic performance using neural network, linear regression and

support vector regression: A case study”, International Journal o f

Computer Applications, 180(40), pp.39-47.

[6] Z Iqbal, J Qadir, A.N Mian, F Kamiran (2017), “Machine learning based student grade prediction: A case study”, https://arxiv org/abs/1708.08744

[7] S.D.ABujang, A Selamat, R Ibrahim, O Krejcar, E Herrera- Viedma, H Fujita, N.A.M Ghani (2021), “Multiclass prediction

model for student grade prediction using machine learning”, IEEE

Access, 9, pp.95608-95621.

[8] S.T Jishan, R.I Rashu, N Haque, R.M Rahman (2015),

“Improving accuracy of students’ final grade prediction model using optimal equal width binning and synthetic minority over-sampling

technique”, Decis Analyst, 2(1), pp.1-25.

[9] A Polyzou, G Karypis (2016), “Grade prediction with models

specific to students and courses”, Int J Data Sci Analyst, 2(3-4),

pp.159-171

[10] I Khan, A A1 Sadiri, A.R Ahmad, N Jabeur (2019),

“Tracking student performance in introductory programming by

means of machine learning”, Proc 4th MEC Int Conf Big Data

Smart City (ICBDSC), pp.1-6.

[11] M.A Al-Barrak, M Al-Razgan (2016), “Predicting students

final GPA using decision trees: A case study”, Int J Inf Educ

Technol., 6(7), pp.528-533.

[12] E.C Abana (2019), “A decision tree approach for predicting

student grades in research project using WEKA”, Int J Adv Comput

Sci Appl., 10(7), pp.285-289.

[13] F Ahmad, N.H Ismail, A.A Aziz (2015), “The prediction

of students’ academic performance using classification data mining techniques”, 4/>/>/ Math Sci., 9, pp.6415-6426.

[14] T Anderson, R Anderson (2017), “Applications of machine learning to student grade prediction in quantitative business courses”,

Glob J Bus Pedagog., 1(3), pp 13-22.

[15] Ng Andrew (2019), Stanford University CS229: Lecture

Notes, 216pp.

[16] C.M Bishop (1997), “Bayesian neural networks”, Journal o f

the Brazilian Computer Society, 4, pp.61-68.

[17] E Bingham, et al (2019), “Pyro: Deep universal probabilistic

programming”, The Journal o f Machine Learning Research, 20(1),

pp.973-978

[18] F Pedregosa, et al (2011), “Scikit-leam: Machine learning

in Python”, The Journal o f Machine Learning Research, 12, pp.2825-

2830

D E C E M B E R 2 0 2 2 V O L U M E 6 4 N U M B E R 4 \ itliiani Jimnial nl Science,

leclnmluiiv mul Fn” incei'ing 15

Ngày đăng: 27/02/2023, 23:57

TÀI LIỆU CÙNG NGƯỜI DÙNG

TÀI LIỆU LIÊN QUAN