This paper presents the method of feature
subset selection in dynamic stability assessment
power system using artificial neural networks.
This paper proposed applying four feature subset
selection algorithms that are FR, SFS, SBS, and
SFFS. The effectiveness of the algorithms was
tested on the GSO-37bus power system. With
the same number of feature, the calculation
results show that SFS algorithm yielded higher
classification rate than FR, SBS algorithm. SFS
algorithm yielded the same classification rate as
SFFS algorithm.
10 trang |
Chia sẻ: linhmy2pp | Ngày: 19/03/2022 | Lượt xem: 202 | Lượt tải: 0
Bạn đang xem nội dung tài liệu Feature subset selection in dynamic stability assessment power system using artificial neural networks, để tải tài liệu về máy bạn click vào nút DOWNLOAD ở trên
TAÏP CHÍ PHAÙT TRIEÅN KH&CN, TAÄP 18, SOÁ K3- 2015
Feature subset selection in dynamic
stability assessment power system
using artificial neural networks
Nguyen Ngoc Au 1
Quyen Huy Anh 1
Phan Thi Thanh Binh 2
1Ho Chi Minh city University of Technical and Education
2Ho Chi Minh city University of Technology, VNU-HCM
(Manuscript Received on October 30nd, 2014, Manuscript Revised July 08nd, 2015)
ABSTRACT
This paper presents method of feature Sequential Backward Selection (SBS),
subset selection in dynamic stability Sequential Forward Floating Selection
assessment (DSA) power system using (SFFS) and Feature Ranking (FR) algorithm
artificial neural networks (ANN). In the to feature subset selection. The
application of ANN on DSA power system, effectiveness of the algorithms was tested
feature subset selection aims to reduce the on the GSO-37bus power system. With the
number of training features, cost and same number of features, the calculation
memory computer. However, the major results show that SFS algorithm yielded
challenge is to reduce the number of higher classification rate than FR, SBS
features but classification rate gets a high algorithm. SFS algorithm yielded the same
accuracy. This paper proposes applying classification rate as SFFS algorithm.
Sequential Forward Selection (SFS),
Key words: feature subset selection, dynamic stability assessment, artificial neural
networks, and power system.
1. INTRODUCTION sources and transmission systems are not
developed to meet the load demand. While
Modern power systems are forced to
operating the power system is always faced with
operate under highly stressed operating
unusual circumstances such as a generator
conditions closer to their stability limits. The
outage, loss of a line, sudden dropping of a large
operation of power systems is challenged
load, switching of station or substation, and
increasingly significant because investment
Trang 15
SCIENCE & TECHNOLOGY DEVELOPMENT, Vol.18, No.K3 - 2015
three-phase sudden short circuit, ... Power The intelligent systems for DSA consist of
system stability is the ability to regain an four basic steps: database generation, feature
equilibrium state after being subjected to a selection, knowledge extraction and model
physical disturbance and maintain the validation. In particular, a very important stage
continuous supply of electricity to customers. is feature selection because it greatly affects
Power system stability is classified [1]: rotor cost, computational time and recognition
angle stability, frequency stability and voltage accuracy of DSA system. Feature selection
stability. Rotor angle stability is divided into two actually reduces features or variables, just select
categories including short-term and long-term. the minimum number of variables but ensure
Short-term stability angle is considered transient recognition accuracy. This paper proposed
dynamic stability and important contribution in applying FR (Feature Ranking), SFFS
power system stability. Long-term stability (Sequential Forward Floating Selection), SFS
angle includes small signal stability and (Sequential Forward Selection), SBS
frequency stability. (Sequential Backward Selection) algorithm for
feature subset selection. The case study was
Due to the complexity of the power system,
done on GSO-37bus power system diagram with
traditional methods to power system analysis
the support of simulation software PowerWorld
take so much time and cause delays in decision
17. The algorithms of feature subset selection
making. However, the relationship between pre-
were programmed on Matlab software.
fault parameters of the power system state and
Multilayer Feed forward Neural Networks
post-fault modes of power system stability has
(MLFN) is supported by Matlab software.
highly nonlinear, extremely difficult to describe
this mathematical relationship. In order to 2. METHOD
overcome such difficulties, intelligent system,
2.1. Mathematical Model of Multimachine
that is ANN, has been proposed for DSA thanks
Power System
to special abilities in pattern classification
[2],[6],[7]. Operating conditions of power The dynamic behavior of a generator power
systems have wide range so that it is difficult system can be described by the following
perform online calculations. ANN is in need of differential equations [1]:
initial off-line data for training. Extensive off-
d 2
i (1)
line simulation is performed so as to acquire a M i 2 Pmi Pei
dt
large enough set of training data to represent the
different operating conditions of typical power d i
It is known that: i (2)
systems. As a pattern classifier, once trained, dt
neural networks not only have extremely fast By substituting (2) in (1), therefore (1) becomes:
solutions but also get the ability to update new
patterns or new operating conditions by d i
M i Pmi Pei (3)
generalizing the training data, improving dt
recognition accuracy [7].
Where: i: rotor angle of machine i; i: rotor
speed of machine i; Pmi: mechanical power of
Trang 16
TAÏP CHÍ PHAÙT TRIEÅN KH&CN, TAÄP 18, SOÁ K3- 2015
machine i; Pei: electrical power of machine i; Step4. Subset feature evaluation.
M : moment of inertia of machine i.
i Step 5. Subset feature selection.
The state of the power system is stable
2.2.2. Data generation, initial feature set
when the rotor angle deviation of any two
selection.
generators not exceeding 1800, and is unstable
when the rotor angle deviation of any two A large number of samples are generated
generators exceed 1800. Status of power system through off-line simulation and the stable status
was performed according to the proposed rules is evaluated for each fault under study. Data for
in [1],]4],[5], as follow: each bus or line fault occurring in the test
systems are recorded in which samples of data
0
If ij < 180 then Stable are kept in a database. The input is the vector of
(4) system state parameters that characterize the
0
If ij 180 then Unstable
current system state, usually called feature, they
can be classified into pre-fault, fault-on and post-
2.2. Feature subset selection
fault features.
2.2.1. General Description
Pre-fault features [2]: steady-state
The MLNF-based DSA power system can operating parameters such as voltage magnitude
be formulated as a mapping yi = f (xi) after and angle of buses, P, Q load, generation and line
learning from a stability database
flow qualities Pflow, Qflow, Pload, Qload, Vbus, and
n before disturbance occurs (P , Q , ,).
D{xi , yi}i1. Where xi is feature; It is n- gen gen bus
dimensional input vector that characterizes the Fault-on features [6]: variables that
system operating state; and yi is output vector. characterize at fault-on state of power system
The feature subset selection consists of selecting occur such as changes in nodal powers, in power
a d dimensional feature vector z. Where d < flows in transmission line, voltage drops in the
n; The d selected features represent the original nodes at instance of fault (Pflow, Qflow, Pload,
data in a new knowledge base Qload, Vbus,).
d
Dnew {zi , yi}i1, and the new mapping Post-fault features [4]: variables that
ynewi=fnew(zi). Thus, feature selection is actually describe system dynamic behavior after
taking away unnecessary features and selecting disturbance occurs such as relative rotor angle,
a candidate subset of features that get rich rotor angular velocity, rotor acceleration, rotor
information with highly accurate identification kinetic energy, and the dynamic voltage
of model. This process includes the following trajectory,
steps: The problem of transient stability is usually
Step 1. Data generation, initial feature set divided into two main categories: assessment
selection. and prediction. Transient stability assessment
usually focuses on the critical clearing time
Step 2. Candidate feature subset selection.
(CCT). In transient stability prediction, the CCT
Step 3. Training and testing classification rate. is not of interest [11]. In this aspect, the progress
Trang 17
SCIENCE & TECHNOLOGY DEVELOPMENT, Vol.18, No.K3 - 2015
of power system transient due to the occurrence
x i m i (5)
z i
of disturbance is monitored. The key question in i
transient stability prediction is: the transient
Where: m is mean value of data. is standard
swings are finally ‘Stable’ or ‘Unstable’ [3], i i
deviation of data.
[10]-[12]. Vector output variables represent the
stable conditions of the power system. Need of 2.2.3. Candidate feature subset selection.
fast DSA power system after the fault is stable or
This step is the process of searching for
unstable, so the output variables are assigned to
potential subset features. The search strategy is
label binary variable y [10, 01]. Class 1 [10] is
divided into a global search and local search.
stable class and class 2 [01] is unstable class.
Global search strategy has the great advantage
The use the post-fault variables can be too that for optimal result, but expensive
long for operators to take timely remedial computation time. Therefore, the optimal search
actions to stop the extremely fast transient strategy is not appropriate when a large number
instability development process. of input variables. In the case of large input
feature, local optimization search strategy will
Found that, pre-fault input features are
spend less time searching because the search
variables that are too difficult to find a clear
process is not through the entire search space.
signal for sampled dataset learning. Post-fault
input features will prolong a warning of 2.2.3.1. Local optimization search strategies
instability power system. Fault-on input features
- Sequential Forward Selection – SFS [8]:
are proposed in [6] to overcome the drawbacks
The SFS method begins with an empty set (k=0),
such as analysis since the changes in the value of
adds one feature at a time to selected subset with
the parameters of input variables are a clear
(k+1) features so that the new subset maximizes
signal for dataset learning. So, this paper did
the cost function J(k+1). It stops when the selected
mining of fault-on input features (V , P ,
bus load subset has the d desired number of features, k<d.
Qload, Pflow, Qflow) as a database for training
neural networks. -Sequential Backward Selection-SBS [8]:
The SBS method begins with all input features
The output variables represent the dynamic D (k=D), removes one feature at a time to
behavior of power system at fault-on. By selected subset with (k-1) features so that the
observation from off-line simulation, these
resultant subset maximizes the cost function J(k-
binary output variables indicate the status of the
1). The algorithm stops when the resultant feature
power system to comply with the law (4). set has the d desired number of features, k<d.
The quantitative variables have different - Sequential Forward Floating Selection-
units of measurement; the value of the variables SFFS [8]: The SFFS is one of two algorithms of
in the different ranges will affect the calculation Floating Search Algorithm (FSA) that are SFFS
results in recognition. Data normalization and SBFS (Sequential Backward Floating
methods commonly applied in accordance with Selection). The SFFS algorithm the search starts
the following formula: with an empty feature set and uses the SFS
algorithm to add one feature at a time to the
Trang 18
TAÏP CHÍ PHAÙT TRIEÅN KH&CN, TAÄP 18, SOÁ K3- 2015
selected feature subset. Every time a new feature Between-class scatter matrix that describes the
is added to the current feature set, the algorithm scatter of the class means about the total mean
tries to backtrack by using the SBS algorithm to is:
remove one feature at a time to find a better c
Ni T
S (m m)(m m) (11)
subset. The algorithm terminates when the size b N i i
of the current feature set is larger than the d i1
S is the covariance matrix of the feature vector
desired number of features. m
with respect to the global mean. Its trace is the
-Feature Ranking-FR [2],[4]: This is a sum of variances of the features around their
simple method which uses less computing time.
respective global mean. Sm is:
By evaluating cost function of a single feature,
S = S +S (12)
then it is ranked by ordering the best of them and m w b
select for a good feature. Goal is to find a feature subset for which the
within-class spread is small and the between-
2.2.3.2. Cost function [8, 9]
class spread is large. The cost function is:
Let the n data samples be x1 , . . . , xn. The
1
sample covariance matrix, Sm, is given by (6): J trace{Sw Sm} (13)
N th
1 T Formula (14), that was written for the k single
S (x m)(x m) (6)
m n n feature, is Fisher distance function:
N n1
The sample mean of all data: S (k )
J (k ) b (14)
S (k )
1 N w
m xn (7) The value of J is bigger means that the feature is
N n1
more important.
The sample mean of class c :
i 2.2.4. Training and testing classification rate
1
(8) To test the studied methods without loss of
mi xn
Ni xnci generality, the database is randomly partition
into k subsets that are D1, D2, , Di,, Dk, each
Where: c is the number of class; N is the number
i equal size. The model is trained on all the
of sample mean of class c ; N is the number of
i subsets except for one that is tested to measuring
all samples.
of validation accuracy. Training and testing are
SW, within-class scatter Matrix, is: performed k times. The validation accuracy or
classification rate is computed for each of the k
c
1 Ni validation sets and averaged to get a final cross-
Sw Si (9)
N i1 N validation accuracy. Classification rate of
1 training or testing is determined by the formula
T (10)
Si (xn mi )(xn mi ) (15):
Ni x c
n n
r(%) r .100 (15)
Si: is the covariance matrix for class i. N
Trang 19
SCIENCE & TECHNOLOGY DEVELOPMENT, Vol.18, No.K3 - 2015
Where: nr is the number of sample for training or (37+25+25+56+56). The number of output
testing with right result; N is the number of variables is 2 (class 1 [10]: stable class, class 2
sample for training or testing. [01]: unstable class). From simulated results and
based on the law (4), there were 240 samples
The expected value (EV) of classification rate of
with 120 stable samples and 120 unstable
the model was proposed in [6] by the formula
samples. Sample set was normalized by formula
(16):
(5). Full feature set was randomly divided into 6
EV 0.9 (16) feature subsets. Each feature subset had 40
2.2.5. Training and testing classification rate samples (20 stable samples and 20 unstable
and subset feature evaluation samples). So, each training subset had 200
samples (100 stable samples, 100 unstable
Applying feature subset selection
samples) and testing subset had 40 samples (20
algorithms were described as above to selecting
stable samples, 20 unstable samples).
feature subsets. Each feature subset was trained
and tested, the classification rates are calculated 3.2 Results of feature subset selection
by the formula (15). In this paper, four search algorithms that are
Feature subset is selected with conditions SFS, SBS, SFFS and FR, were proposed
that have smaller a number of features, agree to applying to feature subset selection. In which,
the formula (16) and get higher classification the SFS, SBS, SFFS algorithms had been applied
rate. in [2]. The objective function (13) was applied
for these three algorithms in this study. FR
3. RESULTS - DISCUSSION algorithm had been applied in [2],[4] with Fisher
3.1 Feature set, samples for training distance function (15). Figure 1 shows the results
of distance measuring value by SFFS, SFS and
The off-line simulation was implemented to
SBS algorithm. Figure 2 shows the results
collection data for training. In this study, the
ranked from large to small according to Fisher's
GSO-37bus system, that is the standard system
distance measuring the value of each single
in the simulation program of PowerWorld 17
feature.
software, [5], was used as case study. It consists
of 37 buses, 9 generators; three different voltage 24
levels are 345kV, 138kV and 69kV, 25 loads, 14 22
transformers, 42 transmission lines. Load level 20
is one hundred percent rated load. Fault types are 18
balanced three-phase, single line to ground, line 16
J value J
to line, double line to ground at buses and along 14 SFFS
SFS
transmission lines. Setting fault clearing time is 12 SBS
25ms [5] with all faults. 10
8
8 10 12 14 16 18 20
Input and output variables are x[Vbus,
Feature (d)
Pload, Qload, Pflow, Qflow] and y[10,01].
Figure 1 distance calculated value of SFFS, SFS and
Totalof input variables is 199
SBS algorithm.
Trang 20
TAÏP CHÍ PHAÙT TRIEÅN KH&CN, TAÄP 18, SOÁ K3- 2015
Table 1. The measured distance (J value) of
SFFS and SFS algorithm of feature subsets with
98
d=13 and d=20.
96
Feature J value J value (SFS) 94
(d) (SFFS)
92
13 15.42795 15.42502 90
Classification 88rate (%) SFFS
20 23.54927 23.48542 SFS
86 SBS
Fisher
1.4 84
8 10 12 14 16 18 20
Feature (d)
1.2
1 Figure 3 clasification rate of testing feature subsets
by MLNF.
0.8
J value 0.6 updating algorithm was selected. These
functions are supported in neural networks tool
0.4
of R2011b Matlab software. Programs were
0.2
performed by laptop with CPU Inter CoreTM i3-
0
0 50 100 150 200 380M, 2GB DDR3 Memory, 500GB HDD.
Feature (d) Figure 3 shows classification rate of testing
Figure 2 Fisher's distance measuring value. feature subsets with algorithms by MLNF.
Table 3 Training time and testing classification
Table 2. Calculating time of SFS, SFFS, SBS rate of algorithms with d=12 and d=199.
algorithm with d=20 and FR with d=199. feature Training r(%)
SFFS SBS FR (d) time (s)
SFS SFS 12 2.03 95.0
Time (s) 1.15 2.58 117.5 0.14 SBS 12 2.18 92.1
Fisher 12 2.16 87.9
Total 199 7.61 95.8
3.3 Results of training
From Table 3, we can observe that SFS
MLNF had three layers: one input layer, algorithm got higher classification rate than
one hidden layer and one output layer. Hidden others. So, Suggested method-based SFS
layer has 10 neurals with activate function algorithm applied to select 12 top of features
tansig. Activate function purelin was used for (VbusWeber69, VbusBLT138,
output layer. Levenberg-Marquardt optimization PloadSHIMKO69, PloadHALE69,
based for weight and bias
QloadPATTEN69, QloadAMANDA69,
QloadBLT138, PloadBLT69,
Trang 21
SCIENCE & TECHNOLOGY DEVELOPMENT, Vol.18, No.K3 - 2015
QloadLYNN138, PflowRAY138-BOB138, According to Table 2, with 20 features, it
PflowBOB138-BLT138, PflowBLT69- can see that calculating time of FR algorithm is
BLT138) in order to reduce the number of inputs the shortest time with 0.14s. Calculating time of
to MLFN. However, we expected to check SBS algorithm is the longest time with 117.5s.
whether SFS is biased to any different Calculating time of SFFS algorithm is 2.58s and
classifiers. Linear Discrimination Analysis was longer 2,2 times than calculating time of SFS
used as the classification algorithm to our testing algorithm. Calculating time of SFFS algorithm is
it. The linear classifier (LC) is one of the 1.15s. Calculating time of SBS algorithm is
simplest discrimination analysis types. This much longer calculating time of SFS, SFFS and
classify function is also supported by Matlab FR algorithm. This can explain that the SBS
software. Figure 4 shows classification rate of algorithm has to through the space search with
testing feature subsets with algorithms by LC. the entire feature set. SFFS algorithm has longer
calculating time than SFS’s time calculating
94 because beside of forward search, SFFS
algorithm has to backward search. The shortest
92
calculating time of FR algorithm has a reason
SFFS
90
SFS that FR algorithm calculated measuring distance
SBF
88 Fisher values only one time respectively for each
feature.
86
Figure 3, classification rates of SFS and
Classification rate (%) rate Classification 84
SFFS algorithm are the same. SFFS and SFS
82 algorithm give better results than SBS and FR
80 algorithm. Classification rates of SFS and SFFS
8 10 12 14 16 18 20
Feature (d) algorithm are more 1,3% to 2,9% than SBS
algorithm and more 4,6% to 8,3% than FR
Figure 4 Classification rate of testing feature subsets
algorithm.
by LC
According to Table 3, SFS algorithm,
3.4 Discussion
subset has 12 features that its classification rate
Figure 1 shows the results of distance got 95% by MLFN. Comparing with feature set
measuring value by SFFS, SFS and SBS has 199 features, SFS algorithm’s feature
algorithm. Figure 2 shows the results ranked number was reduced 16,5 times, its training time
from large to small according to Fisher's distance was reduced 3,74 times. Classification rate of
measuring value of each single feature by FR that feature set has 199 features is 95,8%. By
algorithm. In which, the same distance comparing the calculated results found that SFS
measuring values were caculated by SFS and algorithm has the same results as SFFS
SFFS, but that have very small value difference algorithm. These results can be explained that in
at subsets with 13 features and 20 features as step backward search SFFS algorithm only
Table 1. removes one feature for each execution
algorithm could not search deep enough to find
Trang 22
TAÏP CHÍ PHAÙT TRIEÅN KH&CN, TAÄP 18, SOÁ K3- 2015
better features. SFS algorithm is simpler than 4. CONCLUSION
SFFS algorithm.
This paper presents the method of feature
Classification rates of SFS and SFFS subset selection in dynamic stability assessment
algorithm are also the same and got better results power system using artificial neural networks.
than SBS and FR algorithm by LC. MLFN got This paper proposed applying four feature subset
higher classification rate than LC for the same selection algorithms that are FR, SFS, SBS, and
feature subset selection algorithm. The SFS has SFFS. The effectiveness of the algorithms was
the 12 selected features that its classification rate tested on the GSO-37bus power system. With
got 95% by MLFN. This result was also the same number of feature, the calculation
considered acceptable for some previous studies results show that SFS algorithm yielded higher
applying pattern recognition to power system classification rate than FR, SBS algorithm. SFS
stability. For instance, classification rate got algorithm yielded the same classification rate as
95% [11], 93,6% [12]. SFFS algorithm.
Lựa chọn tập biến trong đánh giá ổn
định động hệ thống điện sử dụng mạng
thần kinh nhân tạo
Nguyễn Ngọc Âu 1
Quyền Huy Ánh 1
Phan Thị Thanh Bình 2
1Trường Đại học Sư Phạm Kỹ Thuật Thành Phố Hồ Chí Minh
2Trường Đại học Bách Khoa, ĐHQG-HCM
TÓM TẮT
Bài báo trình bày phương pháp lựa chọn kiếm tiến (SFS), tìm kiếm lùi (SBS), tìm kiếm
tập biến trong đánh giá ổn định động (DSA) kết hợp tiến lùi (SFFS), xếp hạng (FR) để lựa
hệ thống điện sử dụng mạng thần kinh nhân chọn tập biến. Hiệu quả của các giải thuật đã
tạo (ANN). Trong ứng dụng ANN đánh giá ổn được kiểm tra với sơ đồ hệ thống điện GSO-
định động hệ thống điện, lựa chọn tập biến 37bus. Kết quả tính toán cho thấy với cùng
nhằm mục đích giảm số biến đầu vào, giảm biến đặc trưng (Feature), giải thuật SFS có
chi phí và bộ nhớ máy tính . Tuy nhiên, thách độ chính xác nhận dạng cao hơn giải thuật
thức lớn là cùng với việc giảm số lượng biến FR và SBS, giải thuật SFS và SFFS có kết
nhưng độ chính xác nhận dạng phải cao. Bài quả tính toán như nhau.
báo này đề nghị áp dụng các giải thuật tìm
Trang 23
SCIENCE & TECHNOLOGY DEVELOPMENT, Vol.18, No.K3 - 2015
Từ khóa: lựa chọn tập biến, đánh giá ổn định động, mạng thần kinh nhân tạo, hệ thống
điện.
REFERENCES
[1]. Prabha Kundur, ‘Power System Stability Techniques, Theory and Applications To
and Control’, McGraw-Hill Inc, 1994. Power Systems’, John Wiley & Sons, Inc
[2]. Yan Xu, Zhao Yang Dong, JunHua Zhao, Publication, 2008.
Pei Zhang, Kit Po Wong,’A Reliable [8]. Andrew R.Webb, Keith D.Copsey,
Intelligent System for Real-Time Dynamic ’Statistical Pattern Recognition’, Third
Security Assessment of Power Systems’, Edition, A John Wiley & Sons, Ltd.,
IEEE Transactions On Power Systems, Publication, 2011.
Vol. 27, No. 3, p.1253-1263, August 2012. [9]. Mohamed Cheriet, Nawwaf Kharma,
[3]. Nima Amjady and Seyed Farough Majedi, Cheng-Lin Liu, Ching Y. Suen,’
‘Transient Stability Prediction by a Hybrid ‘Character Recognition systems: A Guide
Intelligent System’, IEEE Transactions On for Students and Practioners’’ A John
Power Systems, Vol. 22, No. 3, p.1275- Wiley & Sons, Ltd., Publication, 2007.
1283, August 2007. [10]. R. Zhang, S. Member, Y. Xu, and Z. Y.
[4]. K. Shanti Swarup, ‘Artificial neural Dong, “Feature Selection For Intelligent
network using pattern recognition for Stability Assessment of Power Systems,”
security assessment and analysis’, 2012 IEEE Power Energy Soc. Gen. Meet.,
Neurocomputing 71, 983–998, Elsevier, pp. 1–7, 2012.
2008. [11]. A. M. a. Haidar, M. W. Mustafa, F. a. F.
[5]. J. Duncan Glover, Mulukutla S.Sarma, Ibrahim, and I. a. Ahmed, “Transient
Thomas J.Overbye,’ Power System stability evaluation of electrical power
Analysis and Design’, Fifth Edition, system using generalized regression neural
Publisher Global Engineering: Christopher networks,” Appl. Soft Comput., vol. 11, no.
M. Shortt, 2012. 4, pp. 3558–3570, 2011.
[6]. Quyen Huy Anh, ‘The applycation of [12]. A. M. El-Arabaty, H. a. Talaat, M. M.
pattern recognition for fast analysis of the Mansour, and a. Y. Abd-Elaziz, “Out-of-
dynamic stability of electrical power step detection based on pattern
system’, Electrical technology, No.2 pp 1- recognition,” Int. J. Electr. Power Energy
13, Perganon, 1994. Syst., vol. 16, no. 4, pp. 269–275, 1994.
[7]. Kwang Y. Lee and Mohamed A. El-
Sharkawi, ‘Modern Heuristic Optimization
Trang 24
Các file đính kèm theo tài liệu này:
- feature_subset_selection_in_dynamic_stability_assessment_pow.pdf