C++ Neural Networks and Fuzzy Logic: Preface



Download 1,14 Mb.
Pdf ko'rish
bet114/443
Sana29.12.2021
Hajmi1,14 Mb.
#77367
1   ...   110   111   112   113   114   115   116   117   ...   443
Bog'liq
C neural networks and fuzzy logic

Weights on Connections

Weight assignments on connections between neurons not only indicate the strength of the signal that is being

fed for aggregation but also the type of interaction between the two neurons. The type of interaction is one of

cooperation or of competition. The cooperative type is suggested by a positive weight, and the competition by

a negative weight, on the connection. The positive weight connection is meant for what is called excitation,

while the negative weight connection is termed an inhibition.



Initialization of Weights

Initializing the network weight structure is part of what is called the encoding phase of a network operation.

The encoding algorithms are several, differing by model and by application. You may have gotten the

impression that the weight matrices used in the examples discussed in detail thus far have been arbitrarily

determined; or if there is a method of setting them up, you are not told what it is.

It is possible to start with randomly chosen values for the weights and to let the weights be adjusted

appropriately as the network is run through successive iterations. This would make it easier also. For example,

under supervised training, if the error between the desired and computed output is used as a criterion in

adjusting weights, then one may as well set the initial weights to zero and let the training process take care of

the rest. The small example that follows illustrates this point.



A Small Example

Suppose you have a network with two input neurons and one output neuron, with forward connections

between the input neurons and the output neuron, as shown in Figure 5.2. The network is required to output a

C++ Neural Networks and Fuzzy Logic:Preface

Instar and Outstar

90



1 for the input patterns (1, 0) and (1, 1), and the value 0 for (0, 1) and (0, 0). There are only two connection

weights w



1

 and w



2

.

Figure 5.2

  Neural network with forward connections.

Let us set initially both weights to 0, but you need a threshold function also. Let us use the following

threshold function, which is slightly different from the one used in a previous example:

1 if x > 0

f(x)

=

{



0 if x d 0

The reason for modifying this function is that if f(x) has value 1 when x = 0, then no matter what the weights

are, the output will work out to 1 with input (0, 0). This makes it impossible to get a correct computation of

any function that takes the value 0 for the arguments (0, 0).

Now we need to know by what procedure we adjust the weights. The procedure we would apply for this

example is as follows.



  If the output with input pattern (a, b) is as desired, then do not adjust the weights.

  If the output with input pattern (a, b) is smaller than what it should be, then increment each of w

1

and w



2

 by 1.


  If the output with input pattern (a, b) is greater than what it should be, then subtract 1 from w

1

 if


the product aw

1

 is smaller than 1, and adjust w



2

 similarly.

Table 5.9 shows what takes place when we follow these procedures, and at what values the weights settle.


Download 1,14 Mb.

Do'stlaringiz bilan baham:
1   ...   110   111   112   113   114   115   116   117   ...   443




Ma'lumotlar bazasi mualliflik huquqi bilan himoyalangan ©hozir.org 2024
ma'muriyatiga murojaat qiling

kiriting | ro'yxatdan o'tish
    Bosh sahifa
юртда тантана
Боғда битган
Бугун юртда
Эшитганлар жилманглар
Эшитмадим деманглар
битган бодомлар
Yangiariq tumani
qitish marakazi
Raqamli texnologiyalar
ilishida muhokamadan
tasdiqqa tavsiya
tavsiya etilgan
iqtisodiyot kafedrasi
steiermarkischen landesregierung
asarlaringizni yuboring
o'zingizning asarlaringizni
Iltimos faqat
faqat o'zingizning
steierm rkischen
landesregierung fachabteilung
rkischen landesregierung
hamshira loyihasi
loyihasi mavsum
faolyatining oqibatlari
asosiy adabiyotlar
fakulteti ahborot
ahborot havfsizligi
havfsizligi kafedrasi
fanidan bo’yicha
fakulteti iqtisodiyot
boshqaruv fakulteti
chiqarishda boshqaruv
ishlab chiqarishda
iqtisodiyot fakultet
multiservis tarmoqlari
fanidan asosiy
Uzbek fanidan
mavzulari potok
asosidagi multiservis
'aliyyil a'ziym
billahil 'aliyyil
illaa billahil
quvvata illaa
falah' deganida
Kompyuter savodxonligi
bo’yicha mustaqil
'alal falah'
Hayya 'alal
'alas soloh
Hayya 'alas
mavsum boyicha


yuklab olish