line |
stmt |
bran |
cond |
sub |
pod |
time |
code |
1
|
|
|
|
|
|
|
#!/usr/bin/perl |
2
|
|
|
|
|
|
|
package AI::ANN::Evolver; |
3
|
|
|
|
|
|
|
BEGIN { |
4
|
2
|
|
|
2
|
|
10238
|
$AI::ANN::Evolver::VERSION = '0.008'; |
5
|
|
|
|
|
|
|
} |
6
|
|
|
|
|
|
|
# ABSTRACT: an evolver for an artificial neural network simulator |
7
|
|
|
|
|
|
|
|
8
|
2
|
|
|
2
|
|
18
|
use strict; |
|
2
|
|
|
|
|
3
|
|
|
2
|
|
|
|
|
66
|
|
9
|
2
|
|
|
2
|
|
135
|
use warnings; |
|
2
|
|
|
|
|
64
|
|
|
2
|
|
|
|
|
60
|
|
10
|
|
|
|
|
|
|
|
11
|
2
|
|
|
2
|
|
1210
|
use Moose; |
|
0
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
|
12
|
|
|
|
|
|
|
|
13
|
|
|
|
|
|
|
use AI::ANN; |
14
|
|
|
|
|
|
|
use Storable qw(dclone); |
15
|
|
|
|
|
|
|
use Math::Libm qw(tan); |
16
|
|
|
|
|
|
|
|
17
|
|
|
|
|
|
|
|
18
|
|
|
|
|
|
|
has 'max_value' => (is => 'rw', isa => 'Num', default => 1); |
19
|
|
|
|
|
|
|
has 'min_value' => (is => 'rw', isa => 'Num', default => 0); |
20
|
|
|
|
|
|
|
has 'mutation_chance' => (is => 'rw', isa => 'Num', default => 0); |
21
|
|
|
|
|
|
|
has 'mutation_amount' => (is => 'rw', isa => 'CodeRef', default => sub{sub{2 * rand() - 1}}); |
22
|
|
|
|
|
|
|
has 'add_link_chance' => (is => 'rw', isa => 'Num', default => 0); |
23
|
|
|
|
|
|
|
has 'kill_link_chance' => (is => 'rw', isa => 'Num', default => 0); |
24
|
|
|
|
|
|
|
has 'sub_crossover_chance' => (is => 'rw', isa => 'Num', default => 0); |
25
|
|
|
|
|
|
|
has 'gaussian_tau' => (is => 'rw', isa => 'CodeRef', default => sub{sub{1/sqrt(2*sqrt(shift))}}); |
26
|
|
|
|
|
|
|
has 'gaussian_tau_prime' => (is => 'rw', isa => 'CodeRef', default => sub{sub{1/sqrt(2*shift)}}); |
27
|
|
|
|
|
|
|
|
28
|
|
|
|
|
|
|
around BUILDARGS => sub { |
29
|
|
|
|
|
|
|
my $orig = shift; |
30
|
|
|
|
|
|
|
my $class = shift; |
31
|
|
|
|
|
|
|
my %data; |
32
|
|
|
|
|
|
|
if ( @_ == 1 && ref $_[0] eq 'HASH' ) { |
33
|
|
|
|
|
|
|
%data = %{$_[0]}; |
34
|
|
|
|
|
|
|
} else { |
35
|
|
|
|
|
|
|
%data = @_; |
36
|
|
|
|
|
|
|
} |
37
|
|
|
|
|
|
|
if ((not (ref $data{'mutation_amount'})) || ref $data{'mutation_amount'} ne 'CODE') { |
38
|
|
|
|
|
|
|
my $range = $data{'mutation_amount'}; |
39
|
|
|
|
|
|
|
$data{'mutation_amount'} = sub { $range * (rand() * 2 - 1) }; |
40
|
|
|
|
|
|
|
} |
41
|
|
|
|
|
|
|
return $class->$orig(%data); |
42
|
|
|
|
|
|
|
}; |
43
|
|
|
|
|
|
|
|
44
|
|
|
|
|
|
|
|
45
|
|
|
|
|
|
|
sub crossover { |
46
|
|
|
|
|
|
|
my $self = shift; |
47
|
|
|
|
|
|
|
my $network1 = shift; |
48
|
|
|
|
|
|
|
my $network2 = shift; |
49
|
|
|
|
|
|
|
my $class = ref($network1); |
50
|
|
|
|
|
|
|
my $inputcount = $network1->input_count(); |
51
|
|
|
|
|
|
|
my $minvalue = $network1->minvalue(); |
52
|
|
|
|
|
|
|
my $maxvalue = $network1->maxvalue(); |
53
|
|
|
|
|
|
|
my $afunc = $network1->afunc(); |
54
|
|
|
|
|
|
|
my $dafunc = $network1->dafunc(); |
55
|
|
|
|
|
|
|
# They better have the same number of inputs |
56
|
|
|
|
|
|
|
$inputcount == $network2->input_count() || return -1; |
57
|
|
|
|
|
|
|
my $networkdata1 = $network1->get_internals(); |
58
|
|
|
|
|
|
|
my $networkdata2 = $network2->get_internals(); |
59
|
|
|
|
|
|
|
my $neuroncount = $#{$networkdata1}; |
60
|
|
|
|
|
|
|
# They better also have the same number of neurons |
61
|
|
|
|
|
|
|
$neuroncount == $#{$networkdata2} || return -1; |
62
|
|
|
|
|
|
|
my $networkdata3 = []; |
63
|
|
|
|
|
|
|
|
64
|
|
|
|
|
|
|
for (my $i = 0; $i <= $neuroncount; $i++) { |
65
|
|
|
|
|
|
|
if (rand() < $self->{'sub_crossover_chance'}) { |
66
|
|
|
|
|
|
|
$networkdata3->[$i] = { 'inputs' => [], 'neurons' => [] }; |
67
|
|
|
|
|
|
|
$networkdata3->[$i]->{'iamanoutput'} = |
68
|
|
|
|
|
|
|
$networkdata1->[$i]->{'iamanoutput'}; |
69
|
|
|
|
|
|
|
for (my $j = 0; $j < $inputcount; $j++) { |
70
|
|
|
|
|
|
|
$networkdata3->[$i]->{'inputs'}->[$j] = |
71
|
|
|
|
|
|
|
(rand() > 0.5) ? |
72
|
|
|
|
|
|
|
$networkdata1->[$i]->{'inputs'}->[$j] : |
73
|
|
|
|
|
|
|
$networkdata2->[$i]->{'inputs'}->[$j]; |
74
|
|
|
|
|
|
|
# Note to self: Don't get any silly ideas about dclone()ing |
75
|
|
|
|
|
|
|
# these, that's a good way to waste half an hour debugging. |
76
|
|
|
|
|
|
|
} |
77
|
|
|
|
|
|
|
for (my $j = 0; $j <= $neuroncount; $j++) { |
78
|
|
|
|
|
|
|
$networkdata3->[$i]->{'neurons'}->[$j] = |
79
|
|
|
|
|
|
|
(rand() > 0.5) ? |
80
|
|
|
|
|
|
|
$networkdata1->[$i]->{'neurons'}->[$j] : |
81
|
|
|
|
|
|
|
$networkdata2->[$i]->{'neurons'}->[$j]; |
82
|
|
|
|
|
|
|
} |
83
|
|
|
|
|
|
|
} else { |
84
|
|
|
|
|
|
|
$networkdata3->[$i] = dclone( |
85
|
|
|
|
|
|
|
(rand() > 0.5) ? |
86
|
|
|
|
|
|
|
$networkdata1->[$i] : |
87
|
|
|
|
|
|
|
$networkdata2->[$i] ); |
88
|
|
|
|
|
|
|
} |
89
|
|
|
|
|
|
|
} |
90
|
|
|
|
|
|
|
my $network3 = $class->new ( 'inputs' => $inputcount, |
91
|
|
|
|
|
|
|
'data' => $networkdata3, |
92
|
|
|
|
|
|
|
'minvalue' => $minvalue, |
93
|
|
|
|
|
|
|
'maxvalue' => $maxvalue, |
94
|
|
|
|
|
|
|
'afunc' => $afunc, |
95
|
|
|
|
|
|
|
'dafunc' => $dafunc); |
96
|
|
|
|
|
|
|
return $network3; |
97
|
|
|
|
|
|
|
} |
98
|
|
|
|
|
|
|
|
99
|
|
|
|
|
|
|
|
100
|
|
|
|
|
|
|
sub mutate { |
101
|
|
|
|
|
|
|
my $self = shift; |
102
|
|
|
|
|
|
|
my $network = shift; |
103
|
|
|
|
|
|
|
my $class = ref($network); |
104
|
|
|
|
|
|
|
my $networkdata = $network->get_internals(); |
105
|
|
|
|
|
|
|
my $inputcount = $network->input_count(); |
106
|
|
|
|
|
|
|
my $minvalue = $network->minvalue(); |
107
|
|
|
|
|
|
|
my $maxvalue = $network->maxvalue(); |
108
|
|
|
|
|
|
|
my $afunc = $network->afunc(); |
109
|
|
|
|
|
|
|
my $dafunc = $network->dafunc(); |
110
|
|
|
|
|
|
|
my $neuroncount = $#{$networkdata}; # BTW did you notice that this |
111
|
|
|
|
|
|
|
# isn't what it says it is? |
112
|
|
|
|
|
|
|
$networkdata = dclone($networkdata); # For safety. |
113
|
|
|
|
|
|
|
for (my $i = 0; $i <= $neuroncount; $i++) { |
114
|
|
|
|
|
|
|
# First each input/neuron pair |
115
|
|
|
|
|
|
|
for (my $j = 0; $j < $inputcount; $j++) { |
116
|
|
|
|
|
|
|
my $weight = $networkdata->[$i]->{'inputs'}->[$j]; |
117
|
|
|
|
|
|
|
if (defined $weight && $weight != 0) { |
118
|
|
|
|
|
|
|
if (rand() < $self->{'mutation_chance'}) { |
119
|
|
|
|
|
|
|
$weight += (rand() * 2 - 1) * $self->{'mutation_amount'}; |
120
|
|
|
|
|
|
|
if ($weight > $self->{'max_value'}) { |
121
|
|
|
|
|
|
|
$weight = $self->{'max_value'}; |
122
|
|
|
|
|
|
|
} |
123
|
|
|
|
|
|
|
if ($weight < $self->{'min_value'}) { |
124
|
|
|
|
|
|
|
$weight = $self->{'min_value'} + 0.000001; |
125
|
|
|
|
|
|
|
} |
126
|
|
|
|
|
|
|
} |
127
|
|
|
|
|
|
|
if (abs($weight) < $self->{'mutation_amount'}) { |
128
|
|
|
|
|
|
|
if (rand() < $self->{'kill_link_chance'}) { |
129
|
|
|
|
|
|
|
$weight = undef; |
130
|
|
|
|
|
|
|
} |
131
|
|
|
|
|
|
|
} |
132
|
|
|
|
|
|
|
} else { |
133
|
|
|
|
|
|
|
if (rand() < $self->{'add_link_chance'}) { |
134
|
|
|
|
|
|
|
$weight = rand() * $self->{'mutation_amount'}; |
135
|
|
|
|
|
|
|
# We want to Do The Right Thing. Here, that means to |
136
|
|
|
|
|
|
|
# detect whether the user is using weights in (0, x), and |
137
|
|
|
|
|
|
|
# if so make sure we don't accidentally give them a |
138
|
|
|
|
|
|
|
# negative weight, because that will become 0.000001. |
139
|
|
|
|
|
|
|
# Instead, we'll generate a positive only value at first |
140
|
|
|
|
|
|
|
# (it's easier) and then, if the user will accept negative |
141
|
|
|
|
|
|
|
# weights, we'll let that happen. |
142
|
|
|
|
|
|
|
if ($self->{'min_value'} < 0) { |
143
|
|
|
|
|
|
|
($weight *= 2) -= $self->{'mutation_amount'}; |
144
|
|
|
|
|
|
|
} |
145
|
|
|
|
|
|
|
# Of course, we have to check to be sure... |
146
|
|
|
|
|
|
|
if ($weight > $self->{'max_value'}) { |
147
|
|
|
|
|
|
|
$weight = $self->{'max_value'}; |
148
|
|
|
|
|
|
|
} |
149
|
|
|
|
|
|
|
if ($weight < $self->{'min_value'}) { |
150
|
|
|
|
|
|
|
$weight = $self->{'min_value'} + 0.000001; |
151
|
|
|
|
|
|
|
} |
152
|
|
|
|
|
|
|
# But we /don't/ need to to a kill_link_chance just yet. |
153
|
|
|
|
|
|
|
} |
154
|
|
|
|
|
|
|
} |
155
|
|
|
|
|
|
|
# This would be a bloody nightmare if we hadn't done that dclone |
156
|
|
|
|
|
|
|
# magic before. But look how easy it is! |
157
|
|
|
|
|
|
|
$networkdata->[$i]->{'inputs'}->[$j] = $weight; |
158
|
|
|
|
|
|
|
} |
159
|
|
|
|
|
|
|
# Now each neuron/neuron pair |
160
|
|
|
|
|
|
|
for (my $j = 0; $j <= $neuroncount; $j++) { |
161
|
|
|
|
|
|
|
# As a reminder to those cursed with the duty of maintaining this code: |
162
|
|
|
|
|
|
|
# This should be an exact copy of the code above, except that 'inputs' |
163
|
|
|
|
|
|
|
# would be replaced with 'neurons'. |
164
|
|
|
|
|
|
|
my $weight = $networkdata->[$i]->{'neurons'}->[$j]; |
165
|
|
|
|
|
|
|
if (defined $weight && $weight != 0) { |
166
|
|
|
|
|
|
|
if (rand() < $self->{'mutation_chance'}) { |
167
|
|
|
|
|
|
|
$weight += (rand() * 2 - 1) * $self->{'mutation_amount'}; |
168
|
|
|
|
|
|
|
if ($weight > $self->{'max_value'}) { |
169
|
|
|
|
|
|
|
$weight = $self->{'max_value'}; |
170
|
|
|
|
|
|
|
} |
171
|
|
|
|
|
|
|
if ($weight < $self->{'min_value'}) { |
172
|
|
|
|
|
|
|
$weight = $self->{'min_value'} + 0.000001; |
173
|
|
|
|
|
|
|
} |
174
|
|
|
|
|
|
|
} |
175
|
|
|
|
|
|
|
if (abs($weight) < $self->{'mutation_amount'}) { |
176
|
|
|
|
|
|
|
if (rand() < $self->{'kill_link_chance'}) { |
177
|
|
|
|
|
|
|
$weight = undef; |
178
|
|
|
|
|
|
|
} |
179
|
|
|
|
|
|
|
} |
180
|
|
|
|
|
|
|
|
181
|
|
|
|
|
|
|
} else { |
182
|
|
|
|
|
|
|
if (rand() < $self->{'add_link_chance'}) { |
183
|
|
|
|
|
|
|
$weight = rand() * $self->{'mutation_amount'}; |
184
|
|
|
|
|
|
|
# We want to Do The Right Thing. Here, that means to |
185
|
|
|
|
|
|
|
# detect whether the user is using weights in (0, x), and |
186
|
|
|
|
|
|
|
# if so make sure we don't accidentally give them a |
187
|
|
|
|
|
|
|
# negative weight, because that will become 0.000001. |
188
|
|
|
|
|
|
|
# Instead, we'll generate a positive only value at first |
189
|
|
|
|
|
|
|
# (it's easier) and then, if the user will accept negative |
190
|
|
|
|
|
|
|
# weights, we'll let that happen. |
191
|
|
|
|
|
|
|
if ($self->{'min_value'} < 0) { |
192
|
|
|
|
|
|
|
($weight *= 2) -= $self->{'mutation_amount'}; |
193
|
|
|
|
|
|
|
} |
194
|
|
|
|
|
|
|
# Of course, we have to check to be sure... |
195
|
|
|
|
|
|
|
if ($weight > $self->{'max_value'}) { |
196
|
|
|
|
|
|
|
$weight = $self->{'max_value'}; |
197
|
|
|
|
|
|
|
} |
198
|
|
|
|
|
|
|
if ($weight < $self->{'min_value'}) { |
199
|
|
|
|
|
|
|
$weight = $self->{'min_value'} + 0.000001; |
200
|
|
|
|
|
|
|
} |
201
|
|
|
|
|
|
|
# But we /don't/ need to to a kill_link_chance just yet. |
202
|
|
|
|
|
|
|
} |
203
|
|
|
|
|
|
|
} |
204
|
|
|
|
|
|
|
# This would be a bloody nightmare if we hadn't done that dclone |
205
|
|
|
|
|
|
|
# magic before. But look how easy it is! |
206
|
|
|
|
|
|
|
$networkdata->[$i]->{'neurons'}->[$j] = $weight; |
207
|
|
|
|
|
|
|
} |
208
|
|
|
|
|
|
|
# That was rather tiring, and that's only for the first neuron!! |
209
|
|
|
|
|
|
|
} |
210
|
|
|
|
|
|
|
# All done. Let's pack it back into an object and let someone else deal |
211
|
|
|
|
|
|
|
# with it. |
212
|
|
|
|
|
|
|
$network = $class->new ( 'inputs' => $inputcount, |
213
|
|
|
|
|
|
|
'data' => $networkdata, |
214
|
|
|
|
|
|
|
'minvalue' => $minvalue, |
215
|
|
|
|
|
|
|
'maxvalue' => $maxvalue, |
216
|
|
|
|
|
|
|
'afunc' => $afunc, |
217
|
|
|
|
|
|
|
'dafunc' => $dafunc); |
218
|
|
|
|
|
|
|
return $network; |
219
|
|
|
|
|
|
|
} |
220
|
|
|
|
|
|
|
|
221
|
|
|
|
|
|
|
|
222
|
|
|
|
|
|
|
sub mutate_gaussian { |
223
|
|
|
|
|
|
|
my $self = shift; |
224
|
|
|
|
|
|
|
my $network = shift; |
225
|
|
|
|
|
|
|
my $class = ref($network); |
226
|
|
|
|
|
|
|
my $networkdata = $network->get_internals(); |
227
|
|
|
|
|
|
|
my $inputcount = $network->input_count(); |
228
|
|
|
|
|
|
|
my $minvalue = $network->minvalue(); |
229
|
|
|
|
|
|
|
my $maxvalue = $network->maxvalue(); |
230
|
|
|
|
|
|
|
my $afunc = $network->afunc(); |
231
|
|
|
|
|
|
|
my $dafunc = $network->dafunc(); |
232
|
|
|
|
|
|
|
my $neuroncount = $#{$networkdata}; # BTW did you notice that this |
233
|
|
|
|
|
|
|
# isn't what it says it is? |
234
|
|
|
|
|
|
|
$networkdata = dclone($networkdata); # For safety. |
235
|
|
|
|
|
|
|
for (my $i = 0; $i <= $neuroncount; $i++) { |
236
|
|
|
|
|
|
|
my $n = 0; |
237
|
|
|
|
|
|
|
for (my $j = 0; $j < $inputcount; $j++) { |
238
|
|
|
|
|
|
|
my $weight = $networkdata->[$i]->{'inputs'}->[$j]; |
239
|
|
|
|
|
|
|
$n++ if $weight; |
240
|
|
|
|
|
|
|
} |
241
|
|
|
|
|
|
|
for (my $j = 0; $j <= $neuroncount; $j++) { |
242
|
|
|
|
|
|
|
my $weight = $networkdata->[$i]->{'neurons'}->[$j]; |
243
|
|
|
|
|
|
|
$n++ if $weight; |
244
|
|
|
|
|
|
|
} |
245
|
|
|
|
|
|
|
next if $n == 0; |
246
|
|
|
|
|
|
|
my $tau = &{$self->{'gaussian_tau'}}($n); |
247
|
|
|
|
|
|
|
my $tau_prime = &{$self->{'gaussian_tau_prime'}}($n); |
248
|
|
|
|
|
|
|
my $random1 = 2 * rand() - 1; |
249
|
|
|
|
|
|
|
for (my $j = 0; $j < $inputcount; $j++) { |
250
|
|
|
|
|
|
|
my $weight = $networkdata->[$i]->{'inputs'}->[$j]; |
251
|
|
|
|
|
|
|
next unless $weight; |
252
|
|
|
|
|
|
|
my $random2 = 2 * rand() - 1; |
253
|
|
|
|
|
|
|
$networkdata->[$i]->{'eta_inputs'}->[$j] *= exp($tau_prime*$random1+$tau*$random2); |
254
|
|
|
|
|
|
|
$networkdata->[$i]->{'inputs'}->[$j] += $networkdata->[$i]->{'eta_inputs'}->[$j]*$random2; |
255
|
|
|
|
|
|
|
} |
256
|
|
|
|
|
|
|
for (my $j = 0; $j <= $neuroncount; $j++) { |
257
|
|
|
|
|
|
|
my $weight = $networkdata->[$i]->{'neurons'}->[$j]; |
258
|
|
|
|
|
|
|
next unless $weight; |
259
|
|
|
|
|
|
|
my $random2 = 2 * rand() - 1; |
260
|
|
|
|
|
|
|
$networkdata->[$i]->{'eta_neurons'}->[$j] *= exp($tau_prime*$random1+$tau*$random2); |
261
|
|
|
|
|
|
|
$networkdata->[$i]->{'neurons'}->[$j] += $networkdata->[$i]->{'eta_neurons'}->[$j]*$random2; |
262
|
|
|
|
|
|
|
} |
263
|
|
|
|
|
|
|
} |
264
|
|
|
|
|
|
|
# All done. Let's pack it back into an object and let someone else deal |
265
|
|
|
|
|
|
|
# with it. |
266
|
|
|
|
|
|
|
$network = $class->new ( 'inputs' => $inputcount, |
267
|
|
|
|
|
|
|
'data' => $networkdata, |
268
|
|
|
|
|
|
|
'minvalue' => $minvalue, |
269
|
|
|
|
|
|
|
'maxvalue' => $maxvalue, |
270
|
|
|
|
|
|
|
'afunc' => $afunc, |
271
|
|
|
|
|
|
|
'dafunc' => $dafunc); |
272
|
|
|
|
|
|
|
return $network; |
273
|
|
|
|
|
|
|
} |
274
|
|
|
|
|
|
|
|
275
|
|
|
|
|
|
|
__PACKAGE__->meta->make_immutable; |
276
|
|
|
|
|
|
|
|
277
|
|
|
|
|
|
|
1; |
278
|
|
|
|
|
|
|
|
279
|
|
|
|
|
|
|
__END__ |
280
|
|
|
|
|
|
|
=pod |
281
|
|
|
|
|
|
|
|
282
|
|
|
|
|
|
|
=head1 NAME |
283
|
|
|
|
|
|
|
|
284
|
|
|
|
|
|
|
AI::ANN::Evolver - an evolver for an artificial neural network simulator |
285
|
|
|
|
|
|
|
|
286
|
|
|
|
|
|
|
=head1 VERSION |
287
|
|
|
|
|
|
|
|
288
|
|
|
|
|
|
|
version 0.008 |
289
|
|
|
|
|
|
|
|
290
|
|
|
|
|
|
|
=head1 METHODS |
291
|
|
|
|
|
|
|
|
292
|
|
|
|
|
|
|
=head2 new |
293
|
|
|
|
|
|
|
|
294
|
|
|
|
|
|
|
AI::ANN::Evolver->new( { mutation_chance => $mutationchance, |
295
|
|
|
|
|
|
|
mutation_amount => $mutationamount, add_link_chance => $addlinkchance, |
296
|
|
|
|
|
|
|
kill_link_chance => $killlinkchance, sub_crossover_chance => |
297
|
|
|
|
|
|
|
$subcrossoverchance, min_value => $minvalue, max_value => $maxvalue } ) |
298
|
|
|
|
|
|
|
|
299
|
|
|
|
|
|
|
All values have a sane default. |
300
|
|
|
|
|
|
|
|
301
|
|
|
|
|
|
|
mutation_chance is the chance that calling mutate() will add a random value |
302
|
|
|
|
|
|
|
on a per-link basis. It only affects existing (nonzero) links. |
303
|
|
|
|
|
|
|
mutation_amount is the maximum change that any single mutation can introduce. |
304
|
|
|
|
|
|
|
It affects the result of successful mutation_chance rolls, the maximum |
305
|
|
|
|
|
|
|
value after an add_link_chance roll, and the maximum strength of a link |
306
|
|
|
|
|
|
|
that can be deleted by kill_link_chance rolls. It can either add or |
307
|
|
|
|
|
|
|
subtract. |
308
|
|
|
|
|
|
|
add_link_chance is the chance that, during a mutate() call, each pair of |
309
|
|
|
|
|
|
|
unconnected neurons or each unconnected neuron => input pair will |
310
|
|
|
|
|
|
|
spontaneously develop a connection. This should be extremely small, as |
311
|
|
|
|
|
|
|
it is not an overall chance, put a chance for each connection that does |
312
|
|
|
|
|
|
|
not yet exist. If you wish to ensure that your neural net does not become |
313
|
|
|
|
|
|
|
recursive, this must be zero. |
314
|
|
|
|
|
|
|
kill_link_chance is the chance that, during a mutate() call, each pair of |
315
|
|
|
|
|
|
|
connected neurons with a weight less than mutation_amount or each |
316
|
|
|
|
|
|
|
neuron => input pair with a weight less than mutation_amount will be |
317
|
|
|
|
|
|
|
disconnected. If add_link_chance is zero, this should also be zero, or |
318
|
|
|
|
|
|
|
your network will just fizzle out. |
319
|
|
|
|
|
|
|
sub_crossover_chance is the chance that, during a crossover() call, each |
320
|
|
|
|
|
|
|
neuron will, rather than being inherited fully from each parent, have |
321
|
|
|
|
|
|
|
each element within it be inherited individually. |
322
|
|
|
|
|
|
|
min_value is the smallest acceptable weight. It must be less than or equal to |
323
|
|
|
|
|
|
|
zero. If a value would be decremented below min_value, it will instead |
324
|
|
|
|
|
|
|
become an epsilon above min_value. This is so that we don't accidentally |
325
|
|
|
|
|
|
|
set a weight to zero, thereby killing the link. |
326
|
|
|
|
|
|
|
max_value is the largest acceptable weight. It must be greater than zero. |
327
|
|
|
|
|
|
|
gaussian_tau and gaussian_tau_prime are the terms to the gaussian mutation |
328
|
|
|
|
|
|
|
method. They are coderefs which accept one parameter, n, the number of |
329
|
|
|
|
|
|
|
non-zero-weight inputs to the given neuron. |
330
|
|
|
|
|
|
|
|
331
|
|
|
|
|
|
|
=head2 crossover |
332
|
|
|
|
|
|
|
|
333
|
|
|
|
|
|
|
$evolver->crossover( $network1, $network2 ) |
334
|
|
|
|
|
|
|
|
335
|
|
|
|
|
|
|
Returns a $network3 consisting of the shuffling of $network1 and $network2 |
336
|
|
|
|
|
|
|
As long as the same neurons in network1 and network2 are outputs, network3 |
337
|
|
|
|
|
|
|
will always have those same outputs. |
338
|
|
|
|
|
|
|
This method, at least if the sub_crossover_chance is nonzero, expects neurons |
339
|
|
|
|
|
|
|
to be labeled from zero to n. |
340
|
|
|
|
|
|
|
You probably don't want to do this. This is the least effective way to evolve |
341
|
|
|
|
|
|
|
neural networks. This is because, due to the hidden intermediate steps, it |
342
|
|
|
|
|
|
|
is possible for two networks which output exactly the same with completely |
343
|
|
|
|
|
|
|
different internal representations. |
344
|
|
|
|
|
|
|
|
345
|
|
|
|
|
|
|
=head2 mutate |
346
|
|
|
|
|
|
|
|
347
|
|
|
|
|
|
|
$evolver->mutate($network) |
348
|
|
|
|
|
|
|
|
349
|
|
|
|
|
|
|
Returns a version of $network mutated according to the parameters set for |
350
|
|
|
|
|
|
|
$evolver, followed by a series of counters. The original is not modified. |
351
|
|
|
|
|
|
|
The counters are, in order, the number of times we compared against the |
352
|
|
|
|
|
|
|
following thresholds: mutation_chance, kill_link_chance, add_link_chance. |
353
|
|
|
|
|
|
|
This is useful if you want to try to normalize your probabilities. For |
354
|
|
|
|
|
|
|
example, if you want to make links be killed about as often as they are |
355
|
|
|
|
|
|
|
added, keep a running total of the counters, and let: |
356
|
|
|
|
|
|
|
$kill_link_chance = $add_link_chance * $add_link_counter / $kill_link_counter |
357
|
|
|
|
|
|
|
This will probably make kill_link_chance much larger than add_link_chance, |
358
|
|
|
|
|
|
|
but in doing so will make links be added at overall the same rate as they |
359
|
|
|
|
|
|
|
are killed. Since new links tend to be killed particularly quickly, it may |
360
|
|
|
|
|
|
|
be wise to add an additional optional multiplier to mutation_amount just |
361
|
|
|
|
|
|
|
for new links. |
362
|
|
|
|
|
|
|
|
363
|
|
|
|
|
|
|
=head2 mutate_gaussian |
364
|
|
|
|
|
|
|
|
365
|
|
|
|
|
|
|
$evolver->mutate_gaussian($network) |
366
|
|
|
|
|
|
|
|
367
|
|
|
|
|
|
|
Returns a version of $network modified according to the Gaussian mutation |
368
|
|
|
|
|
|
|
rules discussed in X. Yao, Evolving Artifical Neural Networks, and X. Yao |
369
|
|
|
|
|
|
|
and Y. Liu, Fast Evolution Strategies. Uses the gaussian_tau and |
370
|
|
|
|
|
|
|
gaussian_tau_prime values from the initializer if they are present, or |
371
|
|
|
|
|
|
|
sane defaults proposed by the above. These are both functions of 'n', the |
372
|
|
|
|
|
|
|
number of inputs to each neuron with nonzero weight. |
373
|
|
|
|
|
|
|
|
374
|
|
|
|
|
|
|
=head1 AUTHOR |
375
|
|
|
|
|
|
|
|
376
|
|
|
|
|
|
|
Dan Collins <DCOLLINS@cpan.org> |
377
|
|
|
|
|
|
|
|
378
|
|
|
|
|
|
|
=head1 COPYRIGHT AND LICENSE |
379
|
|
|
|
|
|
|
|
380
|
|
|
|
|
|
|
This software is Copyright (c) 2011 by Dan Collins. |
381
|
|
|
|
|
|
|
|
382
|
|
|
|
|
|
|
This is free software, licensed under: |
383
|
|
|
|
|
|
|
|
384
|
|
|
|
|
|
|
The GNU General Public License, Version 3, June 2007 |
385
|
|
|
|
|
|
|
|
386
|
|
|
|
|
|
|
=cut |
387
|
|
|
|
|
|
|
|