| line |
stmt |
bran |
cond |
sub |
pod |
time |
code |
|
1
|
|
|
|
|
|
|
#!/usr/bin/perl |
|
2
|
|
|
|
|
|
|
# NanoB2B-NER::NER::Arffman |
|
3
|
|
|
|
|
|
|
# |
|
4
|
|
|
|
|
|
|
# Creates ARFF files from annotated files |
|
5
|
|
|
|
|
|
|
# Version 1.9 |
|
6
|
|
|
|
|
|
|
# |
|
7
|
|
|
|
|
|
|
# Program by Milk |
|
8
|
|
|
|
|
|
|
|
|
9
|
|
|
|
|
|
|
package NanoB2B::NER::Arffman; |
|
10
|
|
|
|
|
|
|
|
|
11
|
1
|
|
|
1
|
|
7
|
use NanoB2B::UniversalRoutines; |
|
|
1
|
|
|
|
|
2
|
|
|
|
1
|
|
|
|
|
27
|
|
|
12
|
1
|
|
|
1
|
|
5
|
use MetaMap::DataStructures; |
|
|
1
|
|
|
|
|
2
|
|
|
|
1
|
|
|
|
|
23
|
|
|
13
|
1
|
|
|
1
|
|
8
|
use File::Path qw(make_path); #makes sub directories |
|
|
1
|
|
|
|
|
3
|
|
|
|
1
|
|
|
|
|
65
|
|
|
14
|
1
|
|
|
1
|
|
5
|
use List::MoreUtils qw(uniq); |
|
|
1
|
|
|
|
|
1
|
|
|
|
1
|
|
|
|
|
11
|
|
|
15
|
|
|
|
|
|
|
|
|
16
|
1
|
|
|
1
|
|
648
|
use strict; |
|
|
1
|
|
|
|
|
2
|
|
|
|
1
|
|
|
|
|
42
|
|
|
17
|
1
|
|
|
1
|
|
9
|
use warnings; |
|
|
1
|
|
|
|
|
3
|
|
|
|
1
|
|
|
|
|
7429
|
|
|
18
|
|
|
|
|
|
|
|
|
19
|
|
|
|
|
|
|
#option variables |
|
20
|
|
|
|
|
|
|
my $debug = 1; |
|
21
|
|
|
|
|
|
|
my $program_dir = ""; |
|
22
|
|
|
|
|
|
|
my $fileIndex = 0; |
|
23
|
|
|
|
|
|
|
my $stopwords_file; |
|
24
|
|
|
|
|
|
|
my $prefix = 3; |
|
25
|
|
|
|
|
|
|
my $suffix = 3; |
|
26
|
|
|
|
|
|
|
my $bucketsNum = 10; |
|
27
|
|
|
|
|
|
|
my $is_cui = 0; |
|
28
|
|
|
|
|
|
|
my $sparse_matrix = 0; |
|
29
|
|
|
|
|
|
|
my $wcs = ""; |
|
30
|
|
|
|
|
|
|
|
|
31
|
|
|
|
|
|
|
#datastructure object |
|
32
|
|
|
|
|
|
|
my %params = (); |
|
33
|
|
|
|
|
|
|
my $dataStructures = MetaMap::DataStructures->new(\%params); |
|
34
|
|
|
|
|
|
|
|
|
35
|
|
|
|
|
|
|
#universal subroutines object |
|
36
|
|
|
|
|
|
|
my %uniParams = (); |
|
37
|
|
|
|
|
|
|
my $uniSub; |
|
38
|
|
|
|
|
|
|
|
|
39
|
|
|
|
|
|
|
#other general global variables |
|
40
|
|
|
|
|
|
|
my @allBuckets; |
|
41
|
|
|
|
|
|
|
my %fileHash; |
|
42
|
|
|
|
|
|
|
my %metamapHash; |
|
43
|
|
|
|
|
|
|
my %featHash; |
|
44
|
|
|
|
|
|
|
my %orthoHash; |
|
45
|
|
|
|
|
|
|
my @features; |
|
46
|
|
|
|
|
|
|
my $selfId = "_self"; |
|
47
|
|
|
|
|
|
|
my $entId = "_e"; |
|
48
|
|
|
|
|
|
|
my $morphID = "_m"; |
|
49
|
|
|
|
|
|
|
|
|
50
|
|
|
|
|
|
|
my $stopRegex; |
|
51
|
|
|
|
|
|
|
|
|
52
|
|
|
|
|
|
|
#### A HERO IS BORN #### |
|
53
|
|
|
|
|
|
|
|
|
54
|
|
|
|
|
|
|
# construction method to create a new Arffman object |
|
55
|
|
|
|
|
|
|
# input : $directory <-- the name of the directory for the files |
|
56
|
|
|
|
|
|
|
# $name <-- name of the file to examine |
|
57
|
|
|
|
|
|
|
# $features <-- the list of features to use [e.g. "ortho morph text pos cui sem"] |
|
58
|
|
|
|
|
|
|
# $bucketsNum <-- the number of buckets to use for k-fold cross validation |
|
59
|
|
|
|
|
|
|
# \$debug <-- run the program with debug print statements |
|
60
|
|
|
|
|
|
|
# \$prefix <-- the number of letters to look at the beginning of each word |
|
61
|
|
|
|
|
|
|
# \$suffix <-- the number of letters to look at the end of each word |
|
62
|
|
|
|
|
|
|
# \$index <-- the index to start metamapping from in the set of files |
|
63
|
|
|
|
|
|
|
# \$no_stopwords <-- exclude examining stop words [imported from the stop word list] |
|
64
|
|
|
|
|
|
|
# output : $self <-- an instance of the Arffman object |
|
65
|
|
|
|
|
|
|
sub new { |
|
66
|
|
|
|
|
|
|
#grab class and parameters |
|
67
|
0
|
|
|
0
|
0
|
|
my $self = {}; |
|
68
|
0
|
|
|
|
|
|
my $class = shift; |
|
69
|
0
|
0
|
|
|
|
|
return undef if(ref $class); |
|
70
|
0
|
|
|
|
|
|
my $params = shift; |
|
71
|
|
|
|
|
|
|
|
|
72
|
|
|
|
|
|
|
#reset all arrays and hashes |
|
73
|
0
|
|
|
|
|
|
@allBuckets = (); |
|
74
|
0
|
|
|
|
|
|
%fileHash = (); |
|
75
|
0
|
|
|
|
|
|
%metamapHash = (); |
|
76
|
0
|
|
|
|
|
|
%featHash = (); |
|
77
|
0
|
|
|
|
|
|
%orthoHash = (); |
|
78
|
0
|
|
|
|
|
|
@features = (); |
|
79
|
|
|
|
|
|
|
|
|
80
|
|
|
|
|
|
|
#bless this object - hallelujah |
|
81
|
0
|
|
|
|
|
|
bless $self, $class; |
|
82
|
0
|
|
|
|
|
|
$self->_init($params); |
|
83
|
0
|
|
|
|
|
|
@allBuckets = (1..$bucketsNum); |
|
84
|
|
|
|
|
|
|
|
|
85
|
|
|
|
|
|
|
#retrieve parameters for universal-routines |
|
86
|
0
|
|
|
|
|
|
$uniParams{'debug'} = $debug; |
|
87
|
0
|
|
|
|
|
|
$uniSub = NanoB2B::UniversalRoutines->new(\%uniParams); |
|
88
|
|
|
|
|
|
|
|
|
89
|
|
|
|
|
|
|
#return the object |
|
90
|
0
|
|
|
|
|
|
return $self; |
|
91
|
|
|
|
|
|
|
} |
|
92
|
|
|
|
|
|
|
|
|
93
|
|
|
|
|
|
|
# method to initialize the NanoB2B::NER::Arffman object. |
|
94
|
|
|
|
|
|
|
# input : $parameters <- reference to a hash |
|
95
|
|
|
|
|
|
|
# output: |
|
96
|
|
|
|
|
|
|
sub _init { |
|
97
|
0
|
|
|
0
|
|
|
my $self = shift; |
|
98
|
0
|
|
|
|
|
|
my $params = shift; |
|
99
|
|
|
|
|
|
|
|
|
100
|
0
|
0
|
|
|
|
|
$params = {} if(!defined $params); |
|
101
|
|
|
|
|
|
|
|
|
102
|
|
|
|
|
|
|
# get some of the parameters |
|
103
|
0
|
|
|
|
|
|
my $diroption = $params->{'directory'}; |
|
104
|
0
|
|
|
|
|
|
my $ftsoption = $params->{'features'}; |
|
105
|
0
|
|
|
|
|
|
my $bucketsNumoption = $params->{'bucketsNum'}; |
|
106
|
0
|
|
|
|
|
|
my $debugoption = $params->{'debug'}; |
|
107
|
0
|
|
|
|
|
|
my $prefixoption = $params->{'prefix'}; |
|
108
|
0
|
|
|
|
|
|
my $suffixoption = $params->{'suffix'}; |
|
109
|
0
|
|
|
|
|
|
my $indexoption = $params->{'index'}; |
|
110
|
0
|
|
|
|
|
|
my $stopwordoption = $params->{'stopwords'}; |
|
111
|
0
|
|
|
|
|
|
my $iscuioption = $params->{'is_cui'}; |
|
112
|
0
|
|
|
|
|
|
my $sparsematrixoption = $params->{'sparse_matrix'}; |
|
113
|
0
|
|
|
|
|
|
my $wcsoption = $params->{'wcs'}; |
|
114
|
|
|
|
|
|
|
|
|
115
|
|
|
|
|
|
|
#set the global variables |
|
116
|
0
|
0
|
|
|
|
|
if(defined $debugoption){$debug = $debugoption;} |
|
|
0
|
|
|
|
|
|
|
|
117
|
0
|
0
|
|
|
|
|
if(defined $diroption){$program_dir = $diroption;} |
|
|
0
|
|
|
|
|
|
|
|
118
|
0
|
0
|
|
|
|
|
if(defined $indexoption){$fileIndex = $indexoption;} |
|
|
0
|
|
|
|
|
|
|
|
119
|
0
|
0
|
|
|
|
|
if(defined $stopwordoption){$stopwords_file = $stopwordoption;} |
|
|
0
|
|
|
|
|
|
|
|
120
|
0
|
0
|
|
|
|
|
if(defined $iscuioption){$is_cui = $iscuioption;} |
|
|
0
|
|
|
|
|
|
|
|
121
|
0
|
0
|
|
|
|
|
if(defined $sparsematrixoption){$sparse_matrix = $sparsematrixoption;} |
|
|
0
|
|
|
|
|
|
|
|
122
|
0
|
0
|
|
|
|
|
if(defined $prefixoption){$prefix = $prefixoption;} |
|
|
0
|
|
|
|
|
|
|
|
123
|
0
|
0
|
|
|
|
|
if(defined $suffixoption){$suffix = $suffixoption;} |
|
|
0
|
|
|
|
|
|
|
|
124
|
0
|
0
|
|
|
|
|
if(defined $wcsoption){$wcs = $wcsoption;} |
|
|
0
|
|
|
|
|
|
|
|
125
|
0
|
0
|
|
|
|
|
if(defined $bucketsNumoption){$bucketsNum = $bucketsNumoption;} |
|
|
0
|
|
|
|
|
|
|
|
126
|
0
|
0
|
|
|
|
|
if(defined $ftsoption){@features = split(' ', $ftsoption);} |
|
|
0
|
|
|
|
|
|
|
|
127
|
|
|
|
|
|
|
} |
|
128
|
|
|
|
|
|
|
|
|
129
|
|
|
|
|
|
|
|
|
130
|
|
|
|
|
|
|
####### ARFFMAN AND THE METHODS OF MADNESS ##### |
|
131
|
|
|
|
|
|
|
|
|
132
|
|
|
|
|
|
|
|
|
133
|
|
|
|
|
|
|
# opens a single file and runs it through the process of creating buckets |
|
134
|
|
|
|
|
|
|
# extracting tokens and concepts, and creating arff files based on the features given |
|
135
|
|
|
|
|
|
|
# input : $file <-- the name of the file to make into arff files |
|
136
|
|
|
|
|
|
|
# output : a set of arff files |
|
137
|
|
|
|
|
|
|
sub arff_file{ |
|
138
|
0
|
|
|
0
|
0
|
|
my $self = shift; |
|
139
|
0
|
|
|
|
|
|
my $file = shift; |
|
140
|
|
|
|
|
|
|
|
|
141
|
|
|
|
|
|
|
#define and reset temp var |
|
142
|
0
|
|
|
|
|
|
my $indexer = 0; |
|
143
|
0
|
|
|
|
|
|
%fileHash = (); |
|
144
|
0
|
|
|
|
|
|
%metamapHash = (); |
|
145
|
0
|
|
|
|
|
|
%featHash = (); |
|
146
|
0
|
|
|
|
|
|
%orthoHash = (); |
|
147
|
|
|
|
|
|
|
|
|
148
|
|
|
|
|
|
|
#get the name of the file |
|
149
|
0
|
|
|
|
|
|
my @n = split '/', $file; |
|
150
|
0
|
|
|
|
|
|
my $l = @n; |
|
151
|
0
|
|
|
|
|
|
my $filename = $n[$l - 1]; |
|
152
|
0
|
|
|
|
|
|
$filename = lc($filename); |
|
153
|
|
|
|
|
|
|
|
|
154
|
0
|
|
|
|
|
|
my $FILE; |
|
155
|
0
|
0
|
|
|
|
|
open ($FILE, "$program_dir/$file") || die ("what is this '$program_dir/$filename' you speak of?\n$!"); |
|
156
|
0
|
|
|
|
|
|
my @fileLines = <$FILE>; |
|
157
|
0
|
|
|
|
|
|
my @orthoLines = @fileLines; |
|
158
|
0
|
|
|
|
|
|
foreach my $l(@fileLines){ |
|
159
|
0
|
|
|
|
|
|
$l = lc($l); |
|
160
|
|
|
|
|
|
|
} |
|
161
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("on_red", "$filename"); |
|
162
|
|
|
|
|
|
|
#$uniSub->printColorDebug("on_cyan", "*** $wcs ***"); |
|
163
|
|
|
|
|
|
|
|
|
164
|
|
|
|
|
|
|
#get the total num of lines |
|
165
|
0
|
|
|
|
|
|
my $totalLines = 0; |
|
166
|
0
|
|
|
|
|
|
$totalLines = @fileLines; |
|
167
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("red", "Lines: $totalLines\n"); |
|
168
|
|
|
|
|
|
|
|
|
169
|
|
|
|
|
|
|
#clean it up for two separate sets |
|
170
|
0
|
|
|
|
|
|
my @tagSet = retagSet($filename, \@fileLines); |
|
171
|
0
|
|
|
|
|
|
my @cleanLines = untagSet($filename, \@fileLines); |
|
172
|
|
|
|
|
|
|
|
|
173
|
|
|
|
|
|
|
#get the orthographic based lines |
|
174
|
|
|
|
|
|
|
#my @orthoLines = ; |
|
175
|
0
|
|
|
|
|
|
@orthoLines = retagSetOrtho($file, \@orthoLines); |
|
176
|
|
|
|
|
|
|
|
|
177
|
|
|
|
|
|
|
#$uniSub->printColorDebug("red", "TAG SET: "); |
|
178
|
|
|
|
|
|
|
#$uniSub->printArr(", ", \@tagSet); |
|
179
|
|
|
|
|
|
|
|
|
180
|
|
|
|
|
|
|
####### ASSIGN THE VALUES TO HASHTABLES O KEEP TRACK OF THEM ####### |
|
181
|
|
|
|
|
|
|
|
|
182
|
|
|
|
|
|
|
#put all the lines in a file hash |
|
183
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("blue", "*Putting all the file lines into a hashtable....\n"); |
|
184
|
0
|
|
|
|
|
|
$indexer = 0; |
|
185
|
0
|
|
|
|
|
|
foreach my $line (@tagSet){ |
|
186
|
0
|
|
|
|
|
|
$fileHash{$indexer} = $line; |
|
187
|
0
|
|
|
|
|
|
$indexer++; |
|
188
|
|
|
|
|
|
|
} |
|
189
|
|
|
|
|
|
|
|
|
190
|
|
|
|
|
|
|
#put the orthographic lines in a hash |
|
191
|
0
|
|
|
|
|
|
$indexer = 0; |
|
192
|
0
|
|
|
|
|
|
foreach my $line (@orthoLines){ |
|
193
|
|
|
|
|
|
|
#$uniSub->printColorDebug("red", "$line\n"); |
|
194
|
0
|
|
|
|
|
|
$orthoHash{$indexer} = $line; |
|
195
|
0
|
|
|
|
|
|
$indexer++; |
|
196
|
|
|
|
|
|
|
} |
|
197
|
|
|
|
|
|
|
|
|
198
|
|
|
|
|
|
|
#import the hashtables from saved data |
|
199
|
0
|
|
|
|
|
|
importMetaData($filename); |
|
200
|
|
|
|
|
|
|
|
|
201
|
|
|
|
|
|
|
#tokenize all the lines --> tokenhash |
|
202
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("blue", "*Tokenizing the lines into a hashtable....\n"); |
|
203
|
0
|
|
|
|
|
|
$indexer = 0; |
|
204
|
0
|
|
|
|
|
|
my $totalTokens = 0; |
|
205
|
0
|
|
|
|
|
|
my $totalConcepts = 0; |
|
206
|
0
|
|
|
|
|
|
foreach my $line (@cleanLines){ |
|
207
|
0
|
|
|
|
|
|
$| = 1; |
|
208
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bright_blue", "\r" . "\t\tLine - $indexer/$totalLines"); |
|
209
|
|
|
|
|
|
|
|
|
210
|
|
|
|
|
|
|
#acquire the necessary variables |
|
211
|
0
|
|
|
|
|
|
my $special_ID = "1.ti.1"; |
|
212
|
0
|
|
|
|
|
|
my $meta = $metamapHash{$indexer}; |
|
213
|
|
|
|
|
|
|
|
|
214
|
|
|
|
|
|
|
#create citation first |
|
215
|
0
|
|
|
|
|
|
$dataStructures->createFromTextWithId($meta, $special_ID); |
|
216
|
0
|
|
|
|
|
|
my $citation = $dataStructures->getCitationWithId($special_ID); |
|
217
|
|
|
|
|
|
|
|
|
218
|
|
|
|
|
|
|
#get tokens |
|
219
|
0
|
|
|
|
|
|
my @tokens = @{$citation->getOrderedTokens()}; |
|
|
0
|
|
|
|
|
|
|
|
220
|
0
|
|
|
|
|
|
my $tnum = @tokens; |
|
221
|
0
|
|
|
|
|
|
$totalTokens += $tnum; |
|
222
|
|
|
|
|
|
|
|
|
223
|
|
|
|
|
|
|
#get concepts |
|
224
|
0
|
|
|
|
|
|
my @concepts = @{$citation->getOrderedConcepts()}; |
|
|
0
|
|
|
|
|
|
|
|
225
|
0
|
|
|
|
|
|
my $cnum = @concepts; |
|
226
|
0
|
|
|
|
|
|
$totalConcepts += $cnum; |
|
227
|
|
|
|
|
|
|
|
|
228
|
|
|
|
|
|
|
#grab EVERYTHING |
|
229
|
|
|
|
|
|
|
|
|
230
|
|
|
|
|
|
|
#get tokens first |
|
231
|
0
|
|
|
|
|
|
foreach my $token (@tokens){ |
|
232
|
|
|
|
|
|
|
|
|
233
|
|
|
|
|
|
|
#print "token "; |
|
234
|
|
|
|
|
|
|
#get text and pos |
|
235
|
0
|
|
|
|
|
|
my $super_text = lc($token->{text}); |
|
236
|
|
|
|
|
|
|
|
|
237
|
0
|
|
|
|
|
|
foreach my $text (split(" ", $super_text)){ #split just in case its a compound word |
|
238
|
|
|
|
|
|
|
#check if already there |
|
239
|
0
|
0
|
|
|
|
|
if($featHash{$text}){ |
|
240
|
0
|
|
|
|
|
|
next; |
|
241
|
|
|
|
|
|
|
} |
|
242
|
|
|
|
|
|
|
|
|
243
|
0
|
|
|
|
|
|
my $pos = $token->{posTag}; |
|
244
|
0
|
|
|
|
|
|
$featHash{$text}->{pos} = $pos; |
|
245
|
|
|
|
|
|
|
} |
|
246
|
|
|
|
|
|
|
} |
|
247
|
|
|
|
|
|
|
#get concepts second |
|
248
|
0
|
|
|
|
|
|
foreach my $cc (@concepts){ |
|
249
|
0
|
|
|
|
|
|
foreach my $concept (@{$cc}[0]){ #two levels for some strange reason |
|
|
0
|
|
|
|
|
|
|
|
250
|
|
|
|
|
|
|
#print "token "; |
|
251
|
0
|
|
|
|
|
|
my $super_text = lc($concept->{text}); |
|
252
|
0
|
|
|
|
|
|
for my $text (split(" ", $super_text)){ #split just in case its a compound word |
|
253
|
|
|
|
|
|
|
#grab original hash |
|
254
|
0
|
|
|
|
|
|
my $hash_ref = $featHash{$text}; |
|
255
|
0
|
0
|
0
|
|
|
|
if (!$hash_ref || $hash_ref->{cui} || $hash_ref->{sem}){ |
|
|
|
|
0
|
|
|
|
|
|
256
|
0
|
|
|
|
|
|
next; |
|
257
|
|
|
|
|
|
|
} |
|
258
|
|
|
|
|
|
|
|
|
259
|
|
|
|
|
|
|
#get sem and cuis v_1 |
|
260
|
|
|
|
|
|
|
#$hash_ref->{cui} = $concept->{cui}; |
|
261
|
|
|
|
|
|
|
#$hash_ref->{sem} = $concept->{semanticTypes}; |
|
262
|
|
|
|
|
|
|
|
|
263
|
|
|
|
|
|
|
#print "$text --> "; |
|
264
|
|
|
|
|
|
|
#$uniSub->printArr(", ", \@sem); |
|
265
|
|
|
|
|
|
|
|
|
266
|
|
|
|
|
|
|
#get sem and cuis v_2 |
|
267
|
0
|
0
|
|
|
|
|
if($hash_ref->{cui}){ |
|
268
|
0
|
|
|
|
|
|
$hash_ref->{cui} = ($hash_ref->{cui} . $concept->{cui}); |
|
269
|
|
|
|
|
|
|
}else{ |
|
270
|
0
|
|
|
|
|
|
$hash_ref->{cui} = $concept->{cui}; |
|
271
|
|
|
|
|
|
|
} |
|
272
|
|
|
|
|
|
|
|
|
273
|
0
|
0
|
|
|
|
|
if($hash_ref->{sem}){ |
|
274
|
0
|
|
|
|
|
|
$hash_ref->{sem} = ($hash_ref->{sem} . $concept->{semanticTypes}); |
|
275
|
|
|
|
|
|
|
}else{ |
|
276
|
0
|
|
|
|
|
|
$hash_ref->{sem} = $concept->{semanticTypes}; |
|
277
|
|
|
|
|
|
|
} |
|
278
|
|
|
|
|
|
|
|
|
279
|
|
|
|
|
|
|
|
|
280
|
|
|
|
|
|
|
#print "$text --> "; |
|
281
|
|
|
|
|
|
|
#$uniSub->printArr(", ", \@sem); |
|
282
|
|
|
|
|
|
|
} |
|
283
|
|
|
|
|
|
|
} |
|
284
|
|
|
|
|
|
|
} |
|
285
|
|
|
|
|
|
|
|
|
286
|
|
|
|
|
|
|
#increment to the next set |
|
287
|
0
|
|
|
|
|
|
$indexer++; |
|
288
|
|
|
|
|
|
|
} |
|
289
|
|
|
|
|
|
|
|
|
290
|
|
|
|
|
|
|
#clean up everything |
|
291
|
0
|
|
|
|
|
|
foreach my $key(keys %featHash){ |
|
292
|
0
|
|
|
|
|
|
my $hash_ref = $featHash{$key}; |
|
293
|
0
|
|
|
|
|
|
my %hash = %$hash_ref; |
|
294
|
0
|
0
|
|
|
|
|
if($hash{cui}){ |
|
295
|
0
|
|
|
|
|
|
$hash{cui} = join(',', uniq(split(',', $hash{cui}))); |
|
296
|
|
|
|
|
|
|
} |
|
297
|
0
|
0
|
|
|
|
|
if($hash{sem}){ |
|
298
|
0
|
|
|
|
|
|
$hash{sem} = join(',', uniq(split(',', $hash{sem}))); |
|
299
|
|
|
|
|
|
|
} |
|
300
|
|
|
|
|
|
|
} |
|
301
|
|
|
|
|
|
|
|
|
302
|
|
|
|
|
|
|
#print it back |
|
303
|
0
|
|
|
|
|
|
print("\n"); |
|
304
|
|
|
|
|
|
|
#hash_debug(); |
|
305
|
|
|
|
|
|
|
|
|
306
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("red", "TOKENS: $totalTokens\n"); |
|
307
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("red", "CONCEPTS: $totalConcepts\n"); |
|
308
|
|
|
|
|
|
|
|
|
309
|
|
|
|
|
|
|
#exit; |
|
310
|
|
|
|
|
|
|
|
|
311
|
|
|
|
|
|
|
|
|
312
|
|
|
|
|
|
|
####### BUCKET SORTING - TRAIN AND TEST DATA ####### |
|
313
|
|
|
|
|
|
|
|
|
314
|
|
|
|
|
|
|
#sort the lines to buckets |
|
315
|
0
|
|
|
|
|
|
my %buckets = (); |
|
316
|
0
|
0
|
|
|
|
|
if($bucketsNum > 1){ |
|
317
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("blue", "*Making buckets....\n"); |
|
318
|
0
|
|
|
|
|
|
%buckets = sort2Buckets($totalLines, $bucketsNum); |
|
319
|
|
|
|
|
|
|
}else{ |
|
320
|
0
|
|
|
|
|
|
push(@{$buckets{1}}, (0..($totalLines-1))); |
|
|
0
|
|
|
|
|
|
|
|
321
|
|
|
|
|
|
|
} |
|
322
|
|
|
|
|
|
|
|
|
323
|
|
|
|
|
|
|
#$uniSub->printArr(" ", $buckets{1}); |
|
324
|
|
|
|
|
|
|
|
|
325
|
|
|
|
|
|
|
|
|
326
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("blue", "*Making train and test files....\n"); |
|
327
|
|
|
|
|
|
|
|
|
328
|
0
|
|
|
|
|
|
process_file($filename, \%buckets); |
|
329
|
|
|
|
|
|
|
|
|
330
|
0
|
|
|
|
|
|
$uniSub->printDebug("\n"); |
|
331
|
|
|
|
|
|
|
|
|
332
|
|
|
|
|
|
|
} |
|
333
|
|
|
|
|
|
|
|
|
334
|
|
|
|
|
|
|
#print out the contents of the feature hash |
|
335
|
|
|
|
|
|
|
sub hash_debug{ |
|
336
|
0
|
|
|
0
|
0
|
|
foreach my $key (sort keys %featHash){ |
|
337
|
0
|
|
|
|
|
|
my $text = $key; |
|
338
|
0
|
|
|
|
|
|
my $hash_ref = $featHash{$key}; |
|
339
|
0
|
|
|
|
|
|
my %hash = %$hash_ref; |
|
340
|
|
|
|
|
|
|
|
|
341
|
0
|
|
|
|
|
|
my $pos = ""; |
|
342
|
0
|
|
|
|
|
|
my $cui = ""; |
|
343
|
0
|
|
|
|
|
|
my $sem = ""; |
|
344
|
|
|
|
|
|
|
|
|
345
|
0
|
0
|
|
|
|
|
if($hash{pos}){ |
|
346
|
0
|
|
|
|
|
|
$pos = $hash{pos}; |
|
347
|
|
|
|
|
|
|
} |
|
348
|
0
|
0
|
|
|
|
|
if($hash{cui}){ |
|
349
|
0
|
|
|
|
|
|
$cui = join(',', @{$hash{cui}}); |
|
|
0
|
|
|
|
|
|
|
|
350
|
|
|
|
|
|
|
} |
|
351
|
0
|
0
|
|
|
|
|
if($hash{sem}){ |
|
352
|
0
|
|
|
|
|
|
$sem = join(',', @{$hash{sem}}); |
|
|
0
|
|
|
|
|
|
|
|
353
|
|
|
|
|
|
|
} |
|
354
|
|
|
|
|
|
|
|
|
355
|
0
|
|
|
|
|
|
print "$text - $pos - $cui - $sem\n"; |
|
356
|
|
|
|
|
|
|
} |
|
357
|
|
|
|
|
|
|
} |
|
358
|
|
|
|
|
|
|
|
|
359
|
|
|
|
|
|
|
|
|
360
|
|
|
|
|
|
|
###################### LINE MANIPULATION ##################### |
|
361
|
|
|
|
|
|
|
|
|
362
|
|
|
|
|
|
|
###### RETAGS THE LINE ###### |
|
363
|
|
|
|
|
|
|
|
|
364
|
|
|
|
|
|
|
# turns the tagged entity words into special words with <> for the context words |
|
365
|
|
|
|
|
|
|
# input : $input <-- the line to retag |
|
366
|
|
|
|
|
|
|
# $id <-- the id within the tag to look for |
|
367
|
|
|
|
|
|
|
# output : (.arff files) |
|
368
|
|
|
|
|
|
|
sub retag{ |
|
369
|
0
|
|
|
0
|
0
|
|
my $input = shift; |
|
370
|
0
|
|
|
|
|
|
my $id = shift; |
|
371
|
|
|
|
|
|
|
|
|
372
|
0
|
|
|
|
|
|
$id = lc($id); |
|
373
|
0
|
|
|
|
|
|
my $line = lc($input); |
|
374
|
|
|
|
|
|
|
|
|
375
|
|
|
|
|
|
|
#get rid of any tags |
|
376
|
0
|
|
|
|
|
|
my @words = split (" ", $line); |
|
377
|
0
|
|
|
|
|
|
my @newSet = (); |
|
378
|
0
|
|
|
|
|
|
my $charact = 0; |
|
379
|
0
|
|
|
|
|
|
foreach my $word (@words){ |
|
380
|
|
|
|
|
|
|
#if word is currently in the annotation |
|
381
|
0
|
0
|
|
|
|
|
if($charact){ |
|
382
|
0
|
0
|
|
|
|
|
if($word =~//){ #reached the end |
|
383
|
0
|
|
|
|
|
|
$charact = 0; |
|
384
|
|
|
|
|
|
|
}else{ #continue adding extension to the annotation set |
|
385
|
0
|
|
|
|
|
|
my $charWord = "$word"."$entId"; |
|
386
|
0
|
|
|
|
|
|
push @newSet, $charWord; |
|
387
|
|
|
|
|
|
|
} |
|
388
|
|
|
|
|
|
|
}else{ |
|
389
|
0
|
0
|
|
|
|
|
if($word =~//){ #start annotation set |
|
390
|
0
|
|
|
|
|
|
$charact = 1; |
|
391
|
|
|
|
|
|
|
}else{ #continue as normal |
|
392
|
0
|
|
|
|
|
|
push @newSet, $word; |
|
393
|
|
|
|
|
|
|
} |
|
394
|
|
|
|
|
|
|
} |
|
395
|
|
|
|
|
|
|
} |
|
396
|
|
|
|
|
|
|
|
|
397
|
|
|
|
|
|
|
#clean up the new line |
|
398
|
0
|
|
|
|
|
|
my $new_line = join " ", @newSet; |
|
399
|
0
|
|
|
|
|
|
$new_line =~s/\b$entId\b//g; |
|
400
|
0
|
|
|
|
|
|
$new_line = $uniSub->cleanWords($new_line); |
|
401
|
0
|
|
|
|
|
|
return $new_line; |
|
402
|
|
|
|
|
|
|
} |
|
403
|
|
|
|
|
|
|
|
|
404
|
|
|
|
|
|
|
# turns the tagged entity words in the entire file into special words with <> for the context words |
|
405
|
|
|
|
|
|
|
# input : $name <-- the name of the file to use as the id tag |
|
406
|
|
|
|
|
|
|
# @lines <-- the set of lines to retag |
|
407
|
|
|
|
|
|
|
# output : @tagSet <-- set of retagged lines |
|
408
|
|
|
|
|
|
|
sub retagSet{ |
|
409
|
0
|
|
|
0
|
0
|
|
my $name = shift; |
|
410
|
0
|
|
|
|
|
|
my $lines_ref = shift; |
|
411
|
0
|
|
|
|
|
|
my @lines = @$lines_ref; |
|
412
|
|
|
|
|
|
|
|
|
413
|
0
|
|
|
|
|
|
my @tagSet = (); |
|
414
|
0
|
|
|
|
|
|
foreach my $line (@lines){ |
|
415
|
|
|
|
|
|
|
#retag the line |
|
416
|
0
|
|
|
|
|
|
chomp($line); |
|
417
|
0
|
|
|
|
|
|
my $tag_line = retag($line, $name); |
|
418
|
|
|
|
|
|
|
|
|
419
|
|
|
|
|
|
|
#add it to the set |
|
420
|
0
|
|
|
|
|
|
push @tagSet, $tag_line; |
|
421
|
|
|
|
|
|
|
} |
|
422
|
0
|
|
|
|
|
|
return @tagSet; |
|
423
|
|
|
|
|
|
|
} |
|
424
|
|
|
|
|
|
|
|
|
425
|
|
|
|
|
|
|
#returns clean line with no tags or retaggings |
|
426
|
|
|
|
|
|
|
# input : $line <-- the line to untag |
|
427
|
|
|
|
|
|
|
# : $id <-- the id label to look for |
|
428
|
|
|
|
|
|
|
# output : $input <-- untagged input line |
|
429
|
|
|
|
|
|
|
sub untag{ |
|
430
|
0
|
|
|
0
|
0
|
|
my $line = shift; |
|
431
|
0
|
|
|
|
|
|
my $id = shift; |
|
432
|
|
|
|
|
|
|
|
|
433
|
|
|
|
|
|
|
#remove and |
|
434
|
0
|
|
|
|
|
|
my $input = lc($line); |
|
435
|
0
|
|
|
|
|
|
$id = lc($id); |
|
436
|
0
|
|
|
|
|
|
$input =~ s///g; |
|
437
|
0
|
|
|
|
|
|
$input =~ s///g; |
|
438
|
0
|
|
|
|
|
|
$input = $uniSub->cleanWords($input); |
|
439
|
0
|
|
|
|
|
|
return $input; |
|
440
|
|
|
|
|
|
|
} |
|
441
|
|
|
|
|
|
|
#returns a clean set of lines |
|
442
|
|
|
|
|
|
|
# input : $filename <-- the name of the file for use in the id tag |
|
443
|
|
|
|
|
|
|
# : @lines <-- the set of lines to untag |
|
444
|
|
|
|
|
|
|
# output : @clean_set <-- untagged set of lines |
|
445
|
|
|
|
|
|
|
sub untagSet{ |
|
446
|
0
|
|
|
0
|
0
|
|
my $filename = shift; |
|
447
|
0
|
|
|
|
|
|
my $lines_ref = shift; |
|
448
|
0
|
|
|
|
|
|
my @lines = @$lines_ref; |
|
449
|
|
|
|
|
|
|
|
|
450
|
0
|
|
|
|
|
|
my @clean_set = (); |
|
451
|
|
|
|
|
|
|
#run untagger on each line |
|
452
|
0
|
|
|
|
|
|
foreach my $line(@lines){ |
|
453
|
0
|
|
|
|
|
|
my $cl = untag($line, $filename); |
|
454
|
0
|
|
|
|
|
|
push @clean_set, $cl; |
|
455
|
|
|
|
|
|
|
} |
|
456
|
0
|
|
|
|
|
|
return @clean_set; |
|
457
|
|
|
|
|
|
|
} |
|
458
|
|
|
|
|
|
|
|
|
459
|
|
|
|
|
|
|
|
|
460
|
|
|
|
|
|
|
#import metamap hashtable data |
|
461
|
|
|
|
|
|
|
# input : $name <-- the name of the file to import from |
|
462
|
|
|
|
|
|
|
# output : (hashmap of metamap lines) |
|
463
|
|
|
|
|
|
|
sub importMetaData{ |
|
464
|
0
|
|
|
0
|
0
|
|
my $name = shift; |
|
465
|
|
|
|
|
|
|
|
|
466
|
|
|
|
|
|
|
#create a directory to save hashtable data |
|
467
|
0
|
|
|
|
|
|
my $META; |
|
468
|
0
|
|
|
|
|
|
my $subdir = "_METAMAPS"; |
|
469
|
0
|
0
|
|
|
|
|
open($META, "<", ("$program_dir/$subdir/" . $name . "_meta")) || die ("Metamap file for " . $name . " not found! Check the _METAMAP folder in the file directory\n$!"); |
|
470
|
|
|
|
|
|
|
|
|
471
|
|
|
|
|
|
|
|
|
472
|
|
|
|
|
|
|
#import metamap data from the file |
|
473
|
0
|
|
|
|
|
|
my @metaLines = <$META>; |
|
474
|
0
|
|
|
|
|
|
my $metaCombo = join("", @metaLines); |
|
475
|
0
|
|
|
|
|
|
my @newMetaLines = split("\n\n", $metaCombo); |
|
476
|
0
|
|
|
|
|
|
my $t = @newMetaLines; |
|
477
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("red", "META LINES: $t\n"); |
|
478
|
0
|
|
|
|
|
|
my $key = 0; |
|
479
|
|
|
|
|
|
|
#assign each metamap line to a hash indexer |
|
480
|
0
|
|
|
|
|
|
foreach my $mm (@newMetaLines){ |
|
481
|
0
|
|
|
|
|
|
$metamapHash{$key} = $mm; |
|
482
|
0
|
|
|
|
|
|
$key++; |
|
483
|
|
|
|
|
|
|
} |
|
484
|
0
|
|
|
|
|
|
close $META; |
|
485
|
|
|
|
|
|
|
} |
|
486
|
|
|
|
|
|
|
|
|
487
|
|
|
|
|
|
|
##### FOR THE ORTHO SET ##### |
|
488
|
|
|
|
|
|
|
|
|
489
|
|
|
|
|
|
|
#turns the tagged entity words into special words with <> for the context words |
|
490
|
|
|
|
|
|
|
# input : $input <-- the line to retag |
|
491
|
|
|
|
|
|
|
# output : $new_line <-- the retagged line |
|
492
|
|
|
|
|
|
|
sub retagOrtho{ |
|
493
|
0
|
|
|
0
|
0
|
|
my $filename = shift; |
|
494
|
0
|
|
|
|
|
|
my $input = shift; |
|
495
|
0
|
|
|
|
|
|
my $line = $input; |
|
496
|
|
|
|
|
|
|
|
|
497
|
|
|
|
|
|
|
#get rid of any tags |
|
498
|
0
|
|
|
|
|
|
my @words = split (" ", $line); |
|
499
|
0
|
|
|
|
|
|
my @newSet = (); |
|
500
|
0
|
|
|
|
|
|
my $charact = 0; |
|
501
|
0
|
|
|
|
|
|
foreach my $word (@words){ |
|
502
|
0
|
0
|
|
|
|
|
if($charact){ #currently annotating |
|
503
|
0
|
0
|
|
|
|
|
if($word =~//){ #stop annotation set |
|
504
|
0
|
|
|
|
|
|
$charact = 0; |
|
505
|
|
|
|
|
|
|
}else{ |
|
506
|
0
|
|
|
|
|
|
my $charWord = "$word"."$entId"; #add extension to annotation word |
|
507
|
0
|
|
|
|
|
|
push @newSet, $charWord; |
|
508
|
|
|
|
|
|
|
} |
|
509
|
|
|
|
|
|
|
}else{ |
|
510
|
0
|
0
|
|
|
|
|
if($word =~//){ #start annotating |
|
511
|
0
|
|
|
|
|
|
$charact = 1; |
|
512
|
|
|
|
|
|
|
}else{ #continue as normal |
|
513
|
0
|
|
|
|
|
|
push @newSet, $word; |
|
514
|
|
|
|
|
|
|
} |
|
515
|
|
|
|
|
|
|
} |
|
516
|
|
|
|
|
|
|
} |
|
517
|
|
|
|
|
|
|
|
|
518
|
|
|
|
|
|
|
#clean up the new line |
|
519
|
0
|
|
|
|
|
|
my $new_line = join " ", @newSet; |
|
520
|
0
|
|
|
|
|
|
$new_line =~s/\s$entId\b//g; |
|
521
|
0
|
|
|
|
|
|
$new_line = noASCIIOrtho($new_line); |
|
522
|
0
|
|
|
|
|
|
return $new_line; |
|
523
|
|
|
|
|
|
|
} |
|
524
|
|
|
|
|
|
|
#turns the tagged entity words in the entire file into special words with <> for the context words |
|
525
|
|
|
|
|
|
|
# input : @lines <-- the set of lines to retag |
|
526
|
|
|
|
|
|
|
# output : @tagSet <-- the retagged line |
|
527
|
|
|
|
|
|
|
sub retagSetOrtho{ |
|
528
|
0
|
|
|
0
|
0
|
|
my $file = shift; |
|
529
|
0
|
|
|
|
|
|
my $lines_ref = shift; |
|
530
|
0
|
|
|
|
|
|
my @lines = @$lines_ref; |
|
531
|
|
|
|
|
|
|
|
|
532
|
0
|
|
|
|
|
|
my @n = split '/', $file; |
|
533
|
0
|
|
|
|
|
|
my $l = @n; |
|
534
|
0
|
|
|
|
|
|
my $filename = $n[$l - 1]; |
|
535
|
|
|
|
|
|
|
|
|
536
|
0
|
|
|
|
|
|
my @tagSet = (); |
|
537
|
0
|
|
|
|
|
|
foreach my $line (@lines){ |
|
538
|
|
|
|
|
|
|
#retag the line |
|
539
|
0
|
|
|
|
|
|
chomp($line); |
|
540
|
0
|
|
|
|
|
|
my $tag_line = retagOrtho($filename, $line); |
|
541
|
|
|
|
|
|
|
|
|
542
|
|
|
|
|
|
|
#add it to the set |
|
543
|
0
|
|
|
|
|
|
push @tagSet, $tag_line; |
|
544
|
|
|
|
|
|
|
} |
|
545
|
0
|
|
|
|
|
|
return @tagSet; |
|
546
|
|
|
|
|
|
|
} |
|
547
|
|
|
|
|
|
|
|
|
548
|
|
|
|
|
|
|
#cleans the line without getting rid of tags |
|
549
|
|
|
|
|
|
|
# input : $line <-- line to clean up |
|
550
|
|
|
|
|
|
|
# output : $new_in <-- the cleaned line |
|
551
|
|
|
|
|
|
|
sub noASCIIOrtho{ |
|
552
|
0
|
|
|
0
|
0
|
|
my $line = shift; |
|
553
|
|
|
|
|
|
|
|
|
554
|
0
|
|
|
|
|
|
my $new_in = $line; |
|
555
|
0
|
|
|
|
|
|
$new_in =~ s/[^[:ascii:]]//g; #remove any words that do not contain ASCII characters |
|
556
|
0
|
|
|
|
|
|
return $new_in |
|
557
|
|
|
|
|
|
|
} |
|
558
|
|
|
|
|
|
|
|
|
559
|
|
|
|
|
|
|
|
|
560
|
|
|
|
|
|
|
####################### TOKENS AND CONCEPT MANIPULATION ####################### |
|
561
|
|
|
|
|
|
|
|
|
562
|
|
|
|
|
|
|
|
|
563
|
|
|
|
|
|
|
#gets rid of any special tokens |
|
564
|
|
|
|
|
|
|
# input : $text <-- the token text to fix |
|
565
|
|
|
|
|
|
|
# output : $tokenText <-- a cleaned up token |
|
566
|
|
|
|
|
|
|
sub cleanToken{ |
|
567
|
0
|
|
|
0
|
0
|
|
my $text = shift; |
|
568
|
|
|
|
|
|
|
|
|
569
|
0
|
|
|
|
|
|
my $tokenText = $text; |
|
570
|
|
|
|
|
|
|
|
|
571
|
|
|
|
|
|
|
#fix "# . #" tokens, fix "__ \' __" tokens, remove any non word based characters |
|
572
|
0
|
|
|
|
|
|
$tokenText =~s/\s\.\s/\./og; |
|
573
|
0
|
|
|
|
|
|
$tokenText =~s/\s\\\'\s//og; |
|
574
|
0
|
|
|
|
|
|
$tokenText =~s/[^a-zA-Z0-9]//og; |
|
575
|
|
|
|
|
|
|
|
|
576
|
0
|
|
|
|
|
|
return $tokenText; |
|
577
|
|
|
|
|
|
|
} |
|
578
|
|
|
|
|
|
|
|
|
579
|
|
|
|
|
|
|
#retrieves the feature for a single word |
|
580
|
|
|
|
|
|
|
# input : $word <-- the word to extract the features from |
|
581
|
|
|
|
|
|
|
# : $type <-- what type of feature to extract [e.g. "pos", "sem", "cui"] |
|
582
|
|
|
|
|
|
|
# output : if "pos" <-- a scalar part-of-speech value |
|
583
|
|
|
|
|
|
|
# : else <-- an array of semantic or cui values (a single text value can have more than one of these) |
|
584
|
|
|
|
|
|
|
sub getFeature{ |
|
585
|
0
|
|
|
0
|
0
|
|
my $orig_word = shift; |
|
586
|
0
|
|
|
|
|
|
my $type = shift; |
|
587
|
|
|
|
|
|
|
|
|
588
|
|
|
|
|
|
|
#clean up thw word |
|
589
|
0
|
|
|
|
|
|
my $word = lc($orig_word); |
|
590
|
0
|
|
|
|
|
|
$word =~s/[^a-zA-Z0-9\s]//; |
|
591
|
|
|
|
|
|
|
|
|
592
|
0
|
0
|
|
|
|
|
if(!$featHash{$word}){ |
|
593
|
0
|
|
|
|
|
|
return ""; |
|
594
|
|
|
|
|
|
|
} |
|
595
|
|
|
|
|
|
|
|
|
596
|
0
|
|
|
|
|
|
my $hash_ref = $featHash{$word}; |
|
597
|
0
|
|
|
|
|
|
my %hash = %$hash_ref; |
|
598
|
|
|
|
|
|
|
|
|
599
|
0
|
0
|
|
|
|
|
if($hash{$type}){ |
|
600
|
0
|
|
|
|
|
|
return $hash{$type}; |
|
601
|
|
|
|
|
|
|
} |
|
602
|
0
|
|
|
|
|
|
return ""; |
|
603
|
|
|
|
|
|
|
} |
|
604
|
|
|
|
|
|
|
|
|
605
|
|
|
|
|
|
|
###################### BUCKETS - TRAIN AND TEST ARFF FILES ##################### |
|
606
|
|
|
|
|
|
|
|
|
607
|
|
|
|
|
|
|
|
|
608
|
|
|
|
|
|
|
#sorts the keys from the hashmaps into buckets so that certain values can be accessed |
|
609
|
|
|
|
|
|
|
# input : $keyAmt <-- the number of lines or "keys" to divvy up into the buckets |
|
610
|
|
|
|
|
|
|
# : $bucketNum <-- how many buckets to use |
|
611
|
|
|
|
|
|
|
# output : %bucketList <-- the set of buckets with keys in them |
|
612
|
|
|
|
|
|
|
sub sort2Buckets{ |
|
613
|
0
|
|
|
0
|
0
|
|
my $keyAmt = shift; |
|
614
|
0
|
|
|
|
|
|
my $bucketNum = shift; |
|
615
|
|
|
|
|
|
|
|
|
616
|
|
|
|
|
|
|
#create sets |
|
617
|
0
|
|
|
|
|
|
my @keySet = (0..$keyAmt - 1); #set of keys |
|
618
|
0
|
|
|
|
|
|
my %bucketList = (); #all of the buckets |
|
619
|
|
|
|
|
|
|
|
|
620
|
|
|
|
|
|
|
#add some buckets to the bucket list |
|
621
|
0
|
|
|
|
|
|
for(my $a = 1; $a <= $bucketNum; $a++){ |
|
622
|
0
|
|
|
|
|
|
$bucketList{$a} = []; |
|
623
|
|
|
|
|
|
|
} |
|
624
|
|
|
|
|
|
|
|
|
625
|
|
|
|
|
|
|
#sort the lines into buckets |
|
626
|
0
|
|
|
|
|
|
my $bucketId = 1; |
|
627
|
0
|
|
|
|
|
|
foreach my $key (@keySet){ |
|
628
|
0
|
|
|
|
|
|
push (@{$bucketList{$bucketId}}, $key); #add the line to the bucket |
|
|
0
|
|
|
|
|
|
|
|
629
|
|
|
|
|
|
|
|
|
630
|
|
|
|
|
|
|
#reset the id if at the max value |
|
631
|
0
|
0
|
|
|
|
|
if($bucketId == $bucketNum){ |
|
632
|
0
|
|
|
|
|
|
$bucketId = 1; |
|
633
|
|
|
|
|
|
|
}else{ |
|
634
|
0
|
|
|
|
|
|
$bucketId++; |
|
635
|
|
|
|
|
|
|
} |
|
636
|
|
|
|
|
|
|
} |
|
637
|
|
|
|
|
|
|
|
|
638
|
|
|
|
|
|
|
#return the list of buckets |
|
639
|
0
|
|
|
|
|
|
return %bucketList; |
|
640
|
|
|
|
|
|
|
} |
|
641
|
|
|
|
|
|
|
|
|
642
|
|
|
|
|
|
|
###################### ARFF STUFF ##################### |
|
643
|
|
|
|
|
|
|
#makes arff files for ortho, morpho, text, pos, cui, and sem attributes |
|
644
|
|
|
|
|
|
|
|
|
645
|
|
|
|
|
|
|
# Processes the file by retrieving attributes, making vectors, and splitting into buckets |
|
646
|
|
|
|
|
|
|
# (formally known as 'zhu li!! Do the thing!!'') |
|
647
|
|
|
|
|
|
|
# input : $name <-- the name of the file |
|
648
|
|
|
|
|
|
|
# : %bucketList <-- the set of buckets with keys in them |
|
649
|
|
|
|
|
|
|
# output : (n arff files; n = # of buckets x (train and test) x # of features being used) |
|
650
|
|
|
|
|
|
|
sub process_file{ |
|
651
|
0
|
|
|
0
|
0
|
|
my $name = shift; |
|
652
|
0
|
|
|
|
|
|
my $bucketList_ref = shift; |
|
653
|
0
|
|
|
|
|
|
my %buckets = %$bucketList_ref; |
|
654
|
|
|
|
|
|
|
|
|
655
|
|
|
|
|
|
|
#grab the attributes |
|
656
|
0
|
|
|
|
|
|
my %attrSets = (); |
|
657
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bold green", "Retrieving attributes...\n"); |
|
658
|
0
|
|
|
|
|
|
foreach my $item(@features){ |
|
659
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bright_green", "\t$item attr\n"); |
|
660
|
0
|
|
|
|
|
|
my %setOfAttr = grabAttr($name, $item, \%buckets); |
|
661
|
0
|
|
|
|
|
|
$attrSets{$item} = \%setOfAttr; #gets both the vector and arff based attributes |
|
662
|
|
|
|
|
|
|
} |
|
663
|
|
|
|
|
|
|
|
|
664
|
|
|
|
|
|
|
#contain the stop words regular expressions if the parameter was defined |
|
665
|
0
|
0
|
|
|
|
|
if(defined $stopwords_file){ |
|
666
|
0
|
|
|
|
|
|
$stopRegex = stop($stopwords_file); |
|
667
|
|
|
|
|
|
|
} |
|
668
|
|
|
|
|
|
|
|
|
669
|
|
|
|
|
|
|
#let's make some vectors! |
|
670
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bold yellow", "Making Vectors...\n-------------------\n"); |
|
671
|
0
|
|
|
|
|
|
my @curFeatSet = (); |
|
672
|
0
|
|
|
|
|
|
my $abbrev = ""; |
|
673
|
|
|
|
|
|
|
|
|
674
|
|
|
|
|
|
|
#run based on wcs |
|
675
|
0
|
|
|
|
|
|
my $wcs_bucket; |
|
676
|
|
|
|
|
|
|
my $wcs_feature; |
|
677
|
0
|
|
|
|
|
|
my $wcs_found = 0; |
|
678
|
0
|
0
|
|
|
|
|
if($wcs){ |
|
679
|
0
|
|
|
|
|
|
my @wcs_parts = split("-", $wcs); |
|
680
|
0
|
|
|
|
|
|
$wcs_feature = $wcs_parts[1]; |
|
681
|
0
|
|
|
|
|
|
$wcs_bucket = $wcs_parts[0]; |
|
682
|
|
|
|
|
|
|
} |
|
683
|
|
|
|
|
|
|
|
|
684
|
|
|
|
|
|
|
|
|
685
|
|
|
|
|
|
|
#iteratively add on the features [e.g. o, om, omt, omtp, omtpc, omtpcs] |
|
686
|
0
|
|
|
|
|
|
foreach my $feature (@features){ |
|
687
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("yellow", "** $feature ** \n"); |
|
688
|
0
|
|
|
|
|
|
push(@curFeatSet, $feature); |
|
689
|
0
|
|
|
|
|
|
$abbrev .= substr($feature, 0, 1); #add to abbreviations for the name |
|
690
|
|
|
|
|
|
|
|
|
691
|
|
|
|
|
|
|
#$uniSub->printColorDebug("on_red", "$wcs - $wcs_found - $abbrev vs. $wcs_feature"); |
|
692
|
0
|
0
|
0
|
|
|
|
if(($wcs) && (!$wcs_found) && ($abbrev ne $wcs_feature)){ |
|
|
|
|
0
|
|
|
|
|
|
693
|
0
|
|
|
|
|
|
print("**SKIP** \n"); |
|
694
|
0
|
|
|
|
|
|
next; |
|
695
|
|
|
|
|
|
|
} |
|
696
|
|
|
|
|
|
|
|
|
697
|
|
|
|
|
|
|
#go through each bucket |
|
698
|
0
|
|
|
|
|
|
foreach my $bucket (sort keys %buckets){ |
|
699
|
|
|
|
|
|
|
#if wcs parameter defined - skip these buckets until the specified bucket and feature is reached |
|
700
|
0
|
0
|
0
|
|
|
|
if(($wcs) && (!$wcs_found) && ($bucket != $wcs_bucket)){ |
|
|
|
|
0
|
|
|
|
|
|
701
|
0
|
|
|
|
|
|
print("\t**SKIP**\n"); |
|
702
|
0
|
|
|
|
|
|
next; |
|
703
|
|
|
|
|
|
|
}else{ |
|
704
|
0
|
|
|
|
|
|
$wcs_found = 1; |
|
705
|
|
|
|
|
|
|
} |
|
706
|
|
|
|
|
|
|
|
|
707
|
|
|
|
|
|
|
#make train-test bucket indexes |
|
708
|
0
|
|
|
|
|
|
my @range = (); |
|
709
|
0
|
0
|
|
|
|
|
if($bucketsNum > 1){ |
|
710
|
0
|
|
|
|
|
|
@range = $uniSub->bully($bucketsNum, $bucket); |
|
711
|
|
|
|
|
|
|
}else{ |
|
712
|
0
|
|
|
|
|
|
@range = (1); |
|
713
|
|
|
|
|
|
|
} |
|
714
|
|
|
|
|
|
|
|
|
715
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("on_green", "BUCKET #$bucket"); |
|
716
|
|
|
|
|
|
|
#retrieve the vector attributes to use |
|
717
|
0
|
|
|
|
|
|
my %vecAttrSet = (); |
|
718
|
0
|
|
|
|
|
|
foreach my $curItem(@curFeatSet){ |
|
719
|
0
|
0
|
|
|
|
|
if($curItem eq "ortho"){ |
|
720
|
0
|
|
|
|
|
|
$vecAttrSet{$curItem} = (); |
|
721
|
|
|
|
|
|
|
}else{ |
|
722
|
|
|
|
|
|
|
#get outer layer (tpcs) |
|
723
|
0
|
|
|
|
|
|
my $a_ref = $attrSets{$curItem}; |
|
724
|
0
|
|
|
|
|
|
my %a = %$a_ref; |
|
725
|
|
|
|
|
|
|
|
|
726
|
|
|
|
|
|
|
#get inner layer (vector) |
|
727
|
0
|
|
|
|
|
|
my $b_ref = $a{vector}; |
|
728
|
0
|
|
|
|
|
|
my %b = %$b_ref; |
|
729
|
|
|
|
|
|
|
|
|
730
|
|
|
|
|
|
|
#foreach my $key (sort keys %b){print "$key\n";} |
|
731
|
|
|
|
|
|
|
|
|
732
|
|
|
|
|
|
|
#finally get the bucket layer (1..$bucketNum) based on range |
|
733
|
0
|
|
|
|
|
|
my $c_ref = $b{$bucket}; |
|
734
|
0
|
|
|
|
|
|
my @c = @$c_ref; |
|
735
|
0
|
|
|
|
|
|
$vecAttrSet{$curItem} = \@c; |
|
736
|
|
|
|
|
|
|
} |
|
737
|
|
|
|
|
|
|
} |
|
738
|
|
|
|
|
|
|
|
|
739
|
|
|
|
|
|
|
### TRAIN ### |
|
740
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bold blue", "\ttraining...\n"); |
|
741
|
|
|
|
|
|
|
#retrieve the lines to use |
|
742
|
0
|
|
|
|
|
|
my @lineSetTrain = (); |
|
743
|
0
|
|
|
|
|
|
my @bucketSetTrain = (); |
|
744
|
0
|
|
|
|
|
|
foreach my $num (@range){push(@bucketSetTrain, @{$buckets{$num}});} |
|
|
0
|
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
|
|
745
|
0
|
|
|
|
|
|
foreach my $key (@bucketSetTrain){push(@lineSetTrain, $orthoHash{$key});} |
|
|
0
|
|
|
|
|
|
|
|
746
|
|
|
|
|
|
|
|
|
747
|
|
|
|
|
|
|
#make the vector |
|
748
|
0
|
|
|
|
|
|
my @vectorSetTrain = vectorMaker(\@lineSetTrain, \@curFeatSet, \%vecAttrSet); |
|
749
|
0
|
|
|
|
|
|
$uniSub->printDebug("\n"); |
|
750
|
|
|
|
|
|
|
|
|
751
|
|
|
|
|
|
|
### TEST ### |
|
752
|
0
|
|
|
|
|
|
my @vectorSetTest = (); |
|
753
|
0
|
0
|
|
|
|
|
if($bucketsNum > 1){ #skip this if only 1 bucket being used (train bucket) |
|
754
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bold magenta", "\ttesting...\n"); |
|
755
|
|
|
|
|
|
|
#retrieve the lines to use |
|
756
|
0
|
|
|
|
|
|
my @lineSetTest = (); |
|
757
|
0
|
|
|
|
|
|
my @bucketSetTest = (); |
|
758
|
0
|
|
|
|
|
|
push(@bucketSetTest, @{$buckets{$bucket}}); |
|
|
0
|
|
|
|
|
|
|
|
759
|
0
|
|
|
|
|
|
foreach my $key (@bucketSetTest){push(@lineSetTest, $orthoHash{$key});} |
|
|
0
|
|
|
|
|
|
|
|
760
|
|
|
|
|
|
|
|
|
761
|
|
|
|
|
|
|
#make the vector |
|
762
|
0
|
|
|
|
|
|
@vectorSetTest = vectorMaker(\@lineSetTest, \@curFeatSet, \%vecAttrSet); |
|
763
|
0
|
|
|
|
|
|
$uniSub->printDebug("\n"); |
|
764
|
|
|
|
|
|
|
} |
|
765
|
|
|
|
|
|
|
|
|
766
|
|
|
|
|
|
|
|
|
767
|
|
|
|
|
|
|
### ARFF ### |
|
768
|
|
|
|
|
|
|
#retrieve the arff attributes to use |
|
769
|
0
|
|
|
|
|
|
my @arffAttrSet = (); |
|
770
|
0
|
|
|
|
|
|
foreach my $curItem(@curFeatSet){ |
|
771
|
0
|
0
|
|
|
|
|
if($curItem eq "ortho"){ |
|
772
|
|
|
|
|
|
|
#get outer layer (ortho) |
|
773
|
0
|
|
|
|
|
|
my $a_ref = $attrSets{$curItem}; |
|
774
|
0
|
|
|
|
|
|
my %a = %$a_ref; |
|
775
|
|
|
|
|
|
|
#get the values from ortho |
|
776
|
0
|
|
|
|
|
|
push(@arffAttrSet, @{$a{arff}}); |
|
|
0
|
|
|
|
|
|
|
|
777
|
|
|
|
|
|
|
}else{ |
|
778
|
|
|
|
|
|
|
#get outer layer (mtpcs) |
|
779
|
0
|
|
|
|
|
|
my $a_ref = $attrSets{$curItem}; |
|
780
|
0
|
|
|
|
|
|
my %a = %$a_ref; |
|
781
|
|
|
|
|
|
|
|
|
782
|
|
|
|
|
|
|
#get inner layer (arff) |
|
783
|
0
|
|
|
|
|
|
my $b_ref = $a{arff}; |
|
784
|
0
|
|
|
|
|
|
my %b = %$b_ref; |
|
785
|
|
|
|
|
|
|
|
|
786
|
|
|
|
|
|
|
#finally get the bucket layer (1..$bucketNum) based on range |
|
787
|
0
|
|
|
|
|
|
my $c_ref = $b{$bucket}; |
|
788
|
0
|
|
|
|
|
|
my @c = @$c_ref; |
|
789
|
0
|
|
|
|
|
|
push(@arffAttrSet, @c); |
|
790
|
|
|
|
|
|
|
} |
|
791
|
|
|
|
|
|
|
} |
|
792
|
|
|
|
|
|
|
|
|
793
|
|
|
|
|
|
|
#create the arff files for the test and train features |
|
794
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bright_yellow", "\tmaking arff files...\n"); |
|
795
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bright_red", "\t\tARFF TRAIN\n"); |
|
796
|
0
|
|
|
|
|
|
createARFF($name, $bucket, $abbrev, "train", \@arffAttrSet, \@vectorSetTrain); |
|
797
|
0
|
0
|
|
|
|
|
if($bucketsNum > 1){ |
|
798
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bright_red", "\t\tARFF TEST\n"); |
|
799
|
0
|
|
|
|
|
|
createARFF($name, $bucket, $abbrev, "test", \@arffAttrSet, \@vectorSetTest); |
|
800
|
|
|
|
|
|
|
} |
|
801
|
|
|
|
|
|
|
|
|
802
|
|
|
|
|
|
|
} |
|
803
|
|
|
|
|
|
|
} |
|
804
|
|
|
|
|
|
|
} |
|
805
|
|
|
|
|
|
|
|
|
806
|
|
|
|
|
|
|
|
|
807
|
|
|
|
|
|
|
#create the arff file |
|
808
|
|
|
|
|
|
|
# input : $name <-- the name of the file |
|
809
|
|
|
|
|
|
|
# : $bucket <-- the index of the bucket you're testing [e.g. bucket #1] |
|
810
|
|
|
|
|
|
|
# : $abbrev <-- the abbreviation label for the set of features |
|
811
|
|
|
|
|
|
|
# : $type <-- train or test ARFF? |
|
812
|
|
|
|
|
|
|
# : @attrARFFSet <-- the set of attributes exclusively for printing to the arff file |
|
813
|
|
|
|
|
|
|
# : @vecSec <-- the set of vectors created |
|
814
|
|
|
|
|
|
|
# output : (an arff file) |
|
815
|
|
|
|
|
|
|
sub createARFF{ |
|
816
|
0
|
|
|
0
|
0
|
|
my $name = shift; |
|
817
|
0
|
|
|
|
|
|
my $bucket = shift; |
|
818
|
0
|
|
|
|
|
|
my $abbrev = shift; |
|
819
|
0
|
|
|
|
|
|
my $type = shift; |
|
820
|
0
|
|
|
|
|
|
my $attr_ref = shift; |
|
821
|
0
|
|
|
|
|
|
my $vec_ref = shift; |
|
822
|
|
|
|
|
|
|
|
|
823
|
0
|
|
|
|
|
|
my $typeDir = "_$type"; |
|
824
|
0
|
|
|
|
|
|
my $ARFF; |
|
825
|
|
|
|
|
|
|
#print to files |
|
826
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("bold cyan", "\t\tcreating $name/$abbrev - BUCKET #$bucket $type ARFF...\n"); |
|
827
|
0
|
0
|
|
|
|
|
if($program_dir ne ""){ |
|
828
|
0
|
|
|
|
|
|
my $subdir = "_ARFF"; |
|
829
|
0
|
|
|
|
|
|
my $arffdir = $name . "_ARFF"; |
|
830
|
0
|
|
|
|
|
|
my $featdir = "_$abbrev"; |
|
831
|
0
|
|
|
|
|
|
make_path("$program_dir/$subdir/$arffdir/$featdir/$typeDir"); |
|
832
|
0
|
0
|
|
|
|
|
open($ARFF, ">", ("$program_dir/$subdir/$arffdir/$featdir/$typeDir/" . $name . "_$type-" . $bucket .".arff")) || die ("Cannot write ARFF file to the directory! Check permissions!\n$!"); |
|
833
|
|
|
|
|
|
|
}else{ |
|
834
|
0
|
|
|
|
|
|
my $arffdir = $name . "_ARFF"; |
|
835
|
0
|
|
|
|
|
|
my $featdir = "_$abbrev"; |
|
836
|
0
|
|
|
|
|
|
make_path("$arffdir/$featdir/$typeDir"); |
|
837
|
0
|
0
|
|
|
|
|
open($ARFF, ">", ("$arffdir/$featdir/$typeDir/" . $name . "_$type-" . $bucket .".arff")) || die ("Cannot write ARFF file to the directory! Check permissions!\n$!"); |
|
838
|
|
|
|
|
|
|
} |
|
839
|
|
|
|
|
|
|
|
|
840
|
|
|
|
|
|
|
#get the attr and vector set |
|
841
|
0
|
|
|
|
|
|
my @attrARFFSet = @$attr_ref; |
|
842
|
0
|
|
|
|
|
|
my @vecSet = @$vec_ref; |
|
843
|
|
|
|
|
|
|
|
|
844
|
|
|
|
|
|
|
#get format for the file |
|
845
|
0
|
|
|
|
|
|
my $relation = "\@RELATION $name"; |
|
846
|
0
|
|
|
|
|
|
my @printAttr = makeAttrData(\@attrARFFSet); |
|
847
|
0
|
|
|
|
|
|
my $entity = "\@ATTRIBUTE Entity {No, Yes}"; #set if the entity word or not |
|
848
|
0
|
|
|
|
|
|
my $data = "\@DATA"; |
|
849
|
|
|
|
|
|
|
|
|
850
|
|
|
|
|
|
|
#print everything to the file |
|
851
|
0
|
|
|
|
|
|
$uniSub->printDebug("\t\tprinting to file...\n"); |
|
852
|
0
|
|
|
|
|
|
$uniSub->print2File($ARFF, $relation); |
|
853
|
0
|
|
|
|
|
|
foreach my $a(@printAttr){$uniSub->print2File($ARFF, $a);} |
|
|
0
|
|
|
|
|
|
|
|
854
|
0
|
|
|
|
|
|
$uniSub->print2File($ARFF, $entity); |
|
855
|
0
|
|
|
|
|
|
$uniSub->print2File($ARFF, $data); |
|
856
|
0
|
|
|
|
|
|
foreach my $d(@vecSet){$uniSub->print2File($ARFF, $d);} |
|
|
0
|
|
|
|
|
|
|
|
857
|
0
|
|
|
|
|
|
close $ARFF; |
|
858
|
|
|
|
|
|
|
} |
|
859
|
|
|
|
|
|
|
|
|
860
|
|
|
|
|
|
|
###################### VECTOR THINGIES ##################### |
|
861
|
|
|
|
|
|
|
|
|
862
|
|
|
|
|
|
|
|
|
863
|
|
|
|
|
|
|
#makes vectors from a set |
|
864
|
|
|
|
|
|
|
# input : @txtLineSet <-- the retagged text lines to make vectors out of |
|
865
|
|
|
|
|
|
|
# : @featureList <-- the list of features to make the vectors out of [e.g. (ortho, morph, text)] |
|
866
|
|
|
|
|
|
|
# : @attrs <-- the attributes to use to make the vectors |
|
867
|
|
|
|
|
|
|
# output : @setVectors <-- the vectors for each word in all of the lines |
|
868
|
|
|
|
|
|
|
sub vectorMaker{ |
|
869
|
0
|
|
|
0
|
0
|
|
my $set_ref = shift; |
|
870
|
0
|
|
|
|
|
|
my $feat_ref = shift; |
|
871
|
0
|
|
|
|
|
|
my $attrib_ref = shift; |
|
872
|
0
|
|
|
|
|
|
my @txtLineSet = @$set_ref; |
|
873
|
0
|
|
|
|
|
|
my @featureList = @$feat_ref; |
|
874
|
0
|
|
|
|
|
|
my %attrs = %$attrib_ref; |
|
875
|
|
|
|
|
|
|
|
|
876
|
0
|
|
|
|
|
|
my @setVectors = (); |
|
877
|
|
|
|
|
|
|
#go through each line of the set |
|
878
|
0
|
|
|
|
|
|
my $setLen = @txtLineSet; |
|
879
|
|
|
|
|
|
|
|
|
880
|
0
|
|
|
|
|
|
for(my $l = 0; $l < $setLen; $l++){ |
|
881
|
0
|
|
|
|
|
|
my $line = $txtLineSet[$l]; |
|
882
|
|
|
|
|
|
|
#$uniSub->printColorDebug("on_red", $line); |
|
883
|
0
|
|
|
|
|
|
my @words = split(' ', $line); |
|
884
|
|
|
|
|
|
|
#$uniSub->printArr(", ", \@words); |
|
885
|
|
|
|
|
|
|
#print "\n"; |
|
886
|
0
|
|
|
|
|
|
my $wordLen = @words; |
|
887
|
|
|
|
|
|
|
|
|
888
|
|
|
|
|
|
|
#go through each word |
|
889
|
0
|
|
|
|
|
|
for(my $a = 0; $a < $wordLen; $a++){ |
|
890
|
|
|
|
|
|
|
|
|
891
|
0
|
|
|
|
|
|
$| = 1; |
|
892
|
|
|
|
|
|
|
|
|
893
|
0
|
|
|
|
|
|
my $wordOrig = $words[$a]; |
|
894
|
|
|
|
|
|
|
#make the words for comparison |
|
895
|
0
|
|
|
|
|
|
my $word = $words[$a]; |
|
896
|
0
|
|
|
|
|
|
my $prevWord = ""; |
|
897
|
0
|
|
|
|
|
|
my $nextWord = ""; |
|
898
|
|
|
|
|
|
|
|
|
899
|
|
|
|
|
|
|
#show progress |
|
900
|
0
|
|
|
|
|
|
my $l2 = $l + 1; |
|
901
|
0
|
|
|
|
|
|
my $a2 = $a + 1; |
|
902
|
0
|
|
|
|
|
|
$uniSub->printDebug("\r" . "\t\tLine - $l2/$setLen ------ Word - $a2/$wordLen ---- "); |
|
903
|
|
|
|
|
|
|
|
|
904
|
|
|
|
|
|
|
#cut-off a word in the display if it is too long |
|
905
|
|
|
|
|
|
|
#(longer than 8 characters) |
|
906
|
0
|
|
|
|
|
|
my $smlword = substr($word, 0, 8); |
|
907
|
0
|
0
|
|
|
|
|
if(length($word) > 8){ |
|
908
|
0
|
|
|
|
|
|
$smlword .= "..."; |
|
909
|
|
|
|
|
|
|
} |
|
910
|
|
|
|
|
|
|
|
|
911
|
|
|
|
|
|
|
#distinguish entity words from normal words |
|
912
|
0
|
0
|
|
|
|
|
if($word =~/$entId/o){ |
|
913
|
0
|
|
|
|
|
|
$uniSub->printColorDebug("red", "$smlword! "); |
|
914
|
|
|
|
|
|
|
}else{ |
|
915
|
0
|
|
|
|
|
|
$uniSub->printDebug("$smlword! ") |
|
916
|
|
|
|
|
|
|
} |
|
917
|
|
|
|
|
|
|
|
|
918
|
0
|
|
|
|
|
|
my @word_cuis = getFeature($word, "cui"); |
|
919
|
0
|
|
|
|
|
|
my $ncui = $word_cuis[0]; |
|
920
|
|
|
|
|
|
|
#$uniSub->printColorDebug("red", "\n\t\t$word - $ncui\n"); |
|
921
|
|
|
|
|
|
|
|
|
922
|
|
|
|
|
|
|
#check if it's a stopword |
|
923
|
0
|
0
|
0
|
|
|
|
if(($stopwords_file and $word=~/$stopRegex/o) || ($is_cui and $word_cuis[0] eq "")){ |
|
|
|
|
0
|
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
924
|
|
|
|
|
|
|
#$uniSub->printColorDebug("on_red", "\t$word\tSKIP!"); |
|
925
|
0
|
0
|
|
|
|
|
if(!($word =~/[a-zA-Z0-9\_\-]+$entId/)){ |
|
926
|
0
|
|
|
|
|
|
next; |
|
927
|
|
|
|
|
|
|
} |
|
928
|
|
|
|
|
|
|
} |
|
929
|
|
|
|
|
|
|
|
|
930
|
|
|
|
|
|
|
#if a weird character word - skip |
|
931
|
0
|
0
|
0
|
|
|
|
if((length($word) eq 1) and ($word =~/[^a-zA-Z0-9\_\-]/)){ |
|
932
|
0
|
|
|
|
|
|
next; |
|
933
|
|
|
|
|
|
|
} |
|
934
|
|
|
|
|
|
|
|
|
935
|
|
|
|
|
|
|
#get the word before and after the current word |
|
936
|
0
|
0
|
|
|
|
|
if($a > 0){$prevWord = $words[$a - 1];} |
|
|
0
|
|
|
|
|
|
|
|
937
|
0
|
0
|
|
|
|
|
if($a < ($wordLen - 1)){$nextWord = $words[$a + 1];} |
|
|
0
|
|
|
|
|
|
|
|
938
|
|
|
|
|
|
|
|
|
939
|
|
|
|
|
|
|
|
|
940
|
|
|
|
|
|
|
|
|
941
|
|
|
|
|
|
|
#get rid of tag if necessary |
|
942
|
0
|
|
|
|
|
|
$prevWord =~s/$entId//og; |
|
943
|
0
|
|
|
|
|
|
$nextWord =~s/$entId//og; |
|
944
|
0
|
|
|
|
|
|
$word =~s/$entId//og; |
|
945
|
|
|
|
|
|
|
|
|
946
|
0
|
|
|
|
|
|
my $vec = ""; |
|
947
|
|
|
|
|
|
|
#use each set of attributes |
|
948
|
0
|
|
|
|
|
|
foreach my $item(@featureList){ |
|
949
|
0
|
|
|
|
|
|
my $addVec = ""; |
|
950
|
0
|
0
|
|
|
|
|
if($item eq "ortho"){$addVec = orthoVec($word);} |
|
|
0
|
0
|
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
|
951
|
0
|
|
|
|
|
|
elsif($item eq "morph"){$addVec = morphVec($word, \@{$attrs{"morph"}});} |
|
|
0
|
|
|
|
|
|
|
|
952
|
0
|
|
|
|
|
|
elsif($item eq "text"){$addVec = textVec($word, $prevWord, $nextWord, \@{$attrs{"text"}});} |
|
|
0
|
|
|
|
|
|
|
|
953
|
0
|
|
|
|
|
|
elsif($item eq "pos"){$addVec = posVec($word, $prevWord, $nextWord, \@{$attrs{"pos"}});} |
|
|
0
|
|
|
|
|
|
|
|
954
|
0
|
|
|
|
|
|
elsif($item eq "cui"){$addVec = cuiVec($word, $prevWord, $nextWord, \@{$attrs{"cui"}});} |
|
|
0
|
|
|
|
|
|
|
|
955
|
0
|
|
|
|
|
|
elsif($item eq "sem"){$addVec = semVec($word, $prevWord, $nextWord, \@{$attrs{"sem"}});} |
|
|
0
|
|
|
|
|
|
|
|
956
|
|
|
|
|
|
|
|
|
957
|
|
|
|
|
|
|
|
|
958
|
0
|
|
|
|
|
|
$vec .= $addVec; |
|
959
|
|
|
|
|
|
|
|
|
960
|
|
|
|
|
|
|
} |
|
961
|
|
|
|
|
|
|
|
|
962
|
|
|
|
|
|
|
#convert binary to sparse if specified |
|
963
|
0
|
0
|
|
|
|
|
if($sparse_matrix){ |
|
964
|
0
|
|
|
|
|
|
$vec = convert2Sparse($vec); |
|
965
|
|
|
|
|
|
|
#$uniSub->printColorDebug("red", "$vec\n"); |
|
966
|
|
|
|
|
|
|
} |
|
967
|
|
|
|
|
|
|
|
|
968
|
|
|
|
|
|
|
#check if the word is an entity or not |
|
969
|
|
|
|
|
|
|
#$uniSub->printColorDebug("red", "\n$wordOrig\n"); |
|
970
|
0
|
0
|
|
|
|
|
$vec .= (($wordOrig =~/\b[\S]+($entId)\b/) ? "Yes " : "No "); |
|
971
|
|
|
|
|
|
|
|
|
972
|
|
|
|
|
|
|
#close it if using sparse matrix |
|
973
|
0
|
0
|
|
|
|
|
if($sparse_matrix){ |
|
974
|
0
|
|
|
|
|
|
$vec .= "}"; |
|
975
|
|
|
|
|
|
|
} |
|
976
|
|
|
|
|
|
|
|
|
977
|
|
|
|
|
|
|
#finally add the word back and add the entire vector to the set |
|
978
|
0
|
|
|
|
|
|
$vec .= "\%$word"; |
|
979
|
0
|
0
|
|
|
|
|
if($word ne ""){ |
|
980
|
0
|
|
|
|
|
|
push(@setVectors, $vec); |
|
981
|
|
|
|
|
|
|
} |
|
982
|
|
|
|
|
|
|
} |
|
983
|
|
|
|
|
|
|
} |
|
984
|
|
|
|
|
|
|
|
|
985
|
0
|
|
|
|
|
|
return @setVectors; |
|
986
|
|
|
|
|
|
|
} |
|
987
|
|
|
|
|
|
|
|
|
988
|
|
|
|
|
|
|
#makes the orthographic based part of the vector |
|
989
|
|
|
|
|
|
|
# input : $word <-- the word to analyze |
|
990
|
|
|
|
|
|
|
# output : $strVec <-- the orthographic vector string |
|
991
|
|
|
|
|
|
|
sub orthoVec{ |
|
992
|
0
|
|
|
0
|
0
|
|
my $word = shift; |
|
993
|
|
|
|
|
|
|
|
|
994
|
|
|
|
|
|
|
## CHECKS ## |
|
995
|
0
|
|
|
|
|
|
my $strVec = ""; |
|
996
|
0
|
|
|
|
|
|
my $addon = ""; |
|
997
|
|
|
|
|
|
|
|
|
998
|
|
|
|
|
|
|
#check if first letter capital |
|
999
|
0
|
0
|
|
|
|
|
$addon = ($word =~ /\b([A-Z])\w+\b/o ? 1 : 0); |
|
1000
|
0
|
|
|
|
|
|
$strVec .= "$addon, "; |
|
1001
|
|
|
|
|
|
|
|
|
1002
|
|
|
|
|
|
|
#check if a single letter word |
|
1003
|
0
|
0
|
|
|
|
|
$addon = (length($word) == 1 ? 1 : 0); |
|
1004
|
0
|
|
|
|
|
|
$strVec .= "$addon, "; |
|
1005
|
|
|
|
|
|
|
|
|
1006
|
|
|
|
|
|
|
#check if all capital letters |
|
1007
|
0
|
0
|
|
|
|
|
$addon = ($word =~ /\b[A-Z]+\b/o ? 1 : 0); |
|
1008
|
0
|
|
|
|
|
|
$strVec .= "$addon, "; |
|
1009
|
|
|
|
|
|
|
|
|
1010
|
|
|
|
|
|
|
#check if contains a digit |
|
1011
|
0
|
0
|
|
|
|
|
$addon = ($word =~ /[0-9]+/o ? 1 : 0); |
|
1012
|
0
|
|
|
|
|
|
$strVec .= "$addon, "; |
|
1013
|
|
|
|
|
|
|
|
|
1014
|
|
|
|
|
|
|
#check if all digits |
|
1015
|
0
|
0
|
|
|
|
|
$addon = ($word =~ /\b[0-9]+\b/o ? 1 : 0); |
|
1016
|
0
|
|
|
|
|
|
$strVec .= "$addon, "; |
|
1017
|
|
|
|
|
|
|
|
|
1018
|
|
|
|
|
|
|
#check if contains a hyphen |
|
1019
|
0
|
0
|
|
|
|
|
$addon = ($word =~ /-/o ? 1 : 0); |
|
1020
|
0
|
|
|
|
|
|
$strVec .= "$addon, "; |
|
1021
|
|
|
|
|
|
|
|
|
1022
|
|
|
|
|
|
|
#check if contains punctuation |
|
1023
|
0
|
0
|
|
|
|
|
$addon = ($word =~ /[^a-zA-Z0-9\s]/o ? 1 : 0); |
|
1024
|
0
|
|
|
|
|
|
$strVec .= "$addon, "; |
|
1025
|
|
|
|
|
|
|
|
|
1026
|
0
|
|
|
|
|
|
return $strVec; |
|
1027
|
|
|
|
|
|
|
} |
|
1028
|
|
|
|
|
|
|
|
|
1029
|
|
|
|
|
|
|
#makes the morphological based part of the vector |
|
1030
|
|
|
|
|
|
|
# input : $word <-- the word to analyze |
|
1031
|
|
|
|
|
|
|
# : @attrs <-- the set of morphological attributes to use |
|
1032
|
|
|
|
|
|
|
# output : $strVec <-- the morphological vector string |
|
1033
|
|
|
|
|
|
|
sub morphVec{ |
|
1034
|
0
|
|
|
0
|
0
|
|
my $word = shift; |
|
1035
|
0
|
|
|
|
|
|
my $attrs_ref = shift; |
|
1036
|
0
|
|
|
|
|
|
my @attrs = @$attrs_ref; |
|
1037
|
|
|
|
|
|
|
|
|
1038
|
0
|
|
|
|
|
|
my $strVec = ""; |
|
1039
|
|
|
|
|
|
|
|
|
1040
|
|
|
|
|
|
|
#grab the first # characters and the last # characters |
|
1041
|
0
|
|
|
|
|
|
my $preWord = substr($word, 0, $prefix); |
|
1042
|
0
|
|
|
|
|
|
my $sufWord = substr($word, -$suffix); |
|
1043
|
|
|
|
|
|
|
|
|
1044
|
|
|
|
|
|
|
#compare and build a binary vector |
|
1045
|
0
|
|
|
|
|
|
foreach my $a (@attrs){ |
|
1046
|
0
|
0
|
|
|
|
|
if($a eq $preWord){ |
|
|
|
0
|
|
|
|
|
|
|
1047
|
0
|
|
|
|
|
|
$strVec .= "1, "; |
|
1048
|
|
|
|
|
|
|
}elsif($a eq $sufWord){ |
|
1049
|
0
|
|
|
|
|
|
$strVec .= "1, "; |
|
1050
|
|
|
|
|
|
|
}else{ |
|
1051
|
0
|
|
|
|
|
|
$strVec .= "0, "; |
|
1052
|
|
|
|
|
|
|
} |
|
1053
|
|
|
|
|
|
|
} |
|
1054
|
|
|
|
|
|
|
|
|
1055
|
0
|
|
|
|
|
|
return $strVec; |
|
1056
|
|
|
|
|
|
|
|
|
1057
|
|
|
|
|
|
|
} |
|
1058
|
|
|
|
|
|
|
|
|
1059
|
|
|
|
|
|
|
#makes the text based part of the vector |
|
1060
|
|
|
|
|
|
|
# input : $w <-- the word to analyze |
|
1061
|
|
|
|
|
|
|
# : $pw <-- the previous word |
|
1062
|
|
|
|
|
|
|
# : $nw <-- the next word |
|
1063
|
|
|
|
|
|
|
# : @attrbts <-- the set of text attributes to use |
|
1064
|
|
|
|
|
|
|
# output : $strVec <-- the text vector string |
|
1065
|
|
|
|
|
|
|
sub textVec{ |
|
1066
|
0
|
|
|
0
|
0
|
|
my $w = shift; |
|
1067
|
0
|
|
|
|
|
|
my $pw = shift; |
|
1068
|
0
|
|
|
|
|
|
my $nw = shift; |
|
1069
|
0
|
|
|
|
|
|
my $at_ref = shift; |
|
1070
|
0
|
|
|
|
|
|
my @attrbts = @$at_ref; |
|
1071
|
|
|
|
|
|
|
|
|
1072
|
0
|
|
|
|
|
|
my $strVec = ""; |
|
1073
|
|
|
|
|
|
|
|
|
1074
|
|
|
|
|
|
|
#clean the words |
|
1075
|
0
|
|
|
|
|
|
$w = $uniSub->cleanWords($w); |
|
1076
|
0
|
|
|
|
|
|
$pw = $uniSub->cleanWords($pw); |
|
1077
|
0
|
|
|
|
|
|
$nw = $uniSub->cleanWords($nw); |
|
1078
|
|
|
|
|
|
|
|
|
1079
|
|
|
|
|
|
|
#check if the word is the attribute or the words adjacent it are the attribute |
|
1080
|
0
|
|
|
|
|
|
foreach my $a(@attrbts){ |
|
1081
|
|
|
|
|
|
|
|
|
1082
|
0
|
|
|
|
|
|
my $pair = ""; |
|
1083
|
0
|
0
|
|
|
|
|
$pair .= ($w eq $a ? "1, " : "0, "); |
|
1084
|
0
|
0
|
0
|
|
|
|
$pair .= (($pw eq $a or $nw eq $a) ? "1, " : "0, "); |
|
1085
|
0
|
|
|
|
|
|
$strVec .= $pair; |
|
1086
|
|
|
|
|
|
|
} |
|
1087
|
|
|
|
|
|
|
|
|
1088
|
0
|
|
|
|
|
|
return $strVec; |
|
1089
|
|
|
|
|
|
|
} |
|
1090
|
|
|
|
|
|
|
|
|
1091
|
|
|
|
|
|
|
#makes the part of speech based part of the vector |
|
1092
|
|
|
|
|
|
|
# input : $w <-- the word to analyze |
|
1093
|
|
|
|
|
|
|
# : $pw <-- the previous word |
|
1094
|
|
|
|
|
|
|
# : $nw <-- the next word |
|
1095
|
|
|
|
|
|
|
# : @attrbts <-- the set of pos attributes to use |
|
1096
|
|
|
|
|
|
|
# output : $strVec <-- the pos vector string |
|
1097
|
|
|
|
|
|
|
sub posVec{ |
|
1098
|
0
|
|
|
0
|
0
|
|
my $w = shift; |
|
1099
|
0
|
|
|
|
|
|
my $pw = shift; |
|
1100
|
0
|
|
|
|
|
|
my $nw = shift; |
|
1101
|
0
|
|
|
|
|
|
my $at_ref = shift; |
|
1102
|
0
|
|
|
|
|
|
my @attrbts = @$at_ref; |
|
1103
|
|
|
|
|
|
|
|
|
1104
|
|
|
|
|
|
|
#clean the words |
|
1105
|
0
|
|
|
|
|
|
$w = $uniSub->cleanWords($w); |
|
1106
|
0
|
|
|
|
|
|
$pw = $uniSub->cleanWords($pw); |
|
1107
|
0
|
|
|
|
|
|
$nw = $uniSub->cleanWords($nw); |
|
1108
|
|
|
|
|
|
|
|
|
1109
|
|
|
|
|
|
|
#alter the words to make them pos types |
|
1110
|
0
|
|
|
|
|
|
$w = getFeature($w, "pos"); |
|
1111
|
0
|
|
|
|
|
|
$pw = getFeature($pw, "pos"); |
|
1112
|
0
|
|
|
|
|
|
$nw = getFeature($nw, "pos"); |
|
1113
|
|
|
|
|
|
|
|
|
1114
|
0
|
|
|
|
|
|
my $strVec = ""; |
|
1115
|
|
|
|
|
|
|
|
|
1116
|
|
|
|
|
|
|
#check if the word is the attribute or the words adjacent it are the attribute |
|
1117
|
0
|
|
|
|
|
|
foreach my $a(@attrbts){ |
|
1118
|
0
|
|
|
|
|
|
my $pair = ""; |
|
1119
|
0
|
0
|
|
|
|
|
$pair .= ($w eq $a ? "1, " : "0, "); |
|
1120
|
0
|
0
|
0
|
|
|
|
$pair .= (($pw eq $a or $nw eq $a) ? "1, " : "0, "); |
|
1121
|
0
|
|
|
|
|
|
$strVec .= $pair; |
|
1122
|
|
|
|
|
|
|
} |
|
1123
|
|
|
|
|
|
|
|
|
1124
|
0
|
|
|
|
|
|
return $strVec; |
|
1125
|
|
|
|
|
|
|
} |
|
1126
|
|
|
|
|
|
|
|
|
1127
|
|
|
|
|
|
|
#makes the cui based part of the vector |
|
1128
|
|
|
|
|
|
|
# input : $w <-- the word to analyze |
|
1129
|
|
|
|
|
|
|
# : $pw <-- the previous word |
|
1130
|
|
|
|
|
|
|
# : $nw <-- the next word |
|
1131
|
|
|
|
|
|
|
# : @attrbts <-- the set of cui attributes to use |
|
1132
|
|
|
|
|
|
|
# output : $strVec <-- the cui vector string |
|
1133
|
|
|
|
|
|
|
sub cuiVec{ |
|
1134
|
0
|
|
|
0
|
0
|
|
my $w = shift; |
|
1135
|
0
|
|
|
|
|
|
my $pw = shift; |
|
1136
|
0
|
|
|
|
|
|
my $nw = shift; |
|
1137
|
0
|
|
|
|
|
|
my $at_ref = shift; |
|
1138
|
0
|
|
|
|
|
|
my @attrbts = @$at_ref; |
|
1139
|
|
|
|
|
|
|
|
|
1140
|
|
|
|
|
|
|
#clean the words |
|
1141
|
0
|
|
|
|
|
|
$w = $uniSub->cleanWords($w); |
|
1142
|
0
|
|
|
|
|
|
$pw = $uniSub->cleanWords($pw); |
|
1143
|
0
|
|
|
|
|
|
$nw = $uniSub->cleanWords($nw); |
|
1144
|
|
|
|
|
|
|
|
|
1145
|
|
|
|
|
|
|
#alter the words to make them cui types |
|
1146
|
0
|
|
|
|
|
|
my @wArr = getFeature($w, "cui"); |
|
1147
|
0
|
|
|
|
|
|
my @pwArr = getFeature($pw, "cui"); |
|
1148
|
0
|
|
|
|
|
|
my @nwArr = getFeature($nw, "cui"); |
|
1149
|
|
|
|
|
|
|
|
|
1150
|
0
|
|
|
|
|
|
my $strVec = ""; |
|
1151
|
|
|
|
|
|
|
#check if the word is the attribute or the words adjacent it are the attribute |
|
1152
|
0
|
|
|
|
|
|
foreach my $a(@attrbts){ |
|
1153
|
0
|
|
|
|
|
|
my $pair = ""; |
|
1154
|
0
|
0
|
|
|
|
|
$pair .= ($uniSub->inArr($a, \@wArr) ? "1, " : "0, "); |
|
1155
|
0
|
0
|
0
|
|
|
|
$pair .= (($uniSub->inArr($a, \@pwArr) or $uniSub->inArr($a, \@nwArr)) ? "1, " : "0, "); |
|
1156
|
0
|
|
|
|
|
|
$strVec .= $pair; |
|
1157
|
|
|
|
|
|
|
} |
|
1158
|
|
|
|
|
|
|
|
|
1159
|
0
|
|
|
|
|
|
return $strVec; |
|
1160
|
|
|
|
|
|
|
} |
|
1161
|
|
|
|
|
|
|
|
|
1162
|
|
|
|
|
|
|
#makes the semantic based part of the vector |
|
1163
|
|
|
|
|
|
|
# input : $w <-- the word to analyze |
|
1164
|
|
|
|
|
|
|
# : $pw <-- the previous word |
|
1165
|
|
|
|
|
|
|
# : $nw <-- the next word |
|
1166
|
|
|
|
|
|
|
# : @attrbts <-- the set of sem attributes to use |
|
1167
|
|
|
|
|
|
|
# output : $strVec <-- the sem vector string |
|
1168
|
|
|
|
|
|
|
sub semVec{ |
|
1169
|
0
|
|
|
0
|
0
|
|
my $w = shift; |
|
1170
|
0
|
|
|
|
|
|
my $pw = shift; |
|
1171
|
0
|
|
|
|
|
|
my $nw = shift; |
|
1172
|
0
|
|
|
|
|
|
my $at_ref = shift; |
|
1173
|
0
|
|
|
|
|
|
my @attrbts = @$at_ref; |
|
1174
|
|
|
|
|
|
|
|
|
1175
|
|
|
|
|
|
|
#clean the words |
|
1176
|
0
|
|
|
|
|
|
$w = $uniSub->cleanWords($w); |
|
1177
|
0
|
|
|
|
|
|
$pw = $uniSub->cleanWords($pw); |
|
1178
|
0
|
|
|
|
|
|
$nw = $uniSub->cleanWords($nw); |
|
1179
|
|
|
|
|
|
|
|
|
1180
|
|
|
|
|
|
|
#alter the words to make them sem types |
|
1181
|
0
|
|
|
|
|
|
my @wArr = getFeature($w, "sem"); |
|
1182
|
0
|
|
|
|
|
|
my @pwArr = getFeature($pw, "sem"); |
|
1183
|
0
|
|
|
|
|
|
my @nwArr = getFeature($nw, "sem"); |
|
1184
|
|
|
|
|
|
|
|
|
1185
|
0
|
|
|
|
|
|
my $strVec = ""; |
|
1186
|
|
|
|
|
|
|
|
|
1187
|
|
|
|
|
|
|
#check if the word is the attribute or the words adjacent it are the attribute |
|
1188
|
0
|
|
|
|
|
|
foreach my $a(@attrbts){ |
|
1189
|
|
|
|
|
|
|
#remove "sem" label |
|
1190
|
0
|
|
|
|
|
|
$a = lc($a); |
|
1191
|
|
|
|
|
|
|
|
|
1192
|
0
|
|
|
|
|
|
my $pair = ""; |
|
1193
|
0
|
0
|
|
|
|
|
$pair .= ($uniSub->inArr($a, \@wArr) ? "1, " : "0, "); |
|
1194
|
0
|
0
|
0
|
|
|
|
$pair .= (($uniSub->inArr($a, \@pwArr) or $uniSub->inArr($a, \@nwArr)) ? "1, " : "0, "); |
|
1195
|
0
|
|
|
|
|
|
$strVec .= $pair; |
|
1196
|
|
|
|
|
|
|
} |
|
1197
|
0
|
|
|
|
|
|
return $strVec; |
|
1198
|
|
|
|
|
|
|
} |
|
1199
|
|
|
|
|
|
|
|
|
1200
|
|
|
|
|
|
|
#converts a binary vector to a sparse vector |
|
1201
|
|
|
|
|
|
|
sub convert2Sparse{ |
|
1202
|
0
|
|
|
0
|
0
|
|
my $bin_vec = shift; |
|
1203
|
0
|
|
|
|
|
|
my @vals = split(", ", $bin_vec); |
|
1204
|
0
|
|
|
|
|
|
my $numVals = @vals; |
|
1205
|
|
|
|
|
|
|
|
|
1206
|
0
|
|
|
|
|
|
my $sparse_vec = "{"; |
|
1207
|
0
|
|
|
|
|
|
for(my $c=0;$c<$numVals;$c++){ |
|
1208
|
0
|
|
|
|
|
|
my $curVal = $vals[$c]; |
|
1209
|
|
|
|
|
|
|
|
|
1210
|
|
|
|
|
|
|
#if a non-zero value is found at the index - add it to the final |
|
1211
|
0
|
0
|
|
|
|
|
if(($curVal eq "1")){ |
|
1212
|
0
|
|
|
|
|
|
$sparse_vec .= "$c $curVal, "; |
|
1213
|
|
|
|
|
|
|
#$uniSub->printColorDebug("red", "$c $curVal, "); |
|
1214
|
|
|
|
|
|
|
} |
|
1215
|
|
|
|
|
|
|
} |
|
1216
|
0
|
|
|
|
|
|
$sparse_vec .= "$numVals, "; |
|
1217
|
|
|
|
|
|
|
|
|
1218
|
0
|
|
|
|
|
|
return $sparse_vec; |
|
1219
|
|
|
|
|
|
|
} |
|
1220
|
|
|
|
|
|
|
|
|
1221
|
|
|
|
|
|
|
|
|
1222
|
|
|
|
|
|
|
###################### ATTRIBUTE BASED METHODS ##################### |
|
1223
|
|
|
|
|
|
|
|
|
1224
|
|
|
|
|
|
|
#gets the attributes based on the item |
|
1225
|
|
|
|
|
|
|
# input : $feature <-- the feature type [e.g. ortho, morph, text] |
|
1226
|
|
|
|
|
|
|
# : %buckets <-- the bucket key set |
|
1227
|
|
|
|
|
|
|
# output : %vecARFFattr <-- the vector set of attributes and arff set of attributes |
|
1228
|
|
|
|
|
|
|
sub grabAttr{ |
|
1229
|
0
|
|
|
0
|
0
|
|
my $name = shift; |
|
1230
|
0
|
|
|
|
|
|
my $feature = shift; |
|
1231
|
0
|
|
|
|
|
|
my $buckets_ref = shift; |
|
1232
|
0
|
|
|
|
|
|
my %buckets = %$buckets_ref; |
|
1233
|
|
|
|
|
|
|
|
|
1234
|
0
|
|
|
|
|
|
my %vecARFFattr = (); |
|
1235
|
|
|
|
|
|
|
#no importing of attributes is needed for ortho since they all check the same features |
|
1236
|
0
|
0
|
|
|
|
|
if($feature eq "ortho"){ |
|
|
|
0
|
|
|
|
|
|
|
1237
|
0
|
|
|
|
|
|
my @vecSet = (); |
|
1238
|
0
|
|
|
|
|
|
my @arffSet = ("first_letter_capital", #output for the ARFF file |
|
1239
|
|
|
|
|
|
|
"single_character", |
|
1240
|
|
|
|
|
|
|
"all_capital", |
|
1241
|
|
|
|
|
|
|
"has_digit", |
|
1242
|
|
|
|
|
|
|
"all_digit", |
|
1243
|
|
|
|
|
|
|
"has_hyphen", |
|
1244
|
|
|
|
|
|
|
"has_punctuation"); |
|
1245
|
0
|
|
|
|
|
|
$vecARFFattr{vector} = \@vecSet; |
|
1246
|
0
|
|
|
|
|
|
$vecARFFattr{arff} = \@arffSet; |
|
1247
|
0
|
|
|
|
|
|
return %vecARFFattr; |
|
1248
|
|
|
|
|
|
|
} |
|
1249
|
|
|
|
|
|
|
#the morphological attributes look at the prefix and suffix of a word not the adjacencies |
|
1250
|
|
|
|
|
|
|
elsif($feature eq "morph"){ |
|
1251
|
0
|
|
|
|
|
|
my %bucketAttr = (); |
|
1252
|
0
|
|
|
|
|
|
my %bucketAttrARFF = (); |
|
1253
|
|
|
|
|
|
|
|
|
1254
|
|
|
|
|
|
|
#get the attributes for each bucket |
|
1255
|
0
|
|
|
|
|
|
foreach my $testBucket (@allBuckets){ |
|
1256
|
0
|
|
|
|
|
|
my @range = (); |
|
1257
|
0
|
0
|
|
|
|
|
if($bucketsNum > 1){ |
|
1258
|
0
|
|
|
|
|
|
@range = $uniSub->bully($bucketsNum, $testBucket); |
|
1259
|
|
|
|
|
|
|
}else{ |
|
1260
|
0
|
|
|
|
|
|
@range = (1); |
|
1261
|
|
|
|
|
|
|
} |
|
1262
|
0
|
|
|
|
|
|
$uniSub->printDebug("\t\t$name BUCKET #$testBucket/$feature MORPHO attributes...\n"); |
|
1263
|
|
|
|
|
|
|
|
|
1264
|
|
|
|
|
|
|
#get attributes [ unique and deluxe ] |
|
1265
|
0
|
|
|
|
|
|
my @attr = getMorphoAttributes(\@range, \%buckets); |
|
1266
|
0
|
|
|
|
|
|
@attr = uniq(@attr); #make unique forms |
|
1267
|
0
|
|
|
|
|
|
$bucketAttr{$testBucket} = \@attr; |
|
1268
|
|
|
|
|
|
|
|
|
1269
|
0
|
|
|
|
|
|
my @attrARFF = @attr; |
|
1270
|
0
|
|
|
|
|
|
foreach my $a(@attrARFF){$a .= $morphID;} |
|
|
0
|
|
|
|
|
|
|
|
1271
|
0
|
|
|
|
|
|
$bucketAttrARFF{$testBucket} = \@attrARFF; |
|
1272
|
|
|
|
|
|
|
} |
|
1273
|
|
|
|
|
|
|
|
|
1274
|
|
|
|
|
|
|
#add to overall |
|
1275
|
0
|
|
|
|
|
|
$vecARFFattr{vector} = \%bucketAttr; |
|
1276
|
0
|
|
|
|
|
|
$vecARFFattr{arff} = \%bucketAttrARFF; |
|
1277
|
|
|
|
|
|
|
|
|
1278
|
0
|
|
|
|
|
|
return %vecARFFattr; |
|
1279
|
|
|
|
|
|
|
} |
|
1280
|
|
|
|
|
|
|
#text, part-of-speech, semantics, cui attributes |
|
1281
|
|
|
|
|
|
|
else{ |
|
1282
|
0
|
|
|
|
|
|
my %bucketAttr = (); |
|
1283
|
0
|
|
|
|
|
|
my %bucketAttrARFF = (); |
|
1284
|
|
|
|
|
|
|
|
|
1285
|
|
|
|
|
|
|
#get the attributes for each bucket |
|
1286
|
0
|
|
|
|
|
|
foreach my $testBucket (@allBuckets){ |
|
1287
|
0
|
|
|
|
|
|
my @range = (); |
|
1288
|
0
|
0
|
|
|
|
|
if($bucketsNum > 1){ |
|
1289
|
0
|
|
|
|
|
|
@range = $uniSub->bully($bucketsNum, $testBucket); |
|
1290
|
|
|
|
|
|
|
}else{ |
|
1291
|
0
|
|
|
|
|
|
@range = (1); |
|
1292
|
|
|
|
|
|
|
} |
|
1293
|
0
|
|
|
|
|
|
$uniSub->printDebug("\t\t$name BUCKET #$testBucket/$feature attributes...\n"); |
|
1294
|
|
|
|
|
|
|
|
|
1295
|
|
|
|
|
|
|
#get attributes [ unique and deluxe ] |
|
1296
|
0
|
|
|
|
|
|
my @attr = getRangeAttributes($feature, \@range, \%buckets); |
|
1297
|
0
|
|
|
|
|
|
$bucketAttr{$testBucket} = \@attr; |
|
1298
|
|
|
|
|
|
|
|
|
1299
|
0
|
|
|
|
|
|
my @attrARFF = getAttrDelux($feature, \@attr); |
|
1300
|
0
|
|
|
|
|
|
$bucketAttrARFF{$testBucket} = \@attrARFF; |
|
1301
|
|
|
|
|
|
|
} |
|
1302
|
|
|
|
|
|
|
|
|
1303
|
|
|
|
|
|
|
#add to overall |
|
1304
|
0
|
|
|
|
|
|
$vecARFFattr{vector} = \%bucketAttr; |
|
1305
|
0
|
|
|
|
|
|
$vecARFFattr{arff} = \%bucketAttrARFF; |
|
1306
|
|
|
|
|
|
|
|
|
1307
|
0
|
|
|
|
|
|
return %vecARFFattr; |
|
1308
|
|
|
|
|
|
|
} |
|
1309
|
|
|
|
|
|
|
} |
|
1310
|
|
|
|
|
|
|
|
|
1311
|
|
|
|
|
|
|
#returns the attribute values of a range of buckets |
|
1312
|
|
|
|
|
|
|
# input : $type <-- the feature type [e.g. ortho, morph, text] |
|
1313
|
|
|
|
|
|
|
# : @bucketRange <-- the range of the buckets to use [e.g.(1-8,10) out of 10 buckets; use "$uniSub->bully" subroutine in UniversalRoutines.pm] |
|
1314
|
|
|
|
|
|
|
# : %buckets <-- the bucket key set |
|
1315
|
|
|
|
|
|
|
# output : @attributes <-- the set of attributes for the specific type and range |
|
1316
|
|
|
|
|
|
|
sub getRangeAttributes{ |
|
1317
|
0
|
|
|
0
|
0
|
|
my $type = shift; |
|
1318
|
0
|
|
|
|
|
|
my $bucketRange_ref = shift; |
|
1319
|
0
|
|
|
|
|
|
my $buckets_ref = shift; |
|
1320
|
0
|
|
|
|
|
|
my @bucketRange = @$bucketRange_ref; |
|
1321
|
0
|
|
|
|
|
|
my %buckets = %$buckets_ref; |
|
1322
|
|
|
|
|
|
|
|
|
1323
|
|
|
|
|
|
|
#collect all the necessary keys |
|
1324
|
0
|
|
|
|
|
|
my @keyRing = (); |
|
1325
|
0
|
|
|
|
|
|
foreach my $bucket (sort { $a <=> $b } keys %buckets){ |
|
|
0
|
|
|
|
|
|
|
|
1326
|
0
|
0
|
|
|
|
|
if($uniSub->inArr($bucket, \@bucketRange)){ |
|
1327
|
0
|
|
|
|
|
|
my @keys = @{$buckets{$bucket}}; |
|
|
0
|
|
|
|
|
|
|
|
1328
|
0
|
|
|
|
|
|
push @keyRing, @keys; |
|
1329
|
|
|
|
|
|
|
} |
|
1330
|
|
|
|
|
|
|
} |
|
1331
|
|
|
|
|
|
|
|
|
1332
|
|
|
|
|
|
|
#get the lines |
|
1333
|
0
|
|
|
|
|
|
my @bucketLines = (); |
|
1334
|
0
|
|
|
|
|
|
foreach my $key (@keyRing){ |
|
1335
|
0
|
|
|
|
|
|
push(@bucketLines, $fileHash{$key}); |
|
1336
|
|
|
|
|
|
|
} |
|
1337
|
|
|
|
|
|
|
|
|
1338
|
|
|
|
|
|
|
#gather the attributes based on the words in the lines |
|
1339
|
0
|
|
|
|
|
|
my @attributes = (); |
|
1340
|
0
|
|
|
|
|
|
foreach my $line (@bucketLines){ #in each line |
|
1341
|
0
|
|
|
|
|
|
foreach my $o_word (split(' ', $line)){ #each word |
|
1342
|
0
|
|
|
|
|
|
my $word = $o_word; |
|
1343
|
0
|
|
|
|
|
|
$word =~s/$entId//; #remove the annotation marker |
|
1344
|
0
|
|
|
|
|
|
$word =~s///; #remove the tag |
|
1345
|
0
|
|
|
|
|
|
$word =~s///; #remove the tag |
|
1346
|
0
|
|
|
|
|
|
$word =~s/[\:\.\-]//; #remove any : . - characters |
|
1347
|
|
|
|
|
|
|
|
|
1348
|
|
|
|
|
|
|
#if word is empty - skip |
|
1349
|
0
|
0
|
|
|
|
|
if($word eq ""){ |
|
1350
|
0
|
|
|
|
|
|
next; |
|
1351
|
|
|
|
|
|
|
} |
|
1352
|
|
|
|
|
|
|
|
|
1353
|
|
|
|
|
|
|
#if only looking for text attributes - use the actual word |
|
1354
|
0
|
0
|
|
|
|
|
if($type eq "text"){ |
|
1355
|
0
|
|
|
|
|
|
push(@attributes, $word); |
|
1356
|
0
|
|
|
|
|
|
next; |
|
1357
|
|
|
|
|
|
|
} |
|
1358
|
|
|
|
|
|
|
|
|
1359
|
|
|
|
|
|
|
#skip if dne |
|
1360
|
0
|
0
|
|
|
|
|
if(!$featHash{$word}){ |
|
1361
|
0
|
|
|
|
|
|
next; |
|
1362
|
|
|
|
|
|
|
} |
|
1363
|
|
|
|
|
|
|
#grab attributes if it does |
|
1364
|
0
|
|
|
|
|
|
my $hash_ref = $featHash{$word}; |
|
1365
|
0
|
|
|
|
|
|
my %hash = %$hash_ref; |
|
1366
|
|
|
|
|
|
|
|
|
1367
|
0
|
0
|
|
|
|
|
if($hash{$type}){ |
|
1368
|
0
|
|
|
|
|
|
push(@attributes, split(',', $hash{$type})); |
|
1369
|
|
|
|
|
|
|
} |
|
1370
|
|
|
|
|
|
|
} |
|
1371
|
|
|
|
|
|
|
} |
|
1372
|
|
|
|
|
|
|
|
|
1373
|
0
|
|
|
|
|
|
@attributes = uniq(@attributes); |
|
1374
|
|
|
|
|
|
|
#$uniSub->printArr(",", \@attributes); |
|
1375
|
0
|
|
|
|
|
|
return @attributes; |
|
1376
|
|
|
|
|
|
|
|
|
1377
|
|
|
|
|
|
|
} |
|
1378
|
|
|
|
|
|
|
|
|
1379
|
|
|
|
|
|
|
#makes the arff version attributes - makes a copy of each attribute but with "_self" at the end |
|
1380
|
|
|
|
|
|
|
# input : $f <-- the feature type (used for special features like POS and morph) |
|
1381
|
|
|
|
|
|
|
# : @attrs <-- the attributes to ready for arff output |
|
1382
|
|
|
|
|
|
|
# output : @attrDelux <-- the delux-arff attribute set |
|
1383
|
|
|
|
|
|
|
sub getAttrDelux{ |
|
1384
|
0
|
|
|
0
|
0
|
|
my $f = shift; |
|
1385
|
0
|
|
|
|
|
|
my $attr_ref = shift; |
|
1386
|
0
|
|
|
|
|
|
my @attr = @$attr_ref; |
|
1387
|
|
|
|
|
|
|
|
|
1388
|
|
|
|
|
|
|
#add the _self copy |
|
1389
|
0
|
|
|
|
|
|
my @attrDelux = (); |
|
1390
|
0
|
|
|
|
|
|
foreach my $word (@attr){ |
|
1391
|
|
|
|
|
|
|
#check if certain type of feature |
|
1392
|
0
|
0
|
|
|
|
|
if($f eq "pos"){ |
|
|
|
0
|
|
|
|
|
|
|
1393
|
0
|
|
|
|
|
|
$word = ($word . "_POS"); |
|
1394
|
|
|
|
|
|
|
}elsif($f eq "sem"){ |
|
1395
|
0
|
|
|
|
|
|
$word = uc $word; |
|
1396
|
|
|
|
|
|
|
} |
|
1397
|
0
|
|
|
|
|
|
$word =~s/$entId//g; |
|
1398
|
|
|
|
|
|
|
|
|
1399
|
|
|
|
|
|
|
#add the copy and then the original |
|
1400
|
0
|
|
|
|
|
|
my $copy = "$word" . "$selfId"; |
|
1401
|
0
|
0
|
|
|
|
|
if(!$uniSub->inArr($word, \@attrDelux)){ |
|
1402
|
0
|
|
|
|
|
|
push (@attrDelux, $copy); |
|
1403
|
0
|
|
|
|
|
|
push(@attrDelux, $word); |
|
1404
|
|
|
|
|
|
|
} |
|
1405
|
|
|
|
|
|
|
} |
|
1406
|
0
|
|
|
|
|
|
return @attrDelux; |
|
1407
|
|
|
|
|
|
|
} |
|
1408
|
|
|
|
|
|
|
|
|
1409
|
|
|
|
|
|
|
#looks at the prefix # and suffix # and returns a substring of each word found in the bucket text set |
|
1410
|
|
|
|
|
|
|
# input : @bucketRange <-- the range of the buckets to use [e.g.(1-8,10) out of 10 buckets; use "$uniSub->bully" subroutine in UniversalRoutines.pm] |
|
1411
|
|
|
|
|
|
|
# : %buckets <-- the bucket key set |
|
1412
|
|
|
|
|
|
|
# output : @attributes <-- the morphological attribute set |
|
1413
|
|
|
|
|
|
|
sub getMorphoAttributes{ |
|
1414
|
0
|
|
|
0
|
0
|
|
my $bucketRange_ref = shift; |
|
1415
|
0
|
|
|
|
|
|
my $buckets_ref = shift; |
|
1416
|
0
|
|
|
|
|
|
my @bucketRange = @$bucketRange_ref; |
|
1417
|
0
|
|
|
|
|
|
my %buckets = %$buckets_ref; |
|
1418
|
|
|
|
|
|
|
|
|
1419
|
|
|
|
|
|
|
#collect all the necessary keys |
|
1420
|
0
|
|
|
|
|
|
my @keyRing = (); |
|
1421
|
0
|
|
|
|
|
|
foreach my $bucket (sort { $a <=> $b } keys %buckets){ |
|
|
0
|
|
|
|
|
|
|
|
1422
|
0
|
0
|
|
|
|
|
if($uniSub->inArr($bucket, \@bucketRange)){ |
|
1423
|
0
|
|
|
|
|
|
my @keys = @{$buckets{$bucket}}; |
|
|
0
|
|
|
|
|
|
|
|
1424
|
0
|
|
|
|
|
|
push @keyRing, @keys; |
|
1425
|
|
|
|
|
|
|
} |
|
1426
|
|
|
|
|
|
|
} |
|
1427
|
|
|
|
|
|
|
|
|
1428
|
0
|
|
|
|
|
|
my @bucketLines = (); |
|
1429
|
|
|
|
|
|
|
#get the lines for each associated key |
|
1430
|
0
|
|
|
|
|
|
foreach my $key (@keyRing){ |
|
1431
|
0
|
|
|
|
|
|
my $line = $fileHash{$key}; |
|
1432
|
0
|
|
|
|
|
|
push @bucketLines, $line; |
|
1433
|
|
|
|
|
|
|
} |
|
1434
|
|
|
|
|
|
|
|
|
1435
|
|
|
|
|
|
|
#get each word from each line |
|
1436
|
0
|
|
|
|
|
|
my @wordSet = (); |
|
1437
|
0
|
|
|
|
|
|
foreach my $line (@bucketLines){ |
|
1438
|
0
|
|
|
|
|
|
my @words = split(" ", $line); |
|
1439
|
0
|
|
|
|
|
|
push(@wordSet, @words); |
|
1440
|
|
|
|
|
|
|
} |
|
1441
|
|
|
|
|
|
|
|
|
1442
|
|
|
|
|
|
|
#get the prefix and suffix from each word |
|
1443
|
0
|
|
|
|
|
|
my @attributes = (); |
|
1444
|
0
|
|
|
|
|
|
foreach my $word (@wordSet){ |
|
1445
|
0
|
|
|
|
|
|
$word =~s/$entId//g; |
|
1446
|
0
|
|
|
|
|
|
push(@attributes, substr($word, 0, $prefix)); #add the word's prefix |
|
1447
|
0
|
|
|
|
|
|
push(@attributes, substr($word, -$suffix)); #add the word's suffix |
|
1448
|
|
|
|
|
|
|
} |
|
1449
|
|
|
|
|
|
|
|
|
1450
|
|
|
|
|
|
|
#my $a = @attributes; |
|
1451
|
|
|
|
|
|
|
#$uniSub->printColorDebug("red", "$type ATTR: #$a\n"); |
|
1452
|
|
|
|
|
|
|
#printArr("\n", @attributes); |
|
1453
|
|
|
|
|
|
|
|
|
1454
|
0
|
|
|
|
|
|
return @attributes; |
|
1455
|
|
|
|
|
|
|
} |
|
1456
|
|
|
|
|
|
|
|
|
1457
|
|
|
|
|
|
|
#formats attributes for the ARFF file |
|
1458
|
|
|
|
|
|
|
# input : @set <-- the attribute set |
|
1459
|
|
|
|
|
|
|
# output : @attributes <-- the arff formatted attributes |
|
1460
|
|
|
|
|
|
|
sub makeAttrData{ |
|
1461
|
0
|
|
|
0
|
0
|
|
my $set_ref = shift; |
|
1462
|
0
|
|
|
|
|
|
my @set = @$set_ref; |
|
1463
|
|
|
|
|
|
|
|
|
1464
|
0
|
|
|
|
|
|
my @attributes = (); |
|
1465
|
0
|
|
|
|
|
|
foreach my $attr (@set){ |
|
1466
|
0
|
|
|
|
|
|
push (@attributes, "\@ATTRIBUTE $attr NUMERIC"); |
|
1467
|
|
|
|
|
|
|
} |
|
1468
|
|
|
|
|
|
|
|
|
1469
|
0
|
|
|
|
|
|
return @attributes; |
|
1470
|
|
|
|
|
|
|
} |
|
1471
|
|
|
|
|
|
|
|
|
1472
|
|
|
|
|
|
|
##new stoplist function - by Dr. McInnes |
|
1473
|
|
|
|
|
|
|
#generates a regex expression searching for the stop list words |
|
1474
|
|
|
|
|
|
|
sub stop { |
|
1475
|
|
|
|
|
|
|
|
|
1476
|
0
|
|
|
0
|
0
|
|
my $stopfile = shift; |
|
1477
|
|
|
|
|
|
|
|
|
1478
|
0
|
|
|
|
|
|
my $stop_regex = ""; |
|
1479
|
0
|
|
|
|
|
|
my $stop_mode = "AND"; |
|
1480
|
|
|
|
|
|
|
|
|
1481
|
0
|
0
|
|
|
|
|
open ( STP, $stopfile ) || |
|
1482
|
|
|
|
|
|
|
die ("Couldn't open the stoplist file $stopfile\n$!"); |
|
1483
|
|
|
|
|
|
|
|
|
1484
|
0
|
|
|
|
|
|
while ( ) { |
|
1485
|
0
|
|
|
|
|
|
chomp; |
|
1486
|
|
|
|
|
|
|
|
|
1487
|
0
|
0
|
|
|
|
|
if(/\@stop.mode\s*=\s*(\w+)\s*$/) { |
|
1488
|
0
|
|
|
|
|
|
$stop_mode=$1; |
|
1489
|
0
|
0
|
|
|
|
|
if(!($stop_mode=~/^(AND|and|OR|or)$/)) { |
|
1490
|
0
|
|
|
|
|
|
print STDERR "Requested Stop Mode $1 is not supported.\n"; |
|
1491
|
0
|
|
|
|
|
|
exit; |
|
1492
|
|
|
|
|
|
|
} |
|
1493
|
0
|
|
|
|
|
|
next; |
|
1494
|
|
|
|
|
|
|
} |
|
1495
|
|
|
|
|
|
|
|
|
1496
|
|
|
|
|
|
|
# accepting Perl Regexs from Stopfile |
|
1497
|
0
|
|
|
|
|
|
s/^\s+//; |
|
1498
|
0
|
|
|
|
|
|
s/\s+$//; |
|
1499
|
|
|
|
|
|
|
|
|
1500
|
|
|
|
|
|
|
#handling a blank lines |
|
1501
|
0
|
0
|
|
|
|
|
if(/^\s*$/) { next; } |
|
|
0
|
|
|
|
|
|
|
|
1502
|
|
|
|
|
|
|
|
|
1503
|
|
|
|
|
|
|
#check if a valid Perl Regex |
|
1504
|
0
|
0
|
|
|
|
|
if(!(/^\//)) { |
|
1505
|
0
|
|
|
|
|
|
print STDERR "Stop token regular expression <$_> should start with '/'\n"; |
|
1506
|
0
|
|
|
|
|
|
exit; |
|
1507
|
|
|
|
|
|
|
} |
|
1508
|
0
|
0
|
|
|
|
|
if(!(/\/$/)) { |
|
1509
|
0
|
|
|
|
|
|
print STDERR "Stop token regular expression <$_> should end with '/'\n"; |
|
1510
|
0
|
|
|
|
|
|
exit; |
|
1511
|
|
|
|
|
|
|
} |
|
1512
|
|
|
|
|
|
|
|
|
1513
|
|
|
|
|
|
|
#remove the / s from beginning and end |
|
1514
|
0
|
|
|
|
|
|
s/^\///; |
|
1515
|
0
|
|
|
|
|
|
s/\/$//; |
|
1516
|
|
|
|
|
|
|
|
|
1517
|
|
|
|
|
|
|
#form a single big regex |
|
1518
|
0
|
|
|
|
|
|
$stop_regex.="(".$_.")|"; |
|
1519
|
|
|
|
|
|
|
} |
|
1520
|
|
|
|
|
|
|
|
|
1521
|
0
|
0
|
|
|
|
|
if(length($stop_regex)<=0) { |
|
1522
|
0
|
|
|
|
|
|
print STDERR "No valid Perl Regular Experssion found in Stop file $stopfile"; |
|
1523
|
0
|
|
|
|
|
|
exit; |
|
1524
|
|
|
|
|
|
|
} |
|
1525
|
|
|
|
|
|
|
|
|
1526
|
0
|
|
|
|
|
|
chop $stop_regex; |
|
1527
|
|
|
|
|
|
|
|
|
1528
|
|
|
|
|
|
|
# making AND a default stop mode |
|
1529
|
0
|
0
|
|
|
|
|
if(!defined $stop_mode) { |
|
1530
|
0
|
|
|
|
|
|
$stop_mode="AND"; |
|
1531
|
|
|
|
|
|
|
} |
|
1532
|
|
|
|
|
|
|
|
|
1533
|
0
|
|
|
|
|
|
close STP; |
|
1534
|
|
|
|
|
|
|
|
|
1535
|
0
|
|
|
|
|
|
return $stop_regex; |
|
1536
|
|
|
|
|
|
|
} |
|
1537
|
|
|
|
|
|
|
|
|
1538
|
|
|
|
|
|
|
1; |