line |
stmt |
bran |
cond |
sub |
pod |
time |
code |
1
|
|
|
|
|
|
|
#!/usr/bin/perl |
2
|
|
|
|
|
|
|
# Copyright (C) 2018–2021 Alex Schroeder |
3
|
|
|
|
|
|
|
|
4
|
|
|
|
|
|
|
# This program is free software: you can redistribute it and/or modify it under |
5
|
|
|
|
|
|
|
# the terms of the GNU General Public License as published by the Free Software |
6
|
|
|
|
|
|
|
# Foundation, either version 3 of the License, or (at your option) any later |
7
|
|
|
|
|
|
|
# version. |
8
|
|
|
|
|
|
|
# |
9
|
|
|
|
|
|
|
# This program is distributed in the hope that it will be useful, but WITHOUT |
10
|
|
|
|
|
|
|
# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS |
11
|
|
|
|
|
|
|
# FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. |
12
|
|
|
|
|
|
|
# |
13
|
|
|
|
|
|
|
# You should have received a copy of the GNU General Public License along with |
14
|
|
|
|
|
|
|
# this program. If not, see . |
15
|
|
|
|
|
|
|
|
16
|
|
|
|
|
|
|
=encoding utf8 |
17
|
|
|
|
|
|
|
|
18
|
|
|
|
|
|
|
=head1 Moku Pona |
19
|
|
|
|
|
|
|
|
20
|
|
|
|
|
|
|
Moku Pona is a Gemini based feed reader. It can monitor URLs to feeds or regular |
21
|
|
|
|
|
|
|
pages for changes and keeps and updated list of these in a Gemini list. Moku |
22
|
|
|
|
|
|
|
Pona knows how to fetch Gopher URLs, Gemini URLs, and regular web URLs. |
23
|
|
|
|
|
|
|
|
24
|
|
|
|
|
|
|
You manage your subscriptions using the command-line, with Moku Pona. |
25
|
|
|
|
|
|
|
|
26
|
|
|
|
|
|
|
You serve the resulting file using a Gemini server. |
27
|
|
|
|
|
|
|
|
28
|
|
|
|
|
|
|
You read it all using your Gemini client. |
29
|
|
|
|
|
|
|
|
30
|
|
|
|
|
|
|
=head2 Limitations |
31
|
|
|
|
|
|
|
|
32
|
|
|
|
|
|
|
Moku Pona only detects changes. Thus, if there is an item that points to a phlog |
33
|
|
|
|
|
|
|
or blog, that's great. Sometimes people put their phlog in a folder per year. If |
34
|
|
|
|
|
|
|
the Gopher menu lists each folder and a date with the latest change, then that's |
35
|
|
|
|
|
|
|
great, you can use it. Without it, you're in trouble: you need to subscribe to |
36
|
|
|
|
|
|
|
the item for the current year in order to see changes, but when the next year |
37
|
|
|
|
|
|
|
comes around, you're subscribed to the wrong item. Sometimes you're lucky and |
38
|
|
|
|
|
|
|
there will be a menu somewhere with a timestamp for the last change. Use that |
39
|
|
|
|
|
|
|
instead. Good luck! |
40
|
|
|
|
|
|
|
|
41
|
|
|
|
|
|
|
=head2 License |
42
|
|
|
|
|
|
|
|
43
|
|
|
|
|
|
|
GNU Affero General Public License |
44
|
|
|
|
|
|
|
|
45
|
|
|
|
|
|
|
=head2 Installation |
46
|
|
|
|
|
|
|
|
47
|
|
|
|
|
|
|
Using C: |
48
|
|
|
|
|
|
|
|
49
|
|
|
|
|
|
|
cpan App::mokupona |
50
|
|
|
|
|
|
|
|
51
|
|
|
|
|
|
|
Manual install: |
52
|
|
|
|
|
|
|
|
53
|
|
|
|
|
|
|
perl Makefile.PL |
54
|
|
|
|
|
|
|
make |
55
|
|
|
|
|
|
|
make install |
56
|
|
|
|
|
|
|
|
57
|
|
|
|
|
|
|
=head2 Dependencies |
58
|
|
|
|
|
|
|
|
59
|
|
|
|
|
|
|
There are some Perl dependencies you need to satisfy in order to run this |
60
|
|
|
|
|
|
|
program: |
61
|
|
|
|
|
|
|
|
62
|
|
|
|
|
|
|
=over |
63
|
|
|
|
|
|
|
|
64
|
|
|
|
|
|
|
=item L, or C |
65
|
|
|
|
|
|
|
|
66
|
|
|
|
|
|
|
=item L, or C |
67
|
|
|
|
|
|
|
|
68
|
|
|
|
|
|
|
=item L, or C |
69
|
|
|
|
|
|
|
|
70
|
|
|
|
|
|
|
=item L, or C |
71
|
|
|
|
|
|
|
|
72
|
|
|
|
|
|
|
=item L, or c |
73
|
|
|
|
|
|
|
|
74
|
|
|
|
|
|
|
=back |
75
|
|
|
|
|
|
|
|
76
|
|
|
|
|
|
|
=cut |
77
|
|
|
|
|
|
|
|
78
|
9
|
|
|
9
|
|
11982
|
use Modern::Perl '2018'; |
|
9
|
|
|
|
|
41252
|
|
|
9
|
|
|
|
|
112
|
|
79
|
9
|
|
|
9
|
|
7253
|
use File::Copy qw(copy); |
|
9
|
|
|
|
|
19744
|
|
|
9
|
|
|
|
|
872
|
|
80
|
9
|
|
|
9
|
|
5577
|
use Encode::Locale qw(decode_argv); |
|
9
|
|
|
|
|
102381
|
|
|
9
|
|
|
|
|
718
|
|
81
|
9
|
|
|
9
|
|
88
|
use Encode qw(decode_utf8); |
|
9
|
|
|
|
|
26
|
|
|
9
|
|
|
|
|
647
|
|
82
|
9
|
|
|
9
|
|
4400
|
use Mojo::IOLoop; |
|
9
|
|
|
|
|
2741509
|
|
|
9
|
|
|
|
|
117
|
|
83
|
9
|
|
|
9
|
|
7066
|
use XML::LibXML; |
|
9
|
|
|
|
|
302974
|
|
|
9
|
|
|
|
|
67
|
|
84
|
9
|
|
|
9
|
|
6587
|
use URI::Escape; |
|
9
|
|
|
|
|
15083
|
|
|
9
|
|
|
|
|
51850
|
|
85
|
|
|
|
|
|
|
|
86
|
|
|
|
|
|
|
decode_argv(); |
87
|
|
|
|
|
|
|
|
88
|
|
|
|
|
|
|
if (-t) { |
89
|
|
|
|
|
|
|
binmode(STDIN, ":encoding(console_in)"); |
90
|
|
|
|
|
|
|
binmode(STDOUT, ":encoding(console_out)"); |
91
|
|
|
|
|
|
|
binmode(STDERR, ":encoding(console_out)"); |
92
|
|
|
|
|
|
|
} |
93
|
|
|
|
|
|
|
|
94
|
|
|
|
|
|
|
=head2 The Data Directory |
95
|
|
|
|
|
|
|
|
96
|
|
|
|
|
|
|
Moku Pona keeps the list of URLs you are subscribed to in directory. It's |
97
|
|
|
|
|
|
|
probably C<~/.moku-pona> on your system. |
98
|
|
|
|
|
|
|
|
99
|
|
|
|
|
|
|
=over |
100
|
|
|
|
|
|
|
|
101
|
|
|
|
|
|
|
=item If you have the C environment variable set, then that's your data |
102
|
|
|
|
|
|
|
directory. |
103
|
|
|
|
|
|
|
|
104
|
|
|
|
|
|
|
=item If you have the C environment variable set, then your data |
105
|
|
|
|
|
|
|
directory is F<$XDG_DATA_HOME/moku-pona>. |
106
|
|
|
|
|
|
|
|
107
|
|
|
|
|
|
|
=item If you you have the C environment variable set, and you have a |
108
|
|
|
|
|
|
|
F<$HOME/.local> directory, then your data directory is |
109
|
|
|
|
|
|
|
F<$HOME/.local/moku-pona>. |
110
|
|
|
|
|
|
|
|
111
|
|
|
|
|
|
|
=item If you have the C environment variable set, then your data directory |
112
|
|
|
|
|
|
|
is F<$HOME/.moku-pona>. |
113
|
|
|
|
|
|
|
|
114
|
|
|
|
|
|
|
=item If you have the C environment variable set (Windows), then your |
115
|
|
|
|
|
|
|
data directory is F<$APPDATA/moku-pona>. |
116
|
|
|
|
|
|
|
|
117
|
|
|
|
|
|
|
=item The last option is to have the C environment variable set. |
118
|
|
|
|
|
|
|
|
119
|
|
|
|
|
|
|
=back |
120
|
|
|
|
|
|
|
|
121
|
|
|
|
|
|
|
The data directory contains a copy of the latest resources. The names of these |
122
|
|
|
|
|
|
|
cache files are simply the URL with all the slashes replaced by a hyphen. |
123
|
|
|
|
|
|
|
|
124
|
|
|
|
|
|
|
=cut |
125
|
|
|
|
|
|
|
|
126
|
|
|
|
|
|
|
our $data_dir = data_dir(); |
127
|
|
|
|
|
|
|
# say "Using $data_dir"; |
128
|
|
|
|
|
|
|
|
129
|
|
|
|
|
|
|
sub data_dir { |
130
|
9
|
50
|
|
9
|
|
48
|
return $ENV{MOKU_PONA} if $ENV{MOKU_PONA}; |
131
|
|
|
|
|
|
|
# find an existing directory |
132
|
9
|
50
|
33
|
|
|
58
|
return $ENV{XDG_DATA_HOME} . '/moku-pona' if $ENV{XDG_DATA_HOME} and -d $ENV{XDG_DATA_HOME} . '/moku-pona'; |
133
|
9
|
50
|
33
|
|
|
252
|
return $ENV{HOME} . '/.local/moku-pona' if $ENV{HOME} and -d $ENV{HOME} . '/.local/moku-pona'; |
134
|
9
|
50
|
33
|
|
|
216
|
return $ENV{HOME} . '/.moku-pona' if $ENV{HOME} and -d $ENV{HOME} . '/.moku-pona'; |
135
|
9
|
50
|
33
|
|
|
57
|
return $ENV{APPDATA} . '/moku-pona' if $ENV{APPDATA} and -d $ENV{APPDATA} . '/.moku-pona'; |
136
|
9
|
50
|
33
|
|
|
43
|
return $ENV{LOGDIR} . '/.moku-pona' if $ENV{LOGDIR} and -d $ENV{LOGDIR} . '/.moku-pona'; |
137
|
|
|
|
|
|
|
# or use a new one |
138
|
9
|
50
|
|
|
|
38
|
return $ENV{XDG_DATA_HOME} . '/moku-pona' if $ENV{XDG_DATA_HOME}; |
139
|
9
|
50
|
33
|
|
|
150
|
return $ENV{HOME} . '/.local/moku-pona' if $ENV{HOME} and -d $ENV{HOME} . '/.local'; |
140
|
9
|
50
|
|
|
|
73
|
return $ENV{HOME} . '/.moku-pona' if $ENV{HOME}; |
141
|
0
|
0
|
|
|
|
0
|
return $ENV{APPDATA} . '/moku-pona' if $ENV{APPDATA}; |
142
|
0
|
0
|
|
|
|
0
|
return $ENV{LOGDIR} . '/.moku-pona' if $ENV{LOGDIR}; |
143
|
0
|
|
|
|
|
0
|
die "Please set the MOKU_PONA environment variable to a directory name\n"; |
144
|
|
|
|
|
|
|
} |
145
|
|
|
|
|
|
|
|
146
|
|
|
|
|
|
|
=pod |
147
|
|
|
|
|
|
|
|
148
|
|
|
|
|
|
|
The C file is a file containing a gemtext list of links, i.e. entries |
149
|
|
|
|
|
|
|
such as these: |
150
|
|
|
|
|
|
|
|
151
|
|
|
|
|
|
|
=> gemini://alexschroeder.ch Alex Schroeder |
152
|
|
|
|
|
|
|
|
153
|
|
|
|
|
|
|
=cut |
154
|
|
|
|
|
|
|
|
155
|
|
|
|
|
|
|
our $site_list = $data_dir . '/sites.txt'; |
156
|
|
|
|
|
|
|
|
157
|
|
|
|
|
|
|
=pod |
158
|
|
|
|
|
|
|
|
159
|
|
|
|
|
|
|
The C file is a file containing a gemtext list of links based on |
160
|
|
|
|
|
|
|
C, but with a timestamp of their last change, and with new updates |
161
|
|
|
|
|
|
|
moved to the top. The ISO date is simply inserted after the URL: |
162
|
|
|
|
|
|
|
|
163
|
|
|
|
|
|
|
=> gemini://alexschroeder.ch 2020-11-07 Alex Schroeder |
164
|
|
|
|
|
|
|
|
165
|
|
|
|
|
|
|
=cut |
166
|
|
|
|
|
|
|
|
167
|
|
|
|
|
|
|
our $updated_list = $data_dir . '/updates.txt'; |
168
|
|
|
|
|
|
|
|
169
|
|
|
|
|
|
|
=pod |
170
|
|
|
|
|
|
|
|
171
|
|
|
|
|
|
|
In order to be at least somewhat backwards compatible with Moku Pona versions |
172
|
|
|
|
|
|
|
1.1 and earlier, C may contain Gopher menu items. These are converted |
173
|
|
|
|
|
|
|
to Gemini URLs during processing and thus the C file still contains |
174
|
|
|
|
|
|
|
regular gemtext. |
175
|
|
|
|
|
|
|
|
176
|
|
|
|
|
|
|
1Alex Schroeder ⭾ ⭾ alexschroeder.ch ⭾ 70 |
177
|
|
|
|
|
|
|
|
178
|
|
|
|
|
|
|
=cut |
179
|
|
|
|
|
|
|
|
180
|
|
|
|
|
|
|
sub convert { |
181
|
25
|
|
|
25
|
|
131
|
for (@_) { |
182
|
44
|
100
|
|
|
|
212
|
next if /^=> /; # is already a gemini link |
183
|
14
|
|
|
|
|
115
|
my ($type, $desc, $selector, $host, $port) = /^([^\t])([^\t]*)\t([^\t]*)\t([^\t]*)\t([^\t\r]*)/; |
184
|
14
|
100
|
66
|
|
|
65
|
if ($host and $port) { |
185
|
11
|
|
50
|
|
|
28
|
$port //= 0; |
186
|
11
|
|
|
|
|
54
|
$_ = "=> gopher://$host:$port/$type$selector $desc"; |
187
|
|
|
|
|
|
|
} |
188
|
|
|
|
|
|
|
} |
189
|
25
|
|
|
|
|
159
|
return @_; |
190
|
|
|
|
|
|
|
} |
191
|
|
|
|
|
|
|
|
192
|
|
|
|
|
|
|
=pod |
193
|
|
|
|
|
|
|
|
194
|
|
|
|
|
|
|
As was said above, however, the recommended format is the use of URLs. Moku Pona |
195
|
|
|
|
|
|
|
supports Gemini, Gopher, and the web (gemini, gopher, gophers, http, and https |
196
|
|
|
|
|
|
|
schemes). |
197
|
|
|
|
|
|
|
|
198
|
|
|
|
|
|
|
=cut |
199
|
|
|
|
|
|
|
|
200
|
|
|
|
|
|
|
sub query_gemini { |
201
|
0
|
|
|
0
|
|
0
|
my $url = shift; |
202
|
0
|
|
|
|
|
0
|
my $responses = shift; |
203
|
0
|
|
|
|
|
0
|
my($scheme, $authority, $path, $query, $fragment) = |
204
|
|
|
|
|
|
|
$url =~ m|(?:([^:/?#]+):)?(?://([^/?#]*))?([^?#]*)(?:\?([^#]*))?(?:#(.*))?|; |
205
|
0
|
0
|
0
|
|
|
0
|
die "⚠ The URL '$url' must use the gemini scheme\n" unless $scheme and $scheme eq 'gemini'; |
206
|
0
|
0
|
|
|
|
0
|
die "⚠ The URL '$url' must have an authority\n" unless $authority; |
207
|
0
|
|
|
|
|
0
|
my ($host, $port) = split(/:/, $authority, 2); |
208
|
0
|
|
0
|
|
|
0
|
$port //= 1965; |
209
|
|
|
|
|
|
|
Mojo::IOLoop->client( |
210
|
|
|
|
|
|
|
{port => $port, address => $host, tls => 1, |
211
|
|
|
|
|
|
|
tls_options => { SSL_verify_mode => 0x00 }} |
212
|
|
|
|
|
|
|
=> sub { |
213
|
0
|
|
|
0
|
|
0
|
my ($loop, $err, $stream) = @_; |
214
|
0
|
0
|
|
|
|
0
|
if (not $stream) { |
215
|
0
|
|
|
|
|
0
|
warn "Cannot connect to $url\n"; |
216
|
0
|
|
|
|
|
0
|
return; |
217
|
|
|
|
|
|
|
} |
218
|
0
|
|
|
|
|
0
|
my $header; |
219
|
|
|
|
|
|
|
$stream->on(read => sub { |
220
|
0
|
|
|
|
|
0
|
my ($stream, $bytes) = @_; |
221
|
0
|
|
|
|
|
0
|
$responses->{$url} .= $bytes; |
222
|
0
|
0
|
|
|
|
0
|
$header = $responses->{$url} =~ s/^.*\r\n// unless $header}); |
|
0
|
|
|
|
|
0
|
|
223
|
0
|
|
|
|
|
0
|
$stream->write("$url\r\n")}) |
224
|
0
|
|
|
|
|
0
|
} |
225
|
|
|
|
|
|
|
|
226
|
|
|
|
|
|
|
sub query_gopher { |
227
|
10
|
|
|
10
|
|
20
|
my $url = shift; |
228
|
10
|
|
|
|
|
13
|
my $responses = shift; |
229
|
10
|
|
|
|
|
21
|
my ($selector, $host, $port) = url_to_gopher($url); |
230
|
10
|
|
|
|
|
26
|
my $tls = $url =~ /^gophers/; |
231
|
|
|
|
|
|
|
Mojo::IOLoop->client( |
232
|
|
|
|
|
|
|
{port => $port, address => $host, tls => $tls } |
233
|
|
|
|
|
|
|
=> sub { |
234
|
10
|
|
|
10
|
|
14760
|
my ($loop, $err, $stream) = @_; |
235
|
10
|
50
|
|
|
|
37
|
if (not $stream) { |
236
|
0
|
|
|
|
|
0
|
warn "Cannot connect to $url\n"; |
237
|
0
|
|
|
|
|
0
|
return; |
238
|
|
|
|
|
|
|
} |
239
|
|
|
|
|
|
|
$stream->on( |
240
|
|
|
|
|
|
|
read => sub { |
241
|
10
|
|
|
|
|
6623
|
my ($stream, $bytes) = @_; |
242
|
10
|
|
|
|
|
72
|
$responses->{$url} .= $bytes}); |
|
10
|
|
|
|
|
107
|
|
243
|
10
|
|
|
|
|
102
|
$stream->write("$selector\r\n")}) |
244
|
10
|
|
|
|
|
229
|
} |
245
|
|
|
|
|
|
|
|
246
|
|
|
|
|
|
|
sub url_to_gopher { |
247
|
10
|
|
|
10
|
|
17
|
my $url = shift; |
248
|
10
|
|
33
|
|
|
38
|
my $name = shift||$url; |
249
|
10
|
|
|
|
|
75
|
my($scheme, $authority, $path, $query, $fragment) = |
250
|
|
|
|
|
|
|
$url =~ m|(?:([^:/?#]+):)?(?://([^/?#]*))?([^?#]*)(?:\?([^#]*))?(?:#(.*))?|; |
251
|
10
|
|
50
|
|
|
35
|
$scheme ||= "gopher"; |
252
|
10
|
50
|
|
|
|
40
|
return unless $scheme =~ /^gophers?$/; |
253
|
10
|
50
|
|
|
|
63
|
$path = substr($path, 1) if substr($path, 0, 1) eq "/"; |
254
|
10
|
50
|
|
|
|
24
|
my $type = $path ? substr($path, 0, 1) : "1"; |
255
|
10
|
50
|
|
|
|
22
|
my $selector = $path ? substr($path, 1) : ""; |
256
|
10
|
|
|
|
|
44
|
my ($host, $port) = split(/:/, $authority, 2); |
257
|
10
|
|
50
|
|
|
52
|
return ($selector, $host, $port||70); |
258
|
|
|
|
|
|
|
} |
259
|
|
|
|
|
|
|
|
260
|
|
|
|
|
|
|
sub query_web { |
261
|
0
|
|
|
0
|
|
0
|
my $url = shift; |
262
|
0
|
|
|
|
|
0
|
my $responses = shift; |
263
|
0
|
|
|
|
|
0
|
my($scheme, $authority, $path, $query, $fragment) = |
264
|
|
|
|
|
|
|
$url =~ m|(?:([^:/?#]+):)?(?://([^/?#]*))?([^?#]*)(?:\?([^#]*))?(?:#(.*))?|; |
265
|
0
|
|
|
|
|
0
|
my $tls = $scheme eq "https"; |
266
|
0
|
|
|
|
|
0
|
my ($host, $port) = split(/:/, $authority, 2); |
267
|
0
|
|
0
|
|
|
0
|
$path ||= "/"; |
268
|
0
|
|
|
|
|
0
|
my $selector = $path; |
269
|
0
|
0
|
|
|
|
0
|
$selector .= "?" . $query if $query; |
270
|
|
|
|
|
|
|
# ignore the fragment |
271
|
0
|
0
|
0
|
|
|
0
|
$port ||= $tls ? 443 : 80; |
272
|
|
|
|
|
|
|
Mojo::IOLoop->client( |
273
|
|
|
|
|
|
|
{port => $port, address => $host, tls => $tls } |
274
|
|
|
|
|
|
|
=> sub { |
275
|
0
|
|
|
0
|
|
0
|
my ($loop, $err, $stream) = @_; |
276
|
0
|
|
|
|
|
0
|
my $header; |
277
|
|
|
|
|
|
|
$stream->on(read => sub { |
278
|
0
|
|
|
|
|
0
|
my ($stream, $bytes) = @_; |
279
|
0
|
|
|
|
|
0
|
$responses->{$url} .= $bytes; |
280
|
0
|
0
|
|
|
|
0
|
$header = $responses->{$url} =~ s/^.*\r\n\r\n//s unless $header}); |
|
0
|
|
|
|
|
0
|
|
281
|
0
|
|
|
|
|
0
|
$stream->write("GET $selector HTTP/1.0\r\n" |
282
|
|
|
|
|
|
|
. "Host: $host:$port\r\n" |
283
|
|
|
|
|
|
|
. "User-Agent: moku-pona\r\n" |
284
|
0
|
|
|
|
|
0
|
. "\r\n")}); |
285
|
|
|
|
|
|
|
} |
286
|
|
|
|
|
|
|
|
287
|
|
|
|
|
|
|
=head2 Migration from 1.1 |
288
|
|
|
|
|
|
|
|
289
|
|
|
|
|
|
|
The best way to migrate your setup is probably to use the C subcommand |
290
|
|
|
|
|
|
|
explained later, and to recreate your list of subscriptions. Then your |
291
|
|
|
|
|
|
|
C file will use gemtext format. |
292
|
|
|
|
|
|
|
|
293
|
|
|
|
|
|
|
moku-pona list | grep "moku-pona add" > commands |
294
|
|
|
|
|
|
|
mv ~/.moku-pona/sites.txt ~/.moku-pona/sites.txt~ |
295
|
|
|
|
|
|
|
sh commands |
296
|
|
|
|
|
|
|
|
297
|
|
|
|
|
|
|
=cut |
298
|
|
|
|
|
|
|
|
299
|
|
|
|
|
|
|
sub load_site { |
300
|
29
|
|
|
29
|
|
1274
|
my $file = $site_list; |
301
|
29
|
100
|
|
|
|
589
|
return [] if not -f $file; |
302
|
21
|
50
|
|
|
|
845
|
open(my $fh, "<:encoding(UTF-8)", $file) or die "Cannot read $file: $!\n"; |
303
|
21
|
|
|
|
|
3414
|
my @lines = <$fh>; |
304
|
21
|
|
|
|
|
468
|
chomp(@lines); |
305
|
21
|
|
|
|
|
86
|
@lines = grep(/^=> /, convert(@lines)); # from gopher |
306
|
21
|
|
|
|
|
422
|
return \@lines; |
307
|
|
|
|
|
|
|
} |
308
|
|
|
|
|
|
|
|
309
|
|
|
|
|
|
|
sub load_file { |
310
|
17
|
|
|
17
|
|
3691
|
my $file = shift; |
311
|
17
|
100
|
|
|
|
418
|
return "" if not -f $file; |
312
|
14
|
50
|
|
|
|
546
|
open(my $fh, "<:encoding(UTF-8)", $file) |
313
|
|
|
|
|
|
|
or die "Cannot read $file: $!\n"; |
314
|
14
|
|
|
|
|
1242
|
local $/ = undef; |
315
|
14
|
|
|
|
|
408
|
return <$fh>; |
316
|
|
|
|
|
|
|
} |
317
|
|
|
|
|
|
|
|
318
|
|
|
|
|
|
|
sub save_file { |
319
|
22
|
|
|
22
|
|
2247
|
my $file = shift; |
320
|
22
|
|
|
|
|
32
|
my $data = shift; |
321
|
22
|
100
|
|
|
|
479
|
mkdir $data_dir unless -d $data_dir; |
322
|
4
|
50
|
|
4
|
|
34
|
open(my $fh, ">:encoding(UTF-8)", $file) |
|
4
|
|
|
|
|
8
|
|
|
4
|
|
|
|
|
33
|
|
|
22
|
|
|
|
|
138739
|
|
323
|
|
|
|
|
|
|
or die "Cannot write $file: $!\n"; |
324
|
22
|
|
|
|
|
166635
|
print $fh $data; |
325
|
|
|
|
|
|
|
} |
326
|
|
|
|
|
|
|
|
327
|
|
|
|
|
|
|
=head2 List your subscriptions |
328
|
|
|
|
|
|
|
|
329
|
|
|
|
|
|
|
moku-pona list |
330
|
|
|
|
|
|
|
|
331
|
|
|
|
|
|
|
This lists all your current subscriptions in a format that is suitable for a |
332
|
|
|
|
|
|
|
shell script! |
333
|
|
|
|
|
|
|
|
334
|
|
|
|
|
|
|
Example: |
335
|
|
|
|
|
|
|
|
336
|
|
|
|
|
|
|
moku-pona list | grep "alexschroeder" |
337
|
|
|
|
|
|
|
|
338
|
|
|
|
|
|
|
In this particular case, since I'm testing my own server, the result would be: |
339
|
|
|
|
|
|
|
|
340
|
|
|
|
|
|
|
moku-pona add https://alexschroeder.ch/wiki?action=rss "rss" |
341
|
|
|
|
|
|
|
moku-pona add gemini://alexschroeder.ch/ "gemini" |
342
|
|
|
|
|
|
|
moku-pona add gopher://alexschroeder.ch/ "gopher" |
343
|
|
|
|
|
|
|
moku-pona add gophers://alexschroeder.ch:7443/ "gophers" |
344
|
|
|
|
|
|
|
|
345
|
|
|
|
|
|
|
=cut |
346
|
|
|
|
|
|
|
|
347
|
|
|
|
|
|
|
sub do_list { |
348
|
1
|
|
|
1
|
|
1576
|
my $site = load_site(); |
349
|
1
|
|
|
|
|
7
|
print("Subscribed items in $site_list:\n"); |
350
|
1
|
50
|
|
|
|
5
|
print("none\n") unless @$site; |
351
|
1
|
|
|
|
|
2
|
for my $line (@$site) { |
352
|
|
|
|
|
|
|
# skip item type |
353
|
2
|
|
|
|
|
12
|
my ($uri, $name) = $line =~ /^=> (\S+)\s+(.*)/; |
354
|
2
|
|
|
|
|
12
|
print(qq{moku-pona add $uri "$name"\n}); |
355
|
|
|
|
|
|
|
} |
356
|
|
|
|
|
|
|
} |
357
|
|
|
|
|
|
|
|
358
|
|
|
|
|
|
|
=head2 Add a subscription |
359
|
|
|
|
|
|
|
|
360
|
|
|
|
|
|
|
moku-pona add url [description] |
361
|
|
|
|
|
|
|
|
362
|
|
|
|
|
|
|
This adds a URL to the list of subscribed items. If the target is an Atom or RSS |
363
|
|
|
|
|
|
|
feed, then that's also supported. You can provide an optional description for |
364
|
|
|
|
|
|
|
this URL. If you don't provide a description, the URL will be used as the item's |
365
|
|
|
|
|
|
|
description. |
366
|
|
|
|
|
|
|
|
367
|
|
|
|
|
|
|
Example: |
368
|
|
|
|
|
|
|
|
369
|
|
|
|
|
|
|
moku-pona add gemini://alexschroeder.ch kensanata |
370
|
|
|
|
|
|
|
|
371
|
|
|
|
|
|
|
=cut |
372
|
|
|
|
|
|
|
|
373
|
|
|
|
|
|
|
sub do_add { |
374
|
8
|
|
|
8
|
|
1576
|
my $uri = shift; |
375
|
8
|
|
|
|
|
17
|
my $name = shift; |
376
|
8
|
|
33
|
|
|
25
|
$name ||= $uri; |
377
|
8
|
|
|
|
|
25
|
my $line = "=> $uri $name"; |
378
|
8
|
|
|
|
|
33
|
my $site = load_site(); |
379
|
8
|
|
|
|
|
29
|
my $uri_re = quotemeta($uri); |
380
|
8
|
|
|
|
|
21
|
my $name_re = quotemeta($name); |
381
|
8
|
50
|
|
|
|
189
|
if (grep(/^=> $uri_re /, @$site)) { |
|
|
50
|
|
|
|
|
|
382
|
0
|
|
|
|
|
0
|
warn("$uri already exists in $site_list\n"); |
383
|
|
|
|
|
|
|
} elsif (grep(/^=> \S+ $name_re/, @$site)) { |
384
|
0
|
|
|
|
|
0
|
warn("$name already exists in $site_list\n"); |
385
|
|
|
|
|
|
|
} else { |
386
|
8
|
|
|
|
|
36
|
push(@$site, $line); |
387
|
|
|
|
|
|
|
} |
388
|
8
|
|
|
|
|
42
|
save_file($site_list, join("\n", @$site, "")); |
389
|
|
|
|
|
|
|
} |
390
|
|
|
|
|
|
|
|
391
|
|
|
|
|
|
|
=head2 Remove a subscription |
392
|
|
|
|
|
|
|
|
393
|
|
|
|
|
|
|
moku-pona remove description |
394
|
|
|
|
|
|
|
|
395
|
|
|
|
|
|
|
This removes one or more URLs from the list of subscribed items. |
396
|
|
|
|
|
|
|
|
397
|
|
|
|
|
|
|
Example: |
398
|
|
|
|
|
|
|
|
399
|
|
|
|
|
|
|
moku-pona remove kensanata |
400
|
|
|
|
|
|
|
|
401
|
|
|
|
|
|
|
=cut |
402
|
|
|
|
|
|
|
|
403
|
|
|
|
|
|
|
sub do_remove { |
404
|
2
|
|
|
2
|
|
2073
|
my @args = @_; |
405
|
2
|
|
|
|
|
8
|
my $site = load_site(); |
406
|
2
|
|
|
|
|
8
|
my $count = 0; |
407
|
2
|
|
|
|
|
5
|
my $i = 0; |
408
|
2
|
|
66
|
|
|
18
|
while (@args and $i < @$site) { |
409
|
4
|
|
|
|
|
29
|
my $line = $site->[$i]; |
410
|
4
|
|
|
|
|
30
|
my ($uri, $name) = $line =~ /^=> (\S+)\s+(.*)/; |
411
|
4
|
|
|
|
|
7
|
my $found = 0; |
412
|
4
|
|
|
|
|
5
|
my $j = 0; |
413
|
4
|
|
|
|
|
12
|
while ($j < @args) { |
414
|
4
|
100
|
|
|
|
12
|
if ($name eq $args[$j]) { |
415
|
2
|
|
|
|
|
4
|
$count++; |
416
|
2
|
|
|
|
|
5
|
$found = 1; |
417
|
2
|
|
|
|
|
7
|
splice(@$site, $i, 1); # remove the site found |
418
|
2
|
|
|
|
|
5
|
splice(@args, $j, 1); # remove the arg found |
419
|
|
|
|
|
|
|
} else { |
420
|
2
|
|
|
|
|
7
|
$j++; |
421
|
|
|
|
|
|
|
} |
422
|
|
|
|
|
|
|
} |
423
|
4
|
100
|
|
|
|
17
|
$i++ unless $found; |
424
|
|
|
|
|
|
|
} |
425
|
2
|
50
|
|
|
|
7
|
if ($count) { |
426
|
2
|
50
|
|
|
|
98
|
printf("Removed %d %s\n", $count, |
427
|
|
|
|
|
|
|
$count == 1 ? "subscription" : "subscriptions"); |
428
|
2
|
|
|
|
|
21
|
save_file($site_list, join("\n", @$site, "")); |
429
|
|
|
|
|
|
|
} else { |
430
|
0
|
|
|
|
|
0
|
warn("No subscriptions matching @args found\n"); |
431
|
0
|
|
|
|
|
0
|
warn("Use moku-pona list to find the correct descriptions.\n"); |
432
|
|
|
|
|
|
|
} |
433
|
|
|
|
|
|
|
} |
434
|
|
|
|
|
|
|
|
435
|
|
|
|
|
|
|
=head2 Clean up the data directory |
436
|
|
|
|
|
|
|
|
437
|
|
|
|
|
|
|
moku-pona cleanup [--confirm] |
438
|
|
|
|
|
|
|
|
439
|
|
|
|
|
|
|
When Moku Pona updates, copies of the URL targets are saved in the data |
440
|
|
|
|
|
|
|
directory. If you remove a subscription (see above), that leaves a cache file in |
441
|
|
|
|
|
|
|
the data directory that is no longer used – and it leaves an entry in |
442
|
|
|
|
|
|
|
C that is no longer wanted. The cleanup command fixes this. It |
443
|
|
|
|
|
|
|
deletes all the cached pages that you are no longer subscribed to, and it |
444
|
|
|
|
|
|
|
removes those entries from C as well. |
445
|
|
|
|
|
|
|
|
446
|
|
|
|
|
|
|
Actually, just to be sure, if you run it without the C<--confirm> argument, it |
447
|
|
|
|
|
|
|
simply prints which files it would trash. Rerun it with the C<--confirm> |
448
|
|
|
|
|
|
|
argument to actually do it. |
449
|
|
|
|
|
|
|
|
450
|
|
|
|
|
|
|
Example: |
451
|
|
|
|
|
|
|
|
452
|
|
|
|
|
|
|
moku-pona cleanup |
453
|
|
|
|
|
|
|
|
454
|
|
|
|
|
|
|
=cut |
455
|
|
|
|
|
|
|
|
456
|
|
|
|
|
|
|
sub do_cleanup { |
457
|
2
|
|
100
|
2
|
|
3665
|
my $confirm = shift||'' eq '--confirm'; |
458
|
2
|
|
|
|
|
5
|
my $todo = 0; |
459
|
|
|
|
|
|
|
# get a hash map telling us the cache files we expect based on our sites |
460
|
2
|
|
|
|
|
8
|
my $site = load_site(); |
461
|
|
|
|
|
|
|
my %caches = map { |
462
|
2
|
|
|
|
|
9
|
my ($uri, $name) = /^=> (\S+)\s+(.*)/; |
|
2
|
|
|
|
|
18
|
|
463
|
2
|
|
|
|
|
18
|
$uri =~ s/[\/:]/-/g; |
464
|
2
|
|
|
|
|
16
|
"$data_dir/$uri" => 1; |
465
|
|
|
|
|
|
|
} @$site; |
466
|
|
|
|
|
|
|
# get a list of text files in the directory |
467
|
2
|
50
|
|
|
|
78
|
opendir(my $dh, $data_dir) or die "Cannot read $data_dir: $!\n"; |
468
|
2
|
|
|
|
|
58
|
my @files = map { "$data_dir/$_" } grep { /^[^.]/ } readdir($dh); |
|
8
|
|
|
|
|
21
|
|
|
12
|
|
|
|
|
31
|
|
469
|
2
|
|
|
|
|
28
|
closedir($dh); |
470
|
|
|
|
|
|
|
# remove unnecessary cache files |
471
|
2
|
|
|
|
|
7
|
for my $file (@files) { |
472
|
8
|
100
|
|
|
|
19
|
next if $file eq $site_list; |
473
|
6
|
100
|
|
|
|
28
|
next if $file eq $updated_list; |
474
|
4
|
100
|
|
|
|
10
|
next if $caches{$file}; |
475
|
2
|
100
|
|
|
|
7
|
if ($confirm) { |
476
|
1
|
|
|
|
|
63
|
unlink $file; |
477
|
|
|
|
|
|
|
} else { |
478
|
1
|
|
|
|
|
47
|
print "trash $file\n"; |
479
|
1
|
|
|
|
|
4
|
$todo++; |
480
|
|
|
|
|
|
|
} |
481
|
|
|
|
|
|
|
} |
482
|
|
|
|
|
|
|
# check updates list |
483
|
2
|
50
|
|
|
|
32
|
if (-f $updated_list) { |
484
|
2
|
50
|
|
|
|
67
|
open(my $fh, "<:encoding(UTF-8)", $updated_list) |
485
|
|
|
|
|
|
|
or die "Cannot read $updated_list: $!\n"; |
486
|
2
|
|
|
|
|
191
|
my @lines = <$fh>; |
487
|
2
|
|
|
|
|
42
|
chomp(@lines); |
488
|
|
|
|
|
|
|
# decide what to do about each line in updates, looking just at the names |
489
|
2
|
|
|
|
|
7
|
my %sites = map { s/^=> (\S+)\s+(.*)/$2/; $_ => 1 } @$site; |
|
2
|
|
|
|
|
21
|
|
|
2
|
|
|
|
|
10
|
|
490
|
2
|
|
|
|
|
5
|
my @deletes; |
491
|
|
|
|
|
|
|
my @keeps; |
492
|
2
|
|
|
|
|
39
|
for my $line (@lines) { |
493
|
4
|
100
|
66
|
|
|
40
|
if ($line =~ /^=> \S+ \d\d\d\d-\d\d-\d\d (.+)/ and not $sites{$1}) { |
494
|
2
|
|
|
|
|
4
|
push(@deletes, $line); |
495
|
2
|
|
|
|
|
5
|
$todo++; |
496
|
|
|
|
|
|
|
} else { |
497
|
2
|
|
|
|
|
5
|
push(@keeps, $line); |
498
|
|
|
|
|
|
|
} |
499
|
|
|
|
|
|
|
} |
500
|
2
|
100
|
66
|
|
|
17
|
print "Removing these entries from updates:\n" |
501
|
|
|
|
|
|
|
. join("\n", @deletes, "") if @deletes and not $confirm; |
502
|
|
|
|
|
|
|
# save |
503
|
2
|
100
|
|
|
|
27
|
save_file($updated_list, join("\n", @keeps, "")) if $confirm; |
504
|
|
|
|
|
|
|
} |
505
|
2
|
100
|
66
|
|
|
46
|
if ($todo && !$confirm) { |
506
|
1
|
|
|
|
|
3
|
print "\n"; |
507
|
1
|
|
|
|
|
6
|
print "Use moku-pona cleanup --confirm to do it.\n"; |
508
|
|
|
|
|
|
|
} |
509
|
|
|
|
|
|
|
} |
510
|
|
|
|
|
|
|
|
511
|
|
|
|
|
|
|
=head2 Update |
512
|
|
|
|
|
|
|
|
513
|
|
|
|
|
|
|
moku-pona update [--quiet] [names...] |
514
|
|
|
|
|
|
|
|
515
|
|
|
|
|
|
|
This updates all the subscribed items and generates a new local page for you to |
516
|
|
|
|
|
|
|
visit: C. |
517
|
|
|
|
|
|
|
|
518
|
|
|
|
|
|
|
Example: |
519
|
|
|
|
|
|
|
|
520
|
|
|
|
|
|
|
moku-pona update |
521
|
|
|
|
|
|
|
|
522
|
|
|
|
|
|
|
If you call it from a cron job, you might want to use the C<--quiet> argument to |
523
|
|
|
|
|
|
|
prevent it from printing all the sites it's contacting (since cron will then |
524
|
|
|
|
|
|
|
mail this to you and you might not care for it unless there's a problem). If |
525
|
|
|
|
|
|
|
there's a problem, you'll still get a message. |
526
|
|
|
|
|
|
|
|
527
|
|
|
|
|
|
|
This is how I call it from my C, for example |
528
|
|
|
|
|
|
|
|
529
|
|
|
|
|
|
|
#m h dom mon dow command |
530
|
|
|
|
|
|
|
11 7,14 * * * /home/alex/bin/moku-pona update --quiet |
531
|
|
|
|
|
|
|
|
532
|
|
|
|
|
|
|
If you're testing things, you can also fetch just a limited number of items by |
533
|
|
|
|
|
|
|
listing them. |
534
|
|
|
|
|
|
|
|
535
|
|
|
|
|
|
|
Example: |
536
|
|
|
|
|
|
|
|
537
|
|
|
|
|
|
|
moku-pona update "RPG Planet" |
538
|
|
|
|
|
|
|
|
539
|
|
|
|
|
|
|
The C files may contain lines that are not links at the top. These |
540
|
|
|
|
|
|
|
will remain untouched. The rest is links. New items are added at the beginning |
541
|
|
|
|
|
|
|
of the links and older copies of such items are removed from the links. |
542
|
|
|
|
|
|
|
|
543
|
|
|
|
|
|
|
=cut |
544
|
|
|
|
|
|
|
|
545
|
|
|
|
|
|
|
sub add_update { |
546
|
4
|
|
|
4
|
|
13
|
my $line = shift; |
547
|
4
|
|
|
|
|
80
|
my ($uri, $name) = $line =~ /^=> (\S+)\s+(.*)/; |
548
|
|
|
|
|
|
|
# add current date |
549
|
4
|
|
|
|
|
64
|
my ($sec, $min, $hour, $mday, $mon, $year) = gmtime(); # UTC |
550
|
4
|
|
|
|
|
47
|
my $date = sprintf('%4d-%02d-%02d', $year + 1900, $mon + 1, $mday); |
551
|
4
|
|
|
|
|
24
|
$line = "=> $uri $date $name"; |
552
|
|
|
|
|
|
|
# load file |
553
|
4
|
|
|
|
|
8
|
my @lines; |
554
|
4
|
100
|
|
|
|
89
|
if (-f $updated_list) { |
555
|
3
|
50
|
|
|
|
155
|
open(my $fh, "<:encoding(UTF-8)", $updated_list) |
556
|
|
|
|
|
|
|
or die "Cannot read $updated_list: $!\n"; |
557
|
3
|
|
|
|
|
440
|
@lines = convert(<$fh>); # from gohper |
558
|
3
|
|
|
|
|
60
|
chomp(@lines); |
559
|
|
|
|
|
|
|
} |
560
|
|
|
|
|
|
|
# start the new list with the non-list links |
561
|
4
|
|
|
|
|
34
|
my @new = grep(!/^=>/, @lines); |
562
|
|
|
|
|
|
|
# add the line to the new list |
563
|
4
|
|
|
|
|
10
|
push(@new, $line); |
564
|
|
|
|
|
|
|
# add the remaining links to the new list, except for the ones matching the name of the new line |
565
|
4
|
|
|
|
|
103
|
push(@new, grep(!/\d\d\d\d-\d\d-\d\d $name$/, grep(/^=>/, @lines))); |
566
|
|
|
|
|
|
|
# save |
567
|
4
|
|
|
|
|
90
|
save_file($updated_list, join("\n", @new, "")); |
568
|
|
|
|
|
|
|
} |
569
|
|
|
|
|
|
|
|
570
|
|
|
|
|
|
|
sub do_update { |
571
|
5
|
|
|
5
|
|
6935
|
my $quiet = grep { $_ eq '--quiet' } @_; |
|
0
|
|
|
|
|
0
|
|
572
|
5
|
|
|
|
|
17
|
my @sites = grep { $_ ne '--quiet' } @_; |
|
0
|
|
|
|
|
0
|
|
573
|
5
|
|
|
|
|
22
|
my $site = load_site(); |
574
|
5
|
|
|
|
|
33
|
my %responses; |
575
|
|
|
|
|
|
|
my @uris; |
576
|
5
|
|
|
|
|
0
|
my %names; |
577
|
5
|
|
|
|
|
0
|
my %lines; |
578
|
5
|
|
|
|
|
35
|
for my $line (@$site) { |
579
|
10
|
|
|
|
|
1321
|
my ($uri, $name) = $line =~ /^=> (\S+)(?:[ \t]+(.*))?/; |
580
|
10
|
|
33
|
|
|
29
|
$name ||= $uri; |
581
|
10
|
50
|
33
|
|
|
38
|
next unless @sites == 0 or grep { $_ eq $name } @sites; |
|
0
|
|
|
|
|
0
|
|
582
|
10
|
50
|
|
|
|
419
|
say("Fetching $name...") unless $quiet; |
583
|
10
|
|
|
|
|
55
|
push(@uris, $uri); |
584
|
10
|
|
|
|
|
37
|
$names{$uri} = $name; |
585
|
10
|
|
|
|
|
25
|
$lines{$uri} = $line; |
586
|
10
|
50
|
|
|
|
51
|
if ($uri =~ /^gopher/) { |
|
|
0
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
587
|
10
|
|
|
|
|
30
|
query_gopher($uri, \%responses); |
588
|
|
|
|
|
|
|
} elsif ($uri =~ /^gemini/) { |
589
|
0
|
|
|
|
|
0
|
query_gemini($uri, \%responses); |
590
|
|
|
|
|
|
|
} elsif ($uri =~ /^http/) { |
591
|
0
|
|
|
|
|
0
|
query_web($uri, \%responses); |
592
|
|
|
|
|
|
|
} else { |
593
|
0
|
|
|
|
|
0
|
warn "Don't know how to fetch $uri\n"; |
594
|
|
|
|
|
|
|
} |
595
|
|
|
|
|
|
|
} |
596
|
|
|
|
|
|
|
|
597
|
5
|
50
|
|
|
|
1650
|
Mojo::IOLoop->start unless Mojo::IOLoop->is_running; |
598
|
|
|
|
|
|
|
|
599
|
5
|
|
|
|
|
2024
|
for my $uri (keys %responses) { |
600
|
10
|
|
|
|
|
47
|
my $name = $names{$uri}; |
601
|
|
|
|
|
|
|
# decode the UTF-8 when we have the entire response |
602
|
10
|
|
|
|
|
64
|
my $new = decode_utf8($responses{$uri}); |
603
|
10
|
50
|
|
|
|
352
|
if (not $new) { |
604
|
0
|
|
|
|
|
0
|
warn("$name returned an empty document\n"); |
605
|
0
|
|
|
|
|
0
|
next; |
606
|
|
|
|
|
|
|
} |
607
|
10
|
|
|
|
|
63
|
my $filename = $uri; |
608
|
10
|
|
|
|
|
131
|
$filename =~ s/[\/:]/-/g; |
609
|
10
|
|
|
|
|
42
|
my $cache = "$data_dir/$filename"; |
610
|
10
|
100
|
|
|
|
55
|
if ($new =~ /^<(\?xml|rss)/) { |
611
|
2
|
|
|
|
|
10
|
$new = to_gemini($new); |
612
|
2
|
|
|
|
|
145
|
my $encoded = uri_escape_utf8($filename); |
613
|
2
|
|
|
|
|
105
|
$lines{$uri} = "=> $encoded $name"; # now referring to the local cache file |
614
|
|
|
|
|
|
|
} |
615
|
10
|
|
|
|
|
34
|
my $old = load_file($cache); |
616
|
10
|
100
|
|
|
|
272
|
if ($new ne $old) { |
617
|
4
|
50
|
|
|
|
214
|
say "$name updated" unless $quiet; |
618
|
4
|
|
|
|
|
42
|
add_update($lines{$uri}); |
619
|
4
|
|
|
|
|
20
|
save_file($cache, $new); |
620
|
|
|
|
|
|
|
} else { |
621
|
6
|
50
|
|
|
|
284
|
say "$name unchanged" unless $quiet; |
622
|
|
|
|
|
|
|
} |
623
|
|
|
|
|
|
|
} |
624
|
|
|
|
|
|
|
} |
625
|
|
|
|
|
|
|
|
626
|
|
|
|
|
|
|
=head2 Subscribing to feeds |
627
|
|
|
|
|
|
|
|
628
|
|
|
|
|
|
|
When the result of an update is an XML document, then it is parsed and the links |
629
|
|
|
|
|
|
|
of its items (if RSS) or entries (if Atom) are extracted and saved in the cache |
630
|
|
|
|
|
|
|
file in the data directory. The effect is this: |
631
|
|
|
|
|
|
|
|
632
|
|
|
|
|
|
|
=over |
633
|
|
|
|
|
|
|
|
634
|
|
|
|
|
|
|
=item If you subscribe to a regular page, then the link to it in C |
635
|
|
|
|
|
|
|
moves to the top when it changes. |
636
|
|
|
|
|
|
|
|
637
|
|
|
|
|
|
|
=item If you subscribe to a feed, then the link in C moves to the |
638
|
|
|
|
|
|
|
top when it changes and it links to a file in the data directory that links to |
639
|
|
|
|
|
|
|
the individual items in the feed. |
640
|
|
|
|
|
|
|
|
641
|
|
|
|
|
|
|
=back |
642
|
|
|
|
|
|
|
|
643
|
|
|
|
|
|
|
Example: |
644
|
|
|
|
|
|
|
|
645
|
|
|
|
|
|
|
moku-pona add https://campaignwiki.org/rpg/feed.xml "RPG" |
646
|
|
|
|
|
|
|
moku-pona update |
647
|
|
|
|
|
|
|
|
648
|
|
|
|
|
|
|
This adds the RPG entry to C as follows: |
649
|
|
|
|
|
|
|
|
650
|
|
|
|
|
|
|
=> https%3A--campaignwiki.org-rpg-feed.xml 2020-11-07 RPG |
651
|
|
|
|
|
|
|
|
652
|
|
|
|
|
|
|
And if you check the file C, you'll see |
653
|
|
|
|
|
|
|
that it's a regular Gemini list. You'll find 100 links like the following: |
654
|
|
|
|
|
|
|
|
655
|
|
|
|
|
|
|
=> https://alexschroeder.ch/wiki/2020-11-05_Episode_34 Episode 34 |
656
|
|
|
|
|
|
|
|
657
|
|
|
|
|
|
|
Now use C (see below) to move the files to the correct |
658
|
|
|
|
|
|
|
directory where your Gemini server expects them. |
659
|
|
|
|
|
|
|
|
660
|
|
|
|
|
|
|
=cut |
661
|
|
|
|
|
|
|
|
662
|
|
|
|
|
|
|
# Convert a RSS or Atom feed to Gemini links |
663
|
|
|
|
|
|
|
sub to_gemini { |
664
|
5
|
|
|
5
|
|
86
|
my $xml = shift; |
665
|
5
|
|
|
|
|
8
|
my $dom = eval { |
666
|
5
|
|
|
|
|
55
|
require XML::LibXML; |
667
|
5
|
|
|
|
|
45
|
my $parser = XML::LibXML->new(recover => 2); # no errors, no warnings |
668
|
5
|
|
|
|
|
480
|
$parser->load_xml(string => $xml); |
669
|
|
|
|
|
|
|
}; |
670
|
5
|
50
|
|
|
|
2614
|
if ($@) { |
671
|
0
|
|
|
|
|
0
|
warn "$@\n"; |
672
|
0
|
|
|
|
|
0
|
return ''; |
673
|
|
|
|
|
|
|
} |
674
|
5
|
|
|
|
|
49
|
my $root = $dom->documentElement(); |
675
|
5
|
|
|
|
|
210
|
my $xpc = XML::LibXML::XPathContext->new; |
676
|
5
|
|
|
|
|
38
|
$xpc->registerNs('atom', 'http://www.w3.org/2005/Atom'); |
677
|
5
|
|
66
|
|
|
32
|
my $nodes = $xpc->findnodes('//atom:entry', $root) || $root->findnodes('//item'); |
678
|
5
|
|
|
|
|
1366
|
my @lines; |
679
|
5
|
|
|
|
|
26
|
for my $node ($nodes->get_nodelist) { |
680
|
7
|
|
66
|
|
|
73
|
my $titles = $xpc->findnodes('atom:title', $node) || $node->getChildrenByTagName('title'); |
681
|
7
|
|
|
|
|
1338
|
my $title = $titles->shift->textContent; # take the first |
682
|
7
|
|
|
|
|
102
|
$title =~ s/\s+$//; # trim right |
683
|
7
|
|
|
|
|
288
|
$title =~ s/^\s+//; # trim left |
684
|
7
|
|
100
|
|
|
23
|
my $links = $xpc->findnodes('atom:link', $node) || $node->getChildrenByTagName('link'); |
685
|
7
|
100
|
|
|
|
1003
|
next unless $links; |
686
|
6
|
|
|
|
|
138
|
my $link = $links->shift; # take the first |
687
|
6
|
|
66
|
|
|
37
|
my $href = $link->getAttribute('href') || $link->textContent; |
688
|
6
|
|
|
|
|
128
|
push(@lines, "=> $href $title"); |
689
|
|
|
|
|
|
|
} |
690
|
5
|
|
|
|
|
94
|
return join("\n", @lines, ""); |
691
|
|
|
|
|
|
|
} |
692
|
|
|
|
|
|
|
|
693
|
|
|
|
|
|
|
=head2 Publishing your subscription |
694
|
|
|
|
|
|
|
|
695
|
|
|
|
|
|
|
moku-pona publish |
696
|
|
|
|
|
|
|
|
697
|
|
|
|
|
|
|
This takes the important files from your data directory and copies them to a |
698
|
|
|
|
|
|
|
target directory. You could just use symbolic links for C and |
699
|
|
|
|
|
|
|
C, of course. But if you've subscribed to actual feeds as described |
700
|
|
|
|
|
|
|
above, then the cache files need to get copied as well! |
701
|
|
|
|
|
|
|
|
702
|
|
|
|
|
|
|
Example: |
703
|
|
|
|
|
|
|
|
704
|
|
|
|
|
|
|
mkdir ~/subs |
705
|
|
|
|
|
|
|
moku-pona publish ~/subs |
706
|
|
|
|
|
|
|
|
707
|
|
|
|
|
|
|
=head2 Serving your subscriptions via Gemini |
708
|
|
|
|
|
|
|
|
709
|
|
|
|
|
|
|
This depends entirely on your Gemini server. If you like it really simple, you |
710
|
|
|
|
|
|
|
can use L. Here's how to create the certificate and key |
711
|
|
|
|
|
|
|
files, copy them to the C<~/subs> directory created above, and run C |
712
|
|
|
|
|
|
|
for a quick test. |
713
|
|
|
|
|
|
|
|
714
|
|
|
|
|
|
|
make cert |
715
|
|
|
|
|
|
|
cp *.pem ~/subs |
716
|
|
|
|
|
|
|
cd ~/subs |
717
|
|
|
|
|
|
|
lupa-pona |
718
|
|
|
|
|
|
|
|
719
|
|
|
|
|
|
|
=cut |
720
|
|
|
|
|
|
|
|
721
|
|
|
|
|
|
|
sub do_publish { |
722
|
1
|
|
|
1
|
|
473
|
my $target = shift; |
723
|
1
|
50
|
|
|
|
18
|
die "Target $target is not a directory\n" unless -d $target; |
724
|
1
|
50
|
|
|
|
14
|
die "Source $site_list does not exist\n" unless -f $site_list; |
725
|
1
|
50
|
|
|
|
14
|
die "Source $updated_list does not exist\n" unless -f $updated_list; |
726
|
1
|
|
|
|
|
2
|
my $path; |
727
|
|
|
|
|
|
|
# copy site list |
728
|
1
|
|
|
|
|
8
|
copy($site_list, "$target/sites.txt"); |
729
|
|
|
|
|
|
|
# copy updates but with local links for the feed files |
730
|
1
|
50
|
|
2
|
|
365
|
open(my $in, "<:encoding(UTF-8)", $updated_list) |
|
2
|
|
|
|
|
21
|
|
|
2
|
|
|
|
|
4
|
|
|
2
|
|
|
|
|
34
|
|
731
|
|
|
|
|
|
|
or die "Cannot read $updated_list: $!\n"; |
732
|
1
|
50
|
|
|
|
1241
|
open(my $out, ">:encoding(UTF-8)", "$target/updates.txt") |
733
|
|
|
|
|
|
|
or die "Cannot write $target/updates.txt: $!\n"; |
734
|
1
|
|
|
|
|
90
|
for my $line (<$in>) { |
735
|
1
|
|
|
|
|
24
|
chomp($line); |
736
|
1
|
|
|
|
|
3
|
($line) = convert($line); |
737
|
1
|
|
|
|
|
8
|
my ($uri, $name) = $line =~ /^=> file:\/\/\/(\S+)\s+(.*)/; |
738
|
|
|
|
|
|
|
# if the target is a local file, then that's because it is the result of a |
739
|
|
|
|
|
|
|
# to_gemini call in do_update, so we need to copy it as well |
740
|
1
|
|
|
|
|
3
|
$uri =~ s/[\/:]/-/g; |
741
|
1
|
50
|
|
|
|
21
|
if (-f "$data_dir/$uri") { |
742
|
1
|
|
|
|
|
8
|
copy("$data_dir/$uri", "$target/$uri"); |
743
|
|
|
|
|
|
|
} |
744
|
1
|
|
|
|
|
379
|
print $out "$line\n"; |
745
|
|
|
|
|
|
|
} |
746
|
|
|
|
|
|
|
} |
747
|
|
|
|
|
|
|
|
748
|
|
|
|
|
|
|
sub do_help { |
749
|
0
|
|
|
0
|
|
0
|
my $parser = Pod::Text->new(); |
750
|
0
|
|
|
|
|
0
|
$parser->parse_file($0); |
751
|
|
|
|
|
|
|
} |
752
|
|
|
|
|
|
|
|
753
|
|
|
|
|
|
|
sub main { |
754
|
0
|
|
0
|
0
|
|
0
|
my $command = shift(@ARGV) || "help"; |
755
|
0
|
0
|
|
|
|
0
|
if ($command eq "add") { do_add(@ARGV) } |
|
0
|
0
|
|
|
|
0
|
|
|
|
0
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
|
|
0
|
|
|
|
|
|
756
|
0
|
|
|
|
|
0
|
elsif ($command eq "remove") { do_remove(@ARGV) } |
757
|
0
|
|
|
|
|
0
|
elsif ($command eq "list") { do_list() } |
758
|
0
|
|
|
|
|
0
|
elsif ($command eq "cleanup") { do_cleanup(@ARGV) } |
759
|
0
|
|
|
|
|
0
|
elsif ($command eq "update") { do_update(@ARGV) } |
760
|
0
|
|
|
|
|
0
|
elsif ($command eq "convert") { do_convert() } |
761
|
0
|
|
|
|
|
0
|
elsif ($command eq "publish") { do_publish(@ARGV) } |
762
|
0
|
|
|
|
|
0
|
else { do_help() } |
763
|
|
|
|
|
|
|
} |
764
|
|
|
|
|
|
|
|
765
|
|
|
|
|
|
|
main() if $0 =~ /\bmoku-pona$/; |
766
|
|
|
|
|
|
|
|
767
|
|
|
|
|
|
|
1; |