| line | stmt | bran | cond | sub | pod | time | code | 
| 1 |  |  |  |  |  |  | #!/home/alex/perl5/perlbrew/perls/perl-5.32.0/bin/perl | 
| 2 |  |  |  |  |  |  | # Copyright (C) 2018–2021  Alex Schroeder | 
| 3 |  |  |  |  |  |  |  | 
| 4 |  |  |  |  |  |  | # This program is free software: you can redistribute it and/or modify it under | 
| 5 |  |  |  |  |  |  | # the terms of the GNU General Public License as published by the Free Software | 
| 6 |  |  |  |  |  |  | # Foundation, either version 3 of the License, or (at your option) any later | 
| 7 |  |  |  |  |  |  | # version. | 
| 8 |  |  |  |  |  |  | # | 
| 9 |  |  |  |  |  |  | # This program is distributed in the hope that it will be useful, but WITHOUT | 
| 10 |  |  |  |  |  |  | # ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS | 
| 11 |  |  |  |  |  |  | # FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. | 
| 12 |  |  |  |  |  |  | # | 
| 13 |  |  |  |  |  |  | # You should have received a copy of the GNU General Public License along with | 
| 14 |  |  |  |  |  |  | # this program. If not, see . | 
| 15 |  |  |  |  |  |  |  | 
| 16 |  |  |  |  |  |  | =encoding utf8 | 
| 17 |  |  |  |  |  |  |  | 
| 18 |  |  |  |  |  |  | =head1 Moku Pona | 
| 19 |  |  |  |  |  |  |  | 
| 20 |  |  |  |  |  |  | Moku Pona is a Gemini based feed reader. It can monitor URLs to feeds or regular | 
| 21 |  |  |  |  |  |  | pages for changes and keeps and updated list of these in a Gemini list. Moku | 
| 22 |  |  |  |  |  |  | Pona knows how to fetch Gopher URLs, Gemini URLs, and regular web URLs. | 
| 23 |  |  |  |  |  |  |  | 
| 24 |  |  |  |  |  |  | You manage your subscriptions using the command-line, with Moku Pona. | 
| 25 |  |  |  |  |  |  |  | 
| 26 |  |  |  |  |  |  | You serve the resulting file using a Gemini server. | 
| 27 |  |  |  |  |  |  |  | 
| 28 |  |  |  |  |  |  | You read it all using your Gemini client. | 
| 29 |  |  |  |  |  |  |  | 
| 30 |  |  |  |  |  |  | =head2 Limitations | 
| 31 |  |  |  |  |  |  |  | 
| 32 |  |  |  |  |  |  | When Moku Pona isn't watching a feed it can only detect changes on a page. Thus, | 
| 33 |  |  |  |  |  |  | if there is an item that points to a phlog or blog, that's great. Sometimes | 
| 34 |  |  |  |  |  |  | people put their phlog in a folder per year. If the Gopher menu lists each | 
| 35 |  |  |  |  |  |  | folder and a date with the latest change, then that's great, you can use it. | 
| 36 |  |  |  |  |  |  | Without it, you're in trouble: you need to subscribe to the item for the current | 
| 37 |  |  |  |  |  |  | year in order to see changes, but when the next year comes around, you're | 
| 38 |  |  |  |  |  |  | subscribed to the wrong item. Sometimes you're lucky and there will be a menu | 
| 39 |  |  |  |  |  |  | somewhere with a timestamp for the last change. Use that instead. Good luck! | 
| 40 |  |  |  |  |  |  |  | 
| 41 |  |  |  |  |  |  | =head2 License | 
| 42 |  |  |  |  |  |  |  | 
| 43 |  |  |  |  |  |  | GNU Affero General Public License | 
| 44 |  |  |  |  |  |  |  | 
| 45 |  |  |  |  |  |  | =head2 Installation | 
| 46 |  |  |  |  |  |  |  | 
| 47 |  |  |  |  |  |  | Using C: | 
| 48 |  |  |  |  |  |  |  | 
| 49 |  |  |  |  |  |  | cpan App::mokupona | 
| 50 |  |  |  |  |  |  |  | 
| 51 |  |  |  |  |  |  | Manual install: | 
| 52 |  |  |  |  |  |  |  | 
| 53 |  |  |  |  |  |  | perl Makefile.PL | 
| 54 |  |  |  |  |  |  | make | 
| 55 |  |  |  |  |  |  | make install | 
| 56 |  |  |  |  |  |  |  | 
| 57 |  |  |  |  |  |  | =head2 Dependencies | 
| 58 |  |  |  |  |  |  |  | 
| 59 |  |  |  |  |  |  | There are some Perl dependencies you need to satisfy in order to run this | 
| 60 |  |  |  |  |  |  | program: | 
| 61 |  |  |  |  |  |  |  | 
| 62 |  |  |  |  |  |  | =over | 
| 63 |  |  |  |  |  |  |  | 
| 64 |  |  |  |  |  |  | =item L, or C | 
| 65 |  |  |  |  |  |  |  | 
| 66 |  |  |  |  |  |  | =item L, or C | 
| 67 |  |  |  |  |  |  |  | 
| 68 |  |  |  |  |  |  | =item L, or C | 
| 69 |  |  |  |  |  |  |  | 
| 70 |  |  |  |  |  |  | =item L, or C | 
| 71 |  |  |  |  |  |  |  | 
| 72 |  |  |  |  |  |  | =item L, or c | 
| 73 |  |  |  |  |  |  |  | 
| 74 |  |  |  |  |  |  | =back | 
| 75 |  |  |  |  |  |  |  | 
| 76 |  |  |  |  |  |  | =cut | 
| 77 |  |  |  |  |  |  |  | 
| 78 | 9 |  |  | 9 |  | 11043 | use Modern::Perl '2018'; | 
|  | 9 |  |  |  |  | 38961 |  | 
|  | 9 |  |  |  |  | 94 |  | 
| 79 | 9 |  |  | 9 |  | 7149 | use File::Copy qw(copy); | 
|  | 9 |  |  |  |  | 19686 |  | 
|  | 9 |  |  |  |  | 1094 |  | 
| 80 | 9 |  |  | 9 |  | 5034 | use Encode::Locale qw(decode_argv); | 
|  | 9 |  |  |  |  | 98874 |  | 
|  | 9 |  |  |  |  | 633 |  | 
| 81 | 9 |  |  | 9 |  | 68 | use Encode qw(decode_utf8); | 
|  | 9 |  |  |  |  | 34 |  | 
|  | 9 |  |  |  |  | 553 |  | 
| 82 | 9 |  |  | 9 |  | 4031 | use Mojo::IOLoop; | 
|  | 9 |  |  |  |  | 2816292 |  | 
|  | 9 |  |  |  |  | 116 |  | 
| 83 | 9 |  |  | 9 |  | 6908 | use XML::LibXML; | 
|  | 9 |  |  |  |  | 294034 |  | 
|  | 9 |  |  |  |  | 61 |  | 
| 84 | 9 |  |  | 9 |  | 5799 | use URI::Escape; | 
|  | 9 |  |  |  |  | 14820 |  | 
|  | 9 |  |  |  |  | 923 |  | 
| 85 | 9 |  |  | 9 |  | 84 | use List::Util qw(none); | 
|  | 9 |  |  |  |  | 19 |  | 
|  | 9 |  |  |  |  | 52047 |  | 
| 86 |  |  |  |  |  |  |  | 
| 87 |  |  |  |  |  |  | decode_argv(); | 
| 88 |  |  |  |  |  |  |  | 
| 89 |  |  |  |  |  |  | if (-t) { | 
| 90 |  |  |  |  |  |  | binmode(STDIN, ":encoding(console_in)"); | 
| 91 |  |  |  |  |  |  | binmode(STDOUT, ":encoding(console_out)"); | 
| 92 |  |  |  |  |  |  | binmode(STDERR, ":encoding(console_out)"); | 
| 93 |  |  |  |  |  |  | } | 
| 94 |  |  |  |  |  |  |  | 
| 95 |  |  |  |  |  |  | =head2 The Data Directory | 
| 96 |  |  |  |  |  |  |  | 
| 97 |  |  |  |  |  |  | Moku Pona keeps the list of URLs you are subscribed to in directory. It's | 
| 98 |  |  |  |  |  |  | probably C<~/.moku-pona> on your system. | 
| 99 |  |  |  |  |  |  |  | 
| 100 |  |  |  |  |  |  | =over | 
| 101 |  |  |  |  |  |  |  | 
| 102 |  |  |  |  |  |  | =item If you have the C environment variable set, then that's your data | 
| 103 |  |  |  |  |  |  | directory. | 
| 104 |  |  |  |  |  |  |  | 
| 105 |  |  |  |  |  |  | =item If you have the C environment variable set, then your data | 
| 106 |  |  |  |  |  |  | directory is F<$XDG_DATA_HOME/moku-pona>. | 
| 107 |  |  |  |  |  |  |  | 
| 108 |  |  |  |  |  |  | =item If you you have the C environment variable set, and you have a | 
| 109 |  |  |  |  |  |  | F<$HOME/.local> directory, then your data directory is | 
| 110 |  |  |  |  |  |  | F<$HOME/.local/moku-pona>. | 
| 111 |  |  |  |  |  |  |  | 
| 112 |  |  |  |  |  |  | =item If you have the C environment variable set, then your data directory | 
| 113 |  |  |  |  |  |  | is F<$HOME/.moku-pona>. | 
| 114 |  |  |  |  |  |  |  | 
| 115 |  |  |  |  |  |  | =item If you have the C environment variable set (Windows), then your | 
| 116 |  |  |  |  |  |  | data directory is F<$APPDATA/moku-pona>. | 
| 117 |  |  |  |  |  |  |  | 
| 118 |  |  |  |  |  |  | =item The last option is to have the C environment variable set. | 
| 119 |  |  |  |  |  |  |  | 
| 120 |  |  |  |  |  |  | =back | 
| 121 |  |  |  |  |  |  |  | 
| 122 |  |  |  |  |  |  | The data directory contains a copy of the latest resources. The names of these | 
| 123 |  |  |  |  |  |  | cache files are simply the URL with all the slashes replaced by a hyphen. | 
| 124 |  |  |  |  |  |  |  | 
| 125 |  |  |  |  |  |  | =cut | 
| 126 |  |  |  |  |  |  |  | 
| 127 |  |  |  |  |  |  | our $data_dir = data_dir(); | 
| 128 |  |  |  |  |  |  | # say "Using $data_dir"; | 
| 129 |  |  |  |  |  |  |  | 
| 130 |  |  |  |  |  |  | sub data_dir { | 
| 131 | 9 | 50 |  | 9 |  | 42 | return $ENV{MOKU_PONA} if $ENV{MOKU_PONA}; | 
| 132 |  |  |  |  |  |  | # find an existing directory | 
| 133 | 9 | 50 | 33 |  |  | 40 | return $ENV{XDG_DATA_HOME} . '/moku-pona' if $ENV{XDG_DATA_HOME} and -d $ENV{XDG_DATA_HOME} . '/moku-pona'; | 
| 134 | 9 | 50 | 33 |  |  | 189 | return $ENV{HOME} . '/.local/moku-pona' if $ENV{HOME} and -d $ENV{HOME} . '/.local/moku-pona'; | 
| 135 | 9 | 50 | 33 |  |  | 182 | return $ENV{HOME} . '/.moku-pona' if $ENV{HOME} and -d $ENV{HOME} . '/.moku-pona'; | 
| 136 | 9 | 50 | 33 |  |  | 80 | return $ENV{APPDATA} . '/moku-pona' if $ENV{APPDATA} and -d $ENV{APPDATA} . '/.moku-pona'; | 
| 137 | 9 | 50 | 33 |  |  | 38 | return $ENV{LOGDIR} . '/.moku-pona' if $ENV{LOGDIR} and -d $ENV{LOGDIR} . '/.moku-pona'; | 
| 138 |  |  |  |  |  |  | # or use a new one | 
| 139 | 9 | 50 |  |  |  | 30 | return $ENV{XDG_DATA_HOME} . '/moku-pona' if $ENV{XDG_DATA_HOME}; | 
| 140 | 9 | 50 | 33 |  |  | 108 | return $ENV{HOME} . '/.local/moku-pona' if $ENV{HOME} and -d $ENV{HOME} . '/.local'; | 
| 141 | 9 | 50 |  |  |  | 71 | return $ENV{HOME} . '/.moku-pona' if $ENV{HOME}; | 
| 142 | 0 | 0 |  |  |  | 0 | return $ENV{APPDATA} . '/moku-pona' if $ENV{APPDATA}; | 
| 143 | 0 | 0 |  |  |  | 0 | return $ENV{LOGDIR} . '/.moku-pona' if $ENV{LOGDIR}; | 
| 144 | 0 |  |  |  |  | 0 | die "Please set the MOKU_PONA environment variable to a directory name\n"; | 
| 145 |  |  |  |  |  |  | } | 
| 146 |  |  |  |  |  |  |  | 
| 147 |  |  |  |  |  |  | =pod | 
| 148 |  |  |  |  |  |  |  | 
| 149 |  |  |  |  |  |  | The C file is a file containing a gemtext list of links, i.e. entries | 
| 150 |  |  |  |  |  |  | such as these: | 
| 151 |  |  |  |  |  |  |  | 
| 152 |  |  |  |  |  |  | => gemini://alexschroeder.ch Alex Schroeder | 
| 153 |  |  |  |  |  |  |  | 
| 154 |  |  |  |  |  |  | =cut | 
| 155 |  |  |  |  |  |  |  | 
| 156 |  |  |  |  |  |  | our $site_list = $data_dir . '/sites.txt'; | 
| 157 |  |  |  |  |  |  |  | 
| 158 |  |  |  |  |  |  | =pod | 
| 159 |  |  |  |  |  |  |  | 
| 160 |  |  |  |  |  |  | The C file is a file containing a gemtext list of links based on | 
| 161 |  |  |  |  |  |  | C, but with a timestamp of their last change, and with new updates | 
| 162 |  |  |  |  |  |  | moved to the top. The ISO date is simply inserted after the URL: | 
| 163 |  |  |  |  |  |  |  | 
| 164 |  |  |  |  |  |  | => gemini://alexschroeder.ch 2020-11-07 Alex Schroeder | 
| 165 |  |  |  |  |  |  |  | 
| 166 |  |  |  |  |  |  | =cut | 
| 167 |  |  |  |  |  |  |  | 
| 168 |  |  |  |  |  |  | our $updated_list = $data_dir . '/updates.txt'; | 
| 169 |  |  |  |  |  |  |  | 
| 170 |  |  |  |  |  |  | =pod | 
| 171 |  |  |  |  |  |  |  | 
| 172 |  |  |  |  |  |  | In order to be at least somewhat backwards compatible with Moku Pona versions | 
| 173 |  |  |  |  |  |  | 1.1 and earlier, C may contain Gopher menu items. These are converted | 
| 174 |  |  |  |  |  |  | to Gemini URLs during processing and thus the C file still contains | 
| 175 |  |  |  |  |  |  | regular gemtext. | 
| 176 |  |  |  |  |  |  |  | 
| 177 |  |  |  |  |  |  | 1Alex Schroeder ⭾ ⭾ alexschroeder.ch ⭾ 70 | 
| 178 |  |  |  |  |  |  |  | 
| 179 |  |  |  |  |  |  | =cut | 
| 180 |  |  |  |  |  |  |  | 
| 181 |  |  |  |  |  |  | sub convert { | 
| 182 | 27 |  |  | 27 |  | 138 | for (@_) { | 
| 183 | 49 | 100 |  |  |  | 236 | next if /^=> /; # is already a gemini link | 
| 184 | 17 |  |  |  |  | 127 | my ($type, $desc, $selector, $host, $port) = /^([^\t])([^\t]*)\t([^\t]*)\t([^\t]*)\t([^\t\r]*)/; | 
| 185 | 17 | 100 | 66 |  |  | 70 | if ($host and $port) { | 
| 186 | 13 |  | 50 |  |  | 27 | $port //= 0; | 
| 187 | 13 |  |  |  |  | 55 | $_ = "=> gopher://$host:$port/$type$selector $desc"; | 
| 188 |  |  |  |  |  |  | } | 
| 189 |  |  |  |  |  |  | } | 
| 190 | 27 |  |  |  |  | 155 | return @_; | 
| 191 |  |  |  |  |  |  | } | 
| 192 |  |  |  |  |  |  |  | 
| 193 |  |  |  |  |  |  | =pod | 
| 194 |  |  |  |  |  |  |  | 
| 195 |  |  |  |  |  |  | As was said above, however, the recommended format is the use of URLs. Moku Pona | 
| 196 |  |  |  |  |  |  | supports Gemini, Gopher, and the web (gemini, gopher, gophers, http, and https | 
| 197 |  |  |  |  |  |  | schemes). | 
| 198 |  |  |  |  |  |  |  | 
| 199 |  |  |  |  |  |  | =cut | 
| 200 |  |  |  |  |  |  |  | 
| 201 |  |  |  |  |  |  | sub query_gemini { | 
| 202 | 0 |  |  | 0 |  | 0 | my $url = shift; | 
| 203 | 0 |  |  |  |  | 0 | my $responses = shift; | 
| 204 | 0 |  |  |  |  | 0 | my($scheme, $authority, $path, $query, $fragment) = | 
| 205 |  |  |  |  |  |  | $url =~ m|(?:([^:/?#]+):)?(?://([^/?#]*))?([^?#]*)(?:\?([^#]*))?(?:#(.*))?|; | 
| 206 | 0 | 0 | 0 |  |  | 0 | die "⚠ The URL '$url' must use the gemini scheme\n" unless $scheme and $scheme eq 'gemini'; | 
| 207 | 0 | 0 |  |  |  | 0 | die "⚠ The URL '$url' must have an authority\n" unless $authority; | 
| 208 | 0 |  |  |  |  | 0 | my ($host, $port) = split(/:/, $authority, 2); | 
| 209 | 0 |  | 0 |  |  | 0 | $port //= 1965; | 
| 210 |  |  |  |  |  |  | Mojo::IOLoop->client( | 
| 211 |  |  |  |  |  |  | {port => $port, address => $host, tls => 1, | 
| 212 |  |  |  |  |  |  | tls_options => { SSL_verify_mode => 0x00 }, timeout => 120} | 
| 213 |  |  |  |  |  |  | => sub { | 
| 214 | 0 |  |  | 0 |  | 0 | my ($loop, $err, $stream) = @_; | 
| 215 | 0 | 0 |  |  |  | 0 | if ($err) { | 
| 216 | 0 |  |  |  |  | 0 | warn "Cannot connect to $url: $err\n"; | 
| 217 | 0 |  |  |  |  | 0 | return; | 
| 218 |  |  |  |  |  |  | } | 
| 219 | 0 |  |  |  |  | 0 | $stream->timeout(300); # 5 min | 
| 220 | 0 |  |  |  |  | 0 | my $header; | 
| 221 |  |  |  |  |  |  | $stream->on(read => sub { | 
| 222 | 0 |  |  |  |  | 0 | my ($stream, $bytes) = @_; | 
| 223 | 0 |  |  |  |  | 0 | $responses->{$url} .= $bytes; | 
| 224 | 0 | 0 |  |  |  | 0 | $header = $responses->{$url} =~ s/^.*\r\n// unless $header}); | 
|  | 0 |  |  |  |  | 0 |  | 
| 225 | 0 |  |  |  |  | 0 | $stream->write("$url\r\n")}) | 
| 226 | 0 |  |  |  |  | 0 | } | 
| 227 |  |  |  |  |  |  |  | 
| 228 |  |  |  |  |  |  | sub query_gopher { | 
| 229 | 12 |  |  | 12 |  | 15 | my $url = shift; | 
| 230 | 12 |  |  |  |  | 18 | my $responses = shift; | 
| 231 | 12 |  |  |  |  | 17 | my ($selector, $host, $port) = url_to_gopher($url); | 
| 232 | 12 |  |  |  |  | 30 | my $tls = $url =~ /^gophers/; | 
| 233 |  |  |  |  |  |  | Mojo::IOLoop->client( | 
| 234 |  |  |  |  |  |  | {port => $port, address => $host, tls => $tls, timeout => 120 } | 
| 235 |  |  |  |  |  |  | => sub { | 
| 236 | 12 |  |  | 12 |  | 13901 | my ($loop, $err, $stream) = @_; | 
| 237 | 12 | 50 |  |  |  | 32 | if ($err) { | 
| 238 | 0 |  |  |  |  | 0 | warn "Cannot connect to $url: $err\n"; | 
| 239 | 0 |  |  |  |  | 0 | return; | 
| 240 |  |  |  |  |  |  | } | 
| 241 | 12 |  |  |  |  | 31 | $stream->timeout(300); # 5 min | 
| 242 |  |  |  |  |  |  | $stream->on( | 
| 243 |  |  |  |  |  |  | read => sub { | 
| 244 | 12 |  |  |  |  | 11087 | my ($stream, $bytes) = @_; | 
| 245 | 12 |  |  |  |  | 406 | $responses->{$url} .= $bytes}); | 
|  | 12 |  |  |  |  | 54 |  | 
| 246 | 12 |  |  |  |  | 105 | $stream->write("$selector\r\n")}) | 
| 247 | 12 |  |  |  |  | 155 | } | 
| 248 |  |  |  |  |  |  |  | 
| 249 |  |  |  |  |  |  | sub url_to_gopher { | 
| 250 | 12 |  |  | 12 |  | 15 | my $url = shift; | 
| 251 | 12 |  | 33 |  |  | 42 | my $name = shift||$url; | 
| 252 | 12 |  |  |  |  | 76 | my($scheme, $authority, $path, $query, $fragment) = | 
| 253 |  |  |  |  |  |  | $url =~ m|(?:([^:/?#]+):)?(?://([^/?#]*))?([^?#]*)(?:\?([^#]*))?(?:#(.*))?|; | 
| 254 | 12 |  | 50 |  |  | 26 | $scheme ||= "gopher"; | 
| 255 | 12 | 50 |  |  |  | 39 | return unless $scheme =~ /^gophers?$/; | 
| 256 | 12 | 50 |  |  |  | 54 | $path = substr($path, 1) if substr($path, 0, 1) eq "/"; | 
| 257 | 12 | 50 |  |  |  | 29 | my $type = $path ? substr($path, 0, 1) : "1"; | 
| 258 | 12 | 50 |  |  |  | 33 | my $selector = $path ? substr($path, 1) : ""; | 
| 259 | 12 |  |  |  |  | 45 | my ($host, $port) = split(/:/, $authority, 2); | 
| 260 | 12 |  | 50 |  |  | 51 | return ($selector, $host, $port||70); | 
| 261 |  |  |  |  |  |  | } | 
| 262 |  |  |  |  |  |  |  | 
| 263 |  |  |  |  |  |  | sub query_web { | 
| 264 | 0 |  |  | 0 |  | 0 | my $url = shift; | 
| 265 | 0 |  |  |  |  | 0 | my $responses = shift; | 
| 266 | 0 |  |  |  |  | 0 | my($scheme, $authority, $path, $query, $fragment) = | 
| 267 |  |  |  |  |  |  | $url =~ m|(?:([^:/?#]+):)?(?://([^/?#]*))?([^?#]*)(?:\?([^#]*))?(?:#(.*))?|; | 
| 268 | 0 |  |  |  |  | 0 | my $tls = $scheme eq "https"; | 
| 269 | 0 |  |  |  |  | 0 | my ($host, $port) = split(/:/, $authority, 2); | 
| 270 | 0 |  | 0 |  |  | 0 | $path ||= "/"; | 
| 271 | 0 |  |  |  |  | 0 | my $selector = $path; | 
| 272 | 0 | 0 |  |  |  | 0 | $selector .= "?" . $query if $query; | 
| 273 |  |  |  |  |  |  | # ignore the fragment | 
| 274 | 0 | 0 | 0 |  |  | 0 | $port ||= $tls ? 443 : 80; | 
| 275 |  |  |  |  |  |  | Mojo::IOLoop->client( | 
| 276 |  |  |  |  |  |  | {port => $port, address => $host, tls => $tls, timeout => 120 } | 
| 277 |  |  |  |  |  |  | => sub { | 
| 278 | 0 |  |  | 0 |  | 0 | my ($loop, $err, $stream) = @_; | 
| 279 | 0 |  |  |  |  | 0 | my $header; | 
| 280 |  |  |  |  |  |  | $stream->on(read => sub { | 
| 281 | 0 |  |  |  |  | 0 | my ($stream, $bytes) = @_; | 
| 282 | 0 |  |  |  |  | 0 | $responses->{$url} .= $bytes; | 
| 283 | 0 | 0 |  |  |  | 0 | $header = $responses->{$url} =~ s/^.*\r\n\r\n//s unless $header}); | 
|  | 0 |  |  |  |  | 0 |  | 
| 284 | 0 |  |  |  |  | 0 | $stream->write("GET $selector HTTP/1.0\r\n" | 
| 285 |  |  |  |  |  |  | . "Host: $host:$port\r\n" | 
| 286 |  |  |  |  |  |  | . "User-Agent: moku-pona\r\n" | 
| 287 | 0 |  |  |  |  | 0 | . "\r\n")}); | 
| 288 |  |  |  |  |  |  | } | 
| 289 |  |  |  |  |  |  |  | 
| 290 |  |  |  |  |  |  | =head2 Migration from 1.1 | 
| 291 |  |  |  |  |  |  |  | 
| 292 |  |  |  |  |  |  | The best way to migrate your setup is probably to use the C  subcommand  | 
| 293 |  |  |  |  |  |  | explained later, and to recreate your list of subscriptions. Then your | 
| 294 |  |  |  |  |  |  | C file will use gemtext format. | 
| 295 |  |  |  |  |  |  |  | 
| 296 |  |  |  |  |  |  | moku-pona list | grep "moku-pona add" > commands | 
| 297 |  |  |  |  |  |  | mv ~/.moku-pona/sites.txt ~/.moku-pona/sites.txt~ | 
| 298 |  |  |  |  |  |  | sh commands | 
| 299 |  |  |  |  |  |  |  | 
| 300 |  |  |  |  |  |  | =cut | 
| 301 |  |  |  |  |  |  |  | 
| 302 |  |  |  |  |  |  | sub load_site { | 
| 303 | 30 |  |  | 30 |  | 1245 | my $file = $site_list; | 
| 304 | 30 | 100 |  |  |  | 496 | return [] if not -f $file; | 
| 305 | 22 | 50 |  |  |  | 796 | open(my $fh, "<:encoding(UTF-8)", $file) or die "Cannot read $file: $!\n"; | 
| 306 | 22 |  |  |  |  | 3165 | my @lines = <$fh>; | 
| 307 | 22 |  |  |  |  | 496 | chomp(@lines); | 
| 308 | 22 |  |  |  |  | 52 | @lines = grep(/^=> /, convert(@lines)); # from gopher | 
| 309 | 22 |  |  |  |  | 371 | return \@lines; | 
| 310 |  |  |  |  |  |  | } | 
| 311 |  |  |  |  |  |  |  | 
| 312 |  |  |  |  |  |  | sub load_file { | 
| 313 | 20 |  |  | 20 |  | 3035 | my $file = shift; | 
| 314 | 20 | 100 |  |  |  | 335 | return "" if not -f $file; | 
| 315 | 17 | 50 |  |  |  | 693 | open(my $fh, "<:encoding(UTF-8)", $file) | 
| 316 |  |  |  |  |  |  | or die "Cannot read $file: $!\n"; | 
| 317 | 17 |  |  |  |  | 1162 | local $/ = undef; | 
| 318 | 17 |  |  |  |  | 470 | return <$fh>; | 
| 319 |  |  |  |  |  |  | } | 
| 320 |  |  |  |  |  |  |  | 
| 321 |  |  |  |  |  |  | sub save_file { | 
| 322 | 24 |  |  | 24 |  | 2022 | my $file = shift; | 
| 323 | 24 |  |  |  |  | 31 | my $data = shift; | 
| 324 | 24 | 100 |  |  |  | 437 | mkdir $data_dir unless -d $data_dir; | 
| 325 | 4 | 50 |  | 4 |  | 27 | open(my $fh, ">:encoding(UTF-8)", $file) | 
|  | 4 |  |  |  |  | 7 |  | 
|  | 4 |  |  |  |  | 29 |  | 
|  | 24 |  |  |  |  | 1462 |  | 
| 326 |  |  |  |  |  |  | or die "Cannot write $file: $!\n"; | 
| 327 | 24 |  |  |  |  | 8596 | print $fh $data; | 
| 328 |  |  |  |  |  |  | } | 
| 329 |  |  |  |  |  |  |  | 
| 330 |  |  |  |  |  |  | =head2 List your subscriptions | 
| 331 |  |  |  |  |  |  |  | 
| 332 |  |  |  |  |  |  | moku-pona list [names...] | 
| 333 |  |  |  |  |  |  |  | 
| 334 |  |  |  |  |  |  | This lists all your current subscriptions in a format that is suitable for a | 
| 335 |  |  |  |  |  |  | shell script. Optionally, only list a subset of the lines. All lines are matched | 
| 336 |  |  |  |  |  |  | against the regular expressions you provide and are only listed if there is at | 
| 337 |  |  |  |  |  |  | least one match, if you provided any. | 
| 338 |  |  |  |  |  |  |  | 
| 339 |  |  |  |  |  |  | Example: | 
| 340 |  |  |  |  |  |  |  | 
| 341 |  |  |  |  |  |  | moku-pona list alex | 
| 342 |  |  |  |  |  |  |  | 
| 343 |  |  |  |  |  |  | In this particular case, since I'm testing my own server, the result would be: | 
| 344 |  |  |  |  |  |  |  | 
| 345 |  |  |  |  |  |  | moku-pona add https://alexschroeder.ch/wiki?action=rss "rss" | 
| 346 |  |  |  |  |  |  | moku-pona add gemini://alexschroeder.ch/ "gemini" | 
| 347 |  |  |  |  |  |  | moku-pona add gopher://alexschroeder.ch/ "gopher" | 
| 348 |  |  |  |  |  |  | moku-pona add gophers://alexschroeder.ch:7443/ "gophers" | 
| 349 |  |  |  |  |  |  |  | 
| 350 |  |  |  |  |  |  | =cut | 
| 351 |  |  |  |  |  |  |  | 
| 352 |  |  |  |  |  |  | sub do_list { | 
| 353 | 1 |  |  | 1 |  | 1508 | my @args = @_; | 
| 354 | 1 |  |  |  |  | 19 | my $site = load_site(); | 
| 355 | 1 | 50 |  |  |  | 11 | if (@args) { | 
| 356 | 0 |  |  |  |  | 0 | print("Subscribed items in $site_list matching @args:\n"); | 
| 357 |  |  |  |  |  |  | } else { | 
| 358 | 1 |  |  |  |  | 6 | print("Subscribed items in $site_list:\n"); | 
| 359 |  |  |  |  |  |  | } | 
| 360 | 1 | 50 |  |  |  | 3 | print("none\n") unless @$site; | 
| 361 | 1 |  |  |  |  | 3 | for my $line (@$site) { | 
| 362 | 2 | 50 | 33 | 0 |  | 6 | next if @args and none { $line =~ /$_/ } @args; | 
|  | 0 |  |  |  |  | 0 |  | 
| 363 |  |  |  |  |  |  | # skip item type | 
| 364 | 2 |  |  |  |  | 14 | my ($uri, $name) = $line =~ /^=> (\S+)\s+(.*)/; | 
| 365 | 2 |  |  |  |  | 10 | print(qq{moku-pona add $uri "$name"\n}); | 
| 366 |  |  |  |  |  |  | } | 
| 367 |  |  |  |  |  |  | } | 
| 368 |  |  |  |  |  |  |  | 
| 369 |  |  |  |  |  |  | =head2 Add a subscription | 
| 370 |  |  |  |  |  |  |  | 
| 371 |  |  |  |  |  |  | moku-pona add url [description] | 
| 372 |  |  |  |  |  |  |  | 
| 373 |  |  |  |  |  |  | This adds a URL to the list of subscribed items. If the target is an Atom or RSS | 
| 374 |  |  |  |  |  |  | feed, then that's also supported. You can provide an optional description for | 
| 375 |  |  |  |  |  |  | this URL. If you don't provide a description, the URL will be used as the item's | 
| 376 |  |  |  |  |  |  | description. | 
| 377 |  |  |  |  |  |  |  | 
| 378 |  |  |  |  |  |  | Example: | 
| 379 |  |  |  |  |  |  |  | 
| 380 |  |  |  |  |  |  | moku-pona add gemini://alexschroeder.ch kensanata | 
| 381 |  |  |  |  |  |  |  | 
| 382 |  |  |  |  |  |  | =cut | 
| 383 |  |  |  |  |  |  |  | 
| 384 |  |  |  |  |  |  | sub do_add { | 
| 385 | 8 |  |  | 8 |  | 1313 | my $uri = shift; | 
| 386 | 8 |  |  |  |  | 20 | my $name = shift; | 
| 387 | 8 |  | 33 |  |  | 23 | $name ||= $uri; | 
| 388 | 8 |  |  |  |  | 21 | my $line = "=> $uri $name"; | 
| 389 | 8 |  |  |  |  | 16 | my $site = load_site(); | 
| 390 | 8 |  |  |  |  | 26 | my $uri_re = quotemeta($uri); | 
| 391 | 8 |  |  |  |  | 17 | my $name_re = quotemeta($name); | 
| 392 | 8 | 50 |  |  |  | 170 | if (grep(/^=> $uri_re /, @$site)) { | 
|  |  | 50 |  |  |  |  |  | 
| 393 | 0 |  |  |  |  | 0 | warn("$uri already exists in $site_list\n"); | 
| 394 |  |  |  |  |  |  | } elsif (grep(/^=> \S+ $name_re$/, @$site)) { | 
| 395 | 0 |  |  |  |  | 0 | warn("$name already exists in $site_list\n"); | 
| 396 |  |  |  |  |  |  | } else { | 
| 397 | 8 |  |  |  |  | 19 | push(@$site, $line); | 
| 398 |  |  |  |  |  |  | } | 
| 399 | 8 |  |  |  |  | 30 | save_file($site_list, join("\n", @$site, "")); | 
| 400 |  |  |  |  |  |  | } | 
| 401 |  |  |  |  |  |  |  | 
| 402 |  |  |  |  |  |  | =head2 Remove a subscription | 
| 403 |  |  |  |  |  |  |  | 
| 404 |  |  |  |  |  |  | moku-pona remove description | 
| 405 |  |  |  |  |  |  |  | 
| 406 |  |  |  |  |  |  | This removes one or more URLs from the list of subscribed items. | 
| 407 |  |  |  |  |  |  |  | 
| 408 |  |  |  |  |  |  | Example: | 
| 409 |  |  |  |  |  |  |  | 
| 410 |  |  |  |  |  |  | moku-pona remove kensanata | 
| 411 |  |  |  |  |  |  |  | 
| 412 |  |  |  |  |  |  | =cut | 
| 413 |  |  |  |  |  |  |  | 
| 414 |  |  |  |  |  |  | sub do_remove { | 
| 415 | 2 |  |  | 2 |  | 1603 | my @args = @_; | 
| 416 | 2 |  |  |  |  | 7 | my $site = load_site(); | 
| 417 | 2 |  |  |  |  | 6 | my $count = 0; | 
| 418 | 2 |  |  |  |  | 4 | my $i = 0; | 
| 419 | 2 |  | 66 |  |  | 15 | while (@args and $i < @$site) { | 
| 420 | 4 |  |  |  |  | 8 | my $line = $site->[$i]; | 
| 421 | 4 |  |  |  |  | 26 | my ($uri, $name) = $line =~ /^=> (\S+)\s+(.*)/; | 
| 422 | 4 |  |  |  |  | 7 | my $found = 0; | 
| 423 | 4 |  |  |  |  | 6 | my $j = 0; | 
| 424 | 4 |  |  |  |  | 10 | while ($j < @args) { | 
| 425 | 4 | 100 |  |  |  | 11 | if ($name eq $args[$j]) { | 
| 426 | 2 |  |  |  |  | 3 | $count++; | 
| 427 | 2 |  |  |  |  | 3 | $found = 1; | 
| 428 | 2 |  |  |  |  | 6 | splice(@$site, $i, 1); # remove the site found | 
| 429 | 2 |  |  |  |  | 4 | splice(@args, $j, 1); # remove the arg found | 
| 430 |  |  |  |  |  |  | } else { | 
| 431 | 2 |  |  |  |  | 5 | $j++; | 
| 432 |  |  |  |  |  |  | } | 
| 433 |  |  |  |  |  |  | } | 
| 434 | 4 | 100 |  |  |  | 16 | $i++ unless $found; | 
| 435 |  |  |  |  |  |  | } | 
| 436 | 2 | 50 |  |  |  | 6 | if ($count) { | 
| 437 | 2 | 50 |  |  |  | 108 | printf("Removed %d %s\n", $count, | 
| 438 |  |  |  |  |  |  | $count == 1 ? "subscription" : "subscriptions"); | 
| 439 | 2 |  |  |  |  | 18 | save_file($site_list, join("\n", @$site, "")); | 
| 440 |  |  |  |  |  |  | } else { | 
| 441 | 0 |  |  |  |  | 0 | warn("No subscriptions matching @args found\n"); | 
| 442 | 0 |  |  |  |  | 0 | warn("Use moku-pona list to find the correct descriptions.\n"); | 
| 443 |  |  |  |  |  |  | } | 
| 444 |  |  |  |  |  |  | } | 
| 445 |  |  |  |  |  |  |  | 
| 446 |  |  |  |  |  |  | =head2 Clean up the data directory | 
| 447 |  |  |  |  |  |  |  | 
| 448 |  |  |  |  |  |  | moku-pona cleanup [--confirm] | 
| 449 |  |  |  |  |  |  |  | 
| 450 |  |  |  |  |  |  | When Moku Pona updates, copies of the URL targets are saved in the data | 
| 451 |  |  |  |  |  |  | directory. If you remove a subscription (see above), that leaves a cache file in | 
| 452 |  |  |  |  |  |  | the data directory that is no longer used – and it leaves an entry in | 
| 453 |  |  |  |  |  |  | C that is no longer wanted. The cleanup command fixes this. It | 
| 454 |  |  |  |  |  |  | deletes all the cached pages that you are no longer subscribed to, and it | 
| 455 |  |  |  |  |  |  | removes those entries from C as well. | 
| 456 |  |  |  |  |  |  |  | 
| 457 |  |  |  |  |  |  | Actually, just to be sure, if you run it without the C<--confirm> argument, it | 
| 458 |  |  |  |  |  |  | simply prints which files it would trash. Rerun it with the C<--confirm> | 
| 459 |  |  |  |  |  |  | argument to actually do it. | 
| 460 |  |  |  |  |  |  |  | 
| 461 |  |  |  |  |  |  | Example: | 
| 462 |  |  |  |  |  |  |  | 
| 463 |  |  |  |  |  |  | moku-pona cleanup | 
| 464 |  |  |  |  |  |  |  | 
| 465 |  |  |  |  |  |  | =cut | 
| 466 |  |  |  |  |  |  |  | 
| 467 |  |  |  |  |  |  | sub do_cleanup { | 
| 468 | 2 |  | 100 | 2 |  | 2901 | my $confirm = shift||'' eq '--confirm'; | 
| 469 | 2 |  |  |  |  | 4 | my $todo = 0; | 
| 470 |  |  |  |  |  |  | # get a hash map telling us the cache files we expect based on our sites | 
| 471 | 2 |  |  |  |  | 4 | my $site = load_site(); | 
| 472 |  |  |  |  |  |  | my %caches = map { | 
| 473 | 2 |  |  |  |  | 7 | my ($uri, $name) = /^=> (\S+)\s+(.*)/; | 
|  | 2 |  |  |  |  | 16 |  | 
| 474 | 2 |  |  |  |  | 15 | $uri =~ s/[\/:]/-/g; | 
| 475 | 2 |  |  |  |  | 14 | "$data_dir/$uri" => 1; | 
| 476 |  |  |  |  |  |  | } @$site; | 
| 477 |  |  |  |  |  |  | # get a list of text files in the directory | 
| 478 | 2 | 50 |  |  |  | 85 | opendir(my $dh, $data_dir) or die "Cannot read $data_dir: $!\n"; | 
| 479 | 2 |  |  |  |  | 60 | my @files = map { "$data_dir/$_" } grep { /^[^.]/ } readdir($dh); | 
|  | 8 |  |  |  |  | 18 |  | 
|  | 12 |  |  |  |  | 34 |  | 
| 480 | 2 |  |  |  |  | 24 | closedir($dh); | 
| 481 |  |  |  |  |  |  | # remove unnecessary cache files | 
| 482 | 2 |  |  |  |  | 6 | for my $file (@files) { | 
| 483 | 8 | 100 |  |  |  | 18 | next if $file eq $site_list; | 
| 484 | 6 | 100 |  |  |  | 12 | next if $file eq $updated_list; | 
| 485 | 4 | 100 |  |  |  | 10 | next if $caches{$file}; | 
| 486 | 2 | 100 |  |  |  | 4 | if ($confirm) { | 
| 487 | 1 |  |  |  |  | 40 | unlink $file; | 
| 488 |  |  |  |  |  |  | } else { | 
| 489 | 1 |  |  |  |  | 6 | print "trash $file\n"; | 
| 490 | 1 |  |  |  |  | 2 | $todo++; | 
| 491 |  |  |  |  |  |  | } | 
| 492 |  |  |  |  |  |  | } | 
| 493 |  |  |  |  |  |  | # check updates list | 
| 494 | 2 | 50 |  |  |  | 26 | if (-f $updated_list) { | 
| 495 | 2 | 50 |  |  |  | 65 | open(my $fh, "<:encoding(UTF-8)", $updated_list) | 
| 496 |  |  |  |  |  |  | or die "Cannot read $updated_list: $!\n"; | 
| 497 | 2 |  |  |  |  | 180 | my @lines = <$fh>; | 
| 498 | 2 |  |  |  |  | 48 | chomp(@lines); | 
| 499 |  |  |  |  |  |  | # decide what to do about each line in updates, looking just at the names | 
| 500 | 2 |  |  |  |  | 30 | my %sites = map { s/^=> (\S+)\s+(.*)/$2/; $_ => 1 } @$site; | 
|  | 2 |  |  |  |  | 20 |  | 
|  | 2 |  |  |  |  | 10 |  | 
| 501 | 2 |  |  |  |  | 5 | my @deletes; | 
| 502 |  |  |  |  |  |  | my @keeps; | 
| 503 | 2 |  |  |  |  | 4 | for my $line (@lines) { | 
| 504 | 4 | 100 | 66 |  |  | 30 | if ($line =~ /^=> \S+ \d\d\d\d-\d\d-\d\d (.+)/ and not $sites{$1}) { | 
| 505 | 2 |  |  |  |  | 5 | push(@deletes, $line); | 
| 506 | 2 |  |  |  |  | 3 | $todo++; | 
| 507 |  |  |  |  |  |  | } else { | 
| 508 | 2 |  |  |  |  | 6 | push(@keeps, $line); | 
| 509 |  |  |  |  |  |  | } | 
| 510 |  |  |  |  |  |  | } | 
| 511 | 2 | 100 | 66 |  |  | 14 | print "Removing these entries from updates:\n" | 
| 512 |  |  |  |  |  |  | . join("\n", @deletes, "") if @deletes and not $confirm; | 
| 513 |  |  |  |  |  |  | # save | 
| 514 | 2 | 100 |  |  |  | 27 | save_file($updated_list, join("\n", @keeps, "")) if $confirm; | 
| 515 |  |  |  |  |  |  | } | 
| 516 | 2 | 100 | 66 |  |  | 18 | if ($todo && !$confirm) { | 
| 517 | 1 |  |  |  |  | 3 | print "\n"; | 
| 518 | 1 |  |  |  |  | 5 | print "Use moku-pona cleanup --confirm to do it.\n"; | 
| 519 |  |  |  |  |  |  | } | 
| 520 |  |  |  |  |  |  | } | 
| 521 |  |  |  |  |  |  |  | 
| 522 |  |  |  |  |  |  | =head2 Update | 
| 523 |  |  |  |  |  |  |  | 
| 524 |  |  |  |  |  |  | moku-pona update [--quiet] [names...] | 
| 525 |  |  |  |  |  |  |  | 
| 526 |  |  |  |  |  |  | This updates all the subscribed items and generates a new local page for you to | 
| 527 |  |  |  |  |  |  | visit: C. | 
| 528 |  |  |  |  |  |  |  | 
| 529 |  |  |  |  |  |  | Example: | 
| 530 |  |  |  |  |  |  |  | 
| 531 |  |  |  |  |  |  | moku-pona update | 
| 532 |  |  |  |  |  |  |  | 
| 533 |  |  |  |  |  |  | If you call it from a cron job, you might want to use the C<--quiet> argument to | 
| 534 |  |  |  |  |  |  | prevent it from printing all the sites it's contacting (since cron will then | 
| 535 |  |  |  |  |  |  | mail this to you and you might not care for it unless there's a problem). If | 
| 536 |  |  |  |  |  |  | there's a problem, you'll still get a message. | 
| 537 |  |  |  |  |  |  |  | 
| 538 |  |  |  |  |  |  | This is how I call it from my C, for example | 
| 539 |  |  |  |  |  |  |  | 
| 540 |  |  |  |  |  |  | #m   h  dom mon dow   command | 
| 541 |  |  |  |  |  |  | 11 7,14 *   *   *     /home/alex/bin/moku-pona update --quiet | 
| 542 |  |  |  |  |  |  |  | 
| 543 |  |  |  |  |  |  | If you're testing things, you can also fetch just a limited number of items by | 
| 544 |  |  |  |  |  |  | listing them. | 
| 545 |  |  |  |  |  |  |  | 
| 546 |  |  |  |  |  |  | Example: | 
| 547 |  |  |  |  |  |  |  | 
| 548 |  |  |  |  |  |  | moku-pona update "RPG Planet" | 
| 549 |  |  |  |  |  |  |  | 
| 550 |  |  |  |  |  |  | The C files may contain lines that are not links at the top. These | 
| 551 |  |  |  |  |  |  | will remain untouched. The rest is links. New items are added at the beginning | 
| 552 |  |  |  |  |  |  | of the links and older copies of such items are removed from the links. | 
| 553 |  |  |  |  |  |  |  | 
| 554 |  |  |  |  |  |  | =cut | 
| 555 |  |  |  |  |  |  |  | 
| 556 |  |  |  |  |  |  | sub add_update { | 
| 557 | 5 |  |  | 5 |  | 9 | my $line = shift; | 
| 558 | 5 |  |  |  |  | 41 | my ($uri, $name) = $line =~ /^=> (\S+)\s+(.*)/; | 
| 559 |  |  |  |  |  |  | # add current date | 
| 560 | 5 |  |  |  |  | 53 | my ($sec, $min, $hour, $mday, $mon, $year) = gmtime(); # UTC | 
| 561 | 5 |  |  |  |  | 34 | my $date = sprintf('%4d-%02d-%02d', $year + 1900, $mon + 1, $mday); | 
| 562 | 5 |  |  |  |  | 18 | $line = "=> $uri $date $name"; | 
| 563 |  |  |  |  |  |  | # load file | 
| 564 | 5 |  |  |  |  | 43 | my @lines; | 
| 565 | 5 | 100 |  |  |  | 76 | if (-f $updated_list) { | 
| 566 | 4 | 50 |  |  |  | 147 | open(my $fh, "<:encoding(UTF-8)", $updated_list) | 
| 567 |  |  |  |  |  |  | or die "Cannot read $updated_list: $!\n"; | 
| 568 | 4 |  |  |  |  | 390 | @lines = convert(<$fh>); # from gohper | 
| 569 | 4 |  |  |  |  | 69 | chomp(@lines); | 
| 570 |  |  |  |  |  |  | } | 
| 571 |  |  |  |  |  |  | # start the new list with the non-list links | 
| 572 | 5 |  |  |  |  | 35 | my @new = grep(!/^=>/, @lines); | 
| 573 |  |  |  |  |  |  | # add the line to the new list | 
| 574 | 5 |  |  |  |  | 15 | push(@new, $line); | 
| 575 |  |  |  |  |  |  | # add the remaining links to the new list, except for the ones matching the name of the new line | 
| 576 | 5 |  |  |  |  | 8 | $name = quotemeta($name); | 
| 577 | 5 |  |  |  |  | 139 | push(@new, grep(!/\d\d\d\d-\d\d-\d\d $name$/, grep(/^=>/, @lines))); | 
| 578 |  |  |  |  |  |  | # save | 
| 579 | 5 |  |  |  |  | 46 | save_file($updated_list, join("\n", @new, "")); | 
| 580 |  |  |  |  |  |  | } | 
| 581 |  |  |  |  |  |  |  | 
| 582 |  |  |  |  |  |  | sub do_update { | 
| 583 | 6 |  |  | 6 |  | 7417 | my $quiet = grep { $_ eq '--quiet' } @_; | 
|  | 0 |  |  |  |  | 0 |  | 
| 584 | 6 |  |  |  |  | 16 | my @sites = grep { $_ ne '--quiet' } @_; | 
|  | 0 |  |  |  |  | 0 |  | 
| 585 | 6 |  |  |  |  | 8 | my $site = load_site(); | 
| 586 | 6 |  |  |  |  | 22 | my %responses; | 
| 587 |  |  |  |  |  |  | my @uris; | 
| 588 | 6 |  |  |  |  | 0 | my %names; | 
| 589 | 6 |  |  |  |  | 0 | my %lines; | 
| 590 | 6 |  |  |  |  | 13 | for my $line (@$site) { | 
| 591 | 12 |  |  |  |  | 1144 | my ($uri, $name) = $line =~ /^=> (\S+)(?:[ \t]+(.*))?/; | 
| 592 | 12 |  | 33 |  |  | 27 | $name ||= $uri; | 
| 593 | 12 | 50 | 33 |  |  | 34 | next unless @sites == 0 or grep { $_ eq $name } @sites; | 
|  | 0 |  |  |  |  | 0 |  | 
| 594 | 12 | 50 |  |  |  | 979 | say("Fetching $name...") unless $quiet; | 
| 595 | 12 |  |  |  |  | 57 | push(@uris, $uri); | 
| 596 | 12 |  |  |  |  | 39 | $names{$uri} = $name; | 
| 597 | 12 |  |  |  |  | 23 | $lines{$uri} = $line; | 
| 598 | 12 | 50 |  |  |  | 62 | if ($uri =~ /^gopher/) { | 
|  |  | 0 |  |  |  |  |  | 
|  |  | 0 |  |  |  |  |  | 
| 599 | 12 |  |  |  |  | 27 | query_gopher($uri, \%responses); | 
| 600 |  |  |  |  |  |  | } elsif ($uri =~ /^gemini/) { | 
| 601 | 0 |  |  |  |  | 0 | query_gemini($uri, \%responses); | 
| 602 |  |  |  |  |  |  | } elsif ($uri =~ /^http/) { | 
| 603 | 0 |  |  |  |  | 0 | query_web($uri, \%responses); | 
| 604 |  |  |  |  |  |  | } else { | 
| 605 | 0 |  |  |  |  | 0 | warn "Don't know how to fetch $uri\n"; | 
| 606 |  |  |  |  |  |  | } | 
| 607 |  |  |  |  |  |  | } | 
| 608 |  |  |  |  |  |  |  | 
| 609 | 6 | 50 |  |  |  | 1510 | Mojo::IOLoop->start unless Mojo::IOLoop->is_running; | 
| 610 |  |  |  |  |  |  |  | 
| 611 | 6 |  |  |  |  | 2481 | for my $uri (keys %responses) { | 
| 612 | 12 |  |  |  |  | 42 | my $name = $names{$uri}; | 
| 613 |  |  |  |  |  |  | # decode the UTF-8 when we have the entire response | 
| 614 | 12 |  |  |  |  | 60 | my $new = decode_utf8($responses{$uri}); | 
| 615 | 12 | 50 |  |  |  | 317 | if (not $new) { | 
| 616 | 0 |  |  |  |  | 0 | warn("$name returned an empty document\n"); | 
| 617 | 0 |  |  |  |  | 0 | next; | 
| 618 |  |  |  |  |  |  | } | 
| 619 | 12 |  |  |  |  | 16 | my $filename = $uri; | 
| 620 | 12 |  |  |  |  | 108 | $filename =~ s/[\/:]/-/g; | 
| 621 | 12 |  |  |  |  | 30 | my $cache = "$data_dir/$filename"; | 
| 622 | 12 | 100 |  |  |  | 45 | if ($new =~ /^<(\?xml|rss)/) { | 
| 623 | 2 |  |  |  |  | 17 | $new = to_gemini($new); | 
| 624 | 2 |  |  |  |  | 63 | my $encoded = uri_escape_utf8($filename); | 
| 625 | 2 |  |  |  |  | 76 | $lines{$uri} = "=> $encoded $name"; # now referring to the local cache file | 
| 626 |  |  |  |  |  |  | } | 
| 627 | 12 |  |  |  |  | 24 | my $old = load_file($cache); | 
| 628 | 12 | 100 |  |  |  | 288 | if ($new ne $old) { | 
| 629 | 5 | 50 |  |  |  | 514 | say "$name updated" unless $quiet; | 
| 630 | 5 |  |  |  |  | 32 | add_update($lines{$uri}); | 
| 631 | 5 |  |  |  |  | 22 | save_file($cache, $new); | 
| 632 |  |  |  |  |  |  | } else { | 
| 633 | 7 | 50 |  |  |  | 773 | say "$name unchanged" unless $quiet; | 
| 634 |  |  |  |  |  |  | } | 
| 635 |  |  |  |  |  |  | } | 
| 636 |  |  |  |  |  |  | } | 
| 637 |  |  |  |  |  |  |  | 
| 638 |  |  |  |  |  |  | =head2 Subscribing to feeds | 
| 639 |  |  |  |  |  |  |  | 
| 640 |  |  |  |  |  |  | When the result of an update is an XML document, then it is parsed and the links | 
| 641 |  |  |  |  |  |  | of its items (if RSS) or entries (if Atom) are extracted and saved in the cache | 
| 642 |  |  |  |  |  |  | file in the data directory. The effect is this: | 
| 643 |  |  |  |  |  |  |  | 
| 644 |  |  |  |  |  |  | =over | 
| 645 |  |  |  |  |  |  |  | 
| 646 |  |  |  |  |  |  | =item If you subscribe to a regular page, then the link to it in C | 
| 647 |  |  |  |  |  |  | moves to the top when it changes. | 
| 648 |  |  |  |  |  |  |  | 
| 649 |  |  |  |  |  |  | =item If you subscribe to a feed, then the link in C moves to the | 
| 650 |  |  |  |  |  |  | top when it changes and it links to a file in the data directory that links to | 
| 651 |  |  |  |  |  |  | the individual items in the feed. | 
| 652 |  |  |  |  |  |  |  | 
| 653 |  |  |  |  |  |  | =back | 
| 654 |  |  |  |  |  |  |  | 
| 655 |  |  |  |  |  |  | Example: | 
| 656 |  |  |  |  |  |  |  | 
| 657 |  |  |  |  |  |  | moku-pona add https://campaignwiki.org/rpg/feed.xml "RPG" | 
| 658 |  |  |  |  |  |  | moku-pona update | 
| 659 |  |  |  |  |  |  |  | 
| 660 |  |  |  |  |  |  | This adds the RPG entry to C as follows: | 
| 661 |  |  |  |  |  |  |  | 
| 662 |  |  |  |  |  |  | => https%3A--campaignwiki.org-rpg-feed.xml 2020-11-07 RPG | 
| 663 |  |  |  |  |  |  |  | 
| 664 |  |  |  |  |  |  | And if you check the file C, you'll see | 
| 665 |  |  |  |  |  |  | that it's a regular Gemini list. You'll find 100 links like the following: | 
| 666 |  |  |  |  |  |  |  | 
| 667 |  |  |  |  |  |  | => https://alexschroeder.ch/wiki/2020-11-05_Episode_34 Episode 34 | 
| 668 |  |  |  |  |  |  |  | 
| 669 |  |  |  |  |  |  | Now use C (see below) to move the files to the correct | 
| 670 |  |  |  |  |  |  | directory where your Gemini server expects them. | 
| 671 |  |  |  |  |  |  |  | 
| 672 |  |  |  |  |  |  | =cut | 
| 673 |  |  |  |  |  |  |  | 
| 674 |  |  |  |  |  |  | # Convert a RSS or Atom feed to Gemini links | 
| 675 |  |  |  |  |  |  | sub to_gemini { | 
| 676 | 5 |  |  | 5 |  | 34 | my $xml = shift; | 
| 677 | 5 |  |  |  |  | 9 | my $dom = eval { | 
| 678 | 5 |  |  |  |  | 32 | require XML::LibXML; | 
| 679 | 5 |  |  |  |  | 24 | my $parser = XML::LibXML->new(recover => 2); # no errors, no warnings | 
| 680 | 5 |  |  |  |  | 364 | $parser->load_xml(string => $xml); | 
| 681 |  |  |  |  |  |  | }; | 
| 682 | 5 | 50 |  |  |  | 2123 | if ($@) { | 
| 683 | 0 |  |  |  |  | 0 | warn "$@\n"; | 
| 684 | 0 |  |  |  |  | 0 | return ''; | 
| 685 |  |  |  |  |  |  | } | 
| 686 | 5 |  |  |  |  | 54 | my $root = $dom->documentElement(); | 
| 687 | 5 |  |  |  |  | 153 | my $xpc = XML::LibXML::XPathContext->new; | 
| 688 | 5 |  |  |  |  | 35 | $xpc->registerNs('atom', 'http://www.w3.org/2005/Atom'); | 
| 689 | 5 |  | 66 |  |  | 18 | my $nodes = $xpc->findnodes('//atom:entry', $root) || $root->findnodes('//item'); | 
| 690 | 5 |  |  |  |  | 1217 | my @lines; | 
| 691 | 5 |  |  |  |  | 23 | for my $node ($nodes->get_nodelist) { | 
| 692 | 7 |  | 66 |  |  | 95 | my $titles = $xpc->findnodes('atom:title', $node) || $node->getChildrenByTagName('title'); | 
| 693 | 7 |  |  |  |  | 1215 | my $first = $titles->shift; | 
| 694 | 7 | 50 |  |  |  | 112 | my $title = $first ? $first->textContent : "untitled"; | 
| 695 | 7 |  |  |  |  | 154 | $title =~ s/\s+$//; # trim right | 
| 696 | 7 |  |  |  |  | 22 | $title =~ s/^\s+//; # trim left | 
| 697 | 7 |  | 100 |  |  | 26 | my $links = $xpc->findnodes('atom:link', $node) || $node->getChildrenByTagName('link'); | 
| 698 | 7 | 100 |  |  |  | 1048 | next unless $links; | 
| 699 | 6 |  |  |  |  | 136 | my $link = $links->shift; # take the first | 
| 700 | 6 |  | 66 |  |  | 33 | my $href = $link->getAttribute('href') || $link->textContent; | 
| 701 | 6 |  |  |  |  | 125 | push(@lines, "=> $href $title"); | 
| 702 |  |  |  |  |  |  | } | 
| 703 | 5 |  |  |  |  | 126 | return join("\n", @lines, ""); | 
| 704 |  |  |  |  |  |  | } | 
| 705 |  |  |  |  |  |  |  | 
| 706 |  |  |  |  |  |  | =head2 Publishing your subscription | 
| 707 |  |  |  |  |  |  |  | 
| 708 |  |  |  |  |  |  | moku-pona publish | 
| 709 |  |  |  |  |  |  |  | 
| 710 |  |  |  |  |  |  | This takes the important files from your data directory and copies them to a | 
| 711 |  |  |  |  |  |  | target directory. You could just use symbolic links for C and | 
| 712 |  |  |  |  |  |  | C, of course. But if you've subscribed to actual feeds as described | 
| 713 |  |  |  |  |  |  | above, then the cache files need to get copied as well! | 
| 714 |  |  |  |  |  |  |  | 
| 715 |  |  |  |  |  |  | Example: | 
| 716 |  |  |  |  |  |  |  | 
| 717 |  |  |  |  |  |  | mkdir ~/subs | 
| 718 |  |  |  |  |  |  | moku-pona publish ~/subs | 
| 719 |  |  |  |  |  |  |  | 
| 720 |  |  |  |  |  |  | =head2 Serving your subscriptions via Gemini | 
| 721 |  |  |  |  |  |  |  | 
| 722 |  |  |  |  |  |  | This depends entirely on your Gemini server. If you like it really simple, you | 
| 723 |  |  |  |  |  |  | can use L. Here's how to create the certificate and key | 
| 724 |  |  |  |  |  |  | files, copy them to the C<~/subs> directory created above, and run C | 
| 725 |  |  |  |  |  |  | for a quick test. | 
| 726 |  |  |  |  |  |  |  | 
| 727 |  |  |  |  |  |  | make cert | 
| 728 |  |  |  |  |  |  | cp *.pem ~/subs | 
| 729 |  |  |  |  |  |  | cd ~/subs | 
| 730 |  |  |  |  |  |  | lupa-pona | 
| 731 |  |  |  |  |  |  |  | 
| 732 |  |  |  |  |  |  | =cut | 
| 733 |  |  |  |  |  |  |  | 
| 734 |  |  |  |  |  |  | sub do_publish { | 
| 735 | 1 |  |  | 1 |  | 538 | my $target = shift; | 
| 736 | 1 | 50 |  |  |  | 16 | die "Target $target is not a directory\n" unless -d $target; | 
| 737 | 1 | 50 |  |  |  | 22 | die "Source $site_list does not exist\n" unless -f $site_list; | 
| 738 | 1 | 50 |  |  |  | 11 | die "Source $updated_list does not exist\n" unless -f $updated_list; | 
| 739 | 1 |  |  |  |  | 2 | my $path; | 
| 740 |  |  |  |  |  |  | # copy site list | 
| 741 | 1 |  |  |  |  | 6 | copy($site_list, "$target/sites.txt"); | 
| 742 |  |  |  |  |  |  | # copy updates but with local links for the feed files | 
| 743 | 1 | 50 |  | 2 |  | 412 | open(my $in, "<:encoding(UTF-8)", $updated_list) | 
|  | 2 |  |  |  |  | 13 |  | 
|  | 2 |  |  |  |  | 3 |  | 
|  | 2 |  |  |  |  | 46 |  | 
| 744 |  |  |  |  |  |  | or die "Cannot read $updated_list: $!\n"; | 
| 745 | 1 | 50 |  |  |  | 1276 | open(my $out, ">:encoding(UTF-8)", "$target/updates.txt") | 
| 746 |  |  |  |  |  |  | or die "Cannot write $target/updates.txt: $!\n"; | 
| 747 | 1 |  |  |  |  | 102 | for my $line (<$in>) { | 
| 748 | 1 |  |  |  |  | 41 | chomp($line); | 
| 749 | 1 |  |  |  |  | 3 | ($line) = convert($line); | 
| 750 | 1 |  |  |  |  | 8 | my ($uri, $name) = $line =~ /^=> file:\/\/\/(\S+)\s+(.*)/; | 
| 751 |  |  |  |  |  |  | # if the target is a local file, then that's because it is the result of a | 
| 752 |  |  |  |  |  |  | # to_gemini call in do_update, so we need to copy it as well | 
| 753 | 1 |  |  |  |  | 4 | $uri =~ s/[\/:]/-/g; | 
| 754 | 1 | 50 |  |  |  | 17 | if (-f "$data_dir/$uri") { | 
| 755 | 1 |  |  |  |  | 9 | copy("$data_dir/$uri", "$target/$uri"); | 
| 756 |  |  |  |  |  |  | } | 
| 757 | 1 |  |  |  |  | 403 | print $out "$line\n"; | 
| 758 |  |  |  |  |  |  | } | 
| 759 |  |  |  |  |  |  | } | 
| 760 |  |  |  |  |  |  |  | 
| 761 |  |  |  |  |  |  | sub do_help { | 
| 762 | 0 |  |  | 0 |  | 0 | my $parser = Pod::Text->new(); | 
| 763 | 0 |  |  |  |  | 0 | $parser->parse_file($0); | 
| 764 |  |  |  |  |  |  | } | 
| 765 |  |  |  |  |  |  |  | 
| 766 |  |  |  |  |  |  | sub main { | 
| 767 | 0 |  | 0 | 0 |  | 0 | my $command = shift(@ARGV) || "help"; | 
| 768 | 0 | 0 |  |  |  | 0 | if ($command eq "add") { do_add(@ARGV) } | 
|  | 0 | 0 |  |  |  | 0 |  | 
|  |  | 0 |  |  |  |  |  | 
|  |  | 0 |  |  |  |  |  | 
|  |  | 0 |  |  |  |  |  | 
|  |  | 0 |  |  |  |  |  | 
|  |  | 0 |  |  |  |  |  | 
| 769 | 0 |  |  |  |  | 0 | elsif ($command eq "remove") { do_remove(@ARGV) } | 
| 770 | 0 |  |  |  |  | 0 | elsif ($command eq "list") { do_list(@ARGV) } | 
| 771 | 0 |  |  |  |  | 0 | elsif ($command eq "cleanup") { do_cleanup(@ARGV) } | 
| 772 | 0 |  |  |  |  | 0 | elsif ($command eq "update") { do_update(@ARGV) } | 
| 773 | 0 |  |  |  |  | 0 | elsif ($command eq "convert") { do_convert() } | 
| 774 | 0 |  |  |  |  | 0 | elsif ($command eq "publish") { do_publish(@ARGV) } | 
| 775 | 0 |  |  |  |  | 0 | else { do_help() } | 
| 776 |  |  |  |  |  |  | } | 
| 777 |  |  |  |  |  |  |  | 
| 778 |  |  |  |  |  |  | main() if $0 =~ /\bmoku-pona$/; | 
| 779 |  |  |  |  |  |  |  | 
| 780 |  |  |  |  |  |  | 1; |