A script for archiving URL sets (HTML, PDF) in the Internet Archive.
© 2015-2025 Ingram Braun https://ingram-braun.net/
This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version.
This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
You should have received a copy of the GNU General Public License along with this program. If not, see http://www.gnu.org/licenses/.
There is an image on Docker Hub now. It is strongly recommended to prefer it over self-install or Windows binaries:
$ docker pull archaeopath/archive.pl-docker:latest
Perl 5.24 (earlier versions not tested but it is likely to work with every build that is capabale of getting the required modules installed). If there are issues with installing the XMLRPC::lite
module, do it with CPAN's notest
pragma.
Collect URLs you want to archive in file urls.txt
separated by one or more line breaks and UTF-8-encoded and call perl archive.pl
. The script does to things: it fetches the URLs and extracts some metadata (works with HTML and PDF). It submits them to Internet Archive. Then it generates a HTML file with a link list that you may post to your blog or use autmoated submission to WordPress. Alternatively you can get the link list as Atom feed. If login credentials are provided but no FTP host, it is posted to WordPress. The WP URL must point to the xmlrpc.php
. Additionally you can post the links on Twitter. Regardless of the format you can upload the result file on a server via FTP.
Internet Archive has a submission limit of 15 URLs per minute per IP. Set an appropriate delay (at least five seconds) to meet it. If you want to be faster, set a proxy server which rotates IPs. This is fi. possible with TOR as service (the TOR browser on Windows does not work here!). Set MaxCircuitDirtiness 10
in the configuration file (/path/to/torrc
) to rotate IPs every ten seconds.
There are several optional switches
-a
output as Atom feed instead of HTML
-c <creator>
name of feed creator (feed only)
-d <path>
FTP path
-D
Debug mode - don't save to Internet Archive
-f <filename>
name of input file if other than urls.txt
-F
Download root URLs by Firefox (recommended)
-h
Show commands
-i <title>
Feed or HTML title
-k <consumer key>
deprecated
-l
save linked documents
-n <username>
FTP or WordPress user
-o <host>
FTP host
-p <password>
FTP or WordPress password
-P <proxy>
A proxy, fi. socks4://localhost:9050 for TOR service
-r
obey robots.txt
-s
save feed in Wayback machine
-t <access token>
deprecated
-T <seconds>
delay per URL in seconds to respect IA's request limit
-u <URL>
Feed or WordPress (xmlrpc.php
) URL
-w
deprecated
-x <secret consumer key>
deprecated
-y <secret access token>
deprecated
-z <time zone>
Time zone (WordPress only)
- Removed the ability to post URLs on Twitter due to Twitter API changes.
- Removed the ability to store Twitter images in various sizes because Twitter has abandoned the corresponding markup since long.
- Blacklist extended.
- Blacklist of URLs (mostly social media sharing URLs which only point to a login mask).
- Docker image on Docker Hub.
- Fixed image data URLs bug.
- Removed TLS from Wayback URLs (too many protocol errors).
- Added
-F
option which downloads root URLs through Firefox to circumvent DDoS protection etc. - Added
mirror.pl
.
- Introduced option
-P
to connect to a proxy server that can rotate IPs (fi. TOR). - User agent bug in
LWP::UserAgent
constructor call fixed. -T
is able to eat floats.- Screen logging enhanced (total execution time and total number of links).
- IA JSON parsing more reliable.
- Script can save all linked URLs, too (IA has restricted this service to logged-in users running JavaScript).
- Debug mode (does not save to IA).
- WordPress bug fixed (8-Bit ASCII in text lead to database error).
- Ampersand bug in 'URL available' request fixed.
- Trim metadata.
- Disregard
robots.txt
by default.
- Post HTML outfile to WordPress
- Wayback machine saves all documents linked in the URL if it is HTML (Windows only).
- Time delay between processing of URLs because Internet Archive set up a request limit.
- Version and help switches.
- Tweet URLs.
- Enhanced handling of PDF metadata.
- Always save biggest Twitter image.
Not published.
- Supports
wget
andPowerShell
(w flag). - Displays the closest Wayback copy date.
- Better URL parsing.
- Windows executable only 64 bit since not all modules install properly on 32.
- Enhanced metadata scraping.
- Archive images from Twitter in different sizes.
- Added project page link to outfile.
- Remove UTF-8 BOM from infile.
- User agent avoids strings
archiv
andwayback
. - Internet Archive via TLS URL.
- Thumbnail if URL points to an image.
- Debugging messages removed.
- Archive.Org URL changed.
- Internationalized domain names (IDN) allowed in URLs.
- Blank spaces allowed in URLs.
- URL list MUST be in UTF-8 now!
- Only line breaks allowed as list separator in URL list.
- Added workaround for ampersand bug on Windows in module
Browser::Open
https://rt.cpan.org/Ticket/Display.html?id=117917&results=035ab18171a4a673f347e0ca5a8629f4
Binaries Windows 64 bit (XP or higher) can get obtained from the project page https://ingram-braun.net/erga/archive-pl-a-perl-script-for-archiving-url-sets-in-the-internet-archive/