Original link: https://blog.save-web.org/blog/2023/04/16/stwp-2023-%E7%AC%AC-15-%E5%91%A8%E5%91%A8%E6 %8A%A5/
Week 15 Project Summary
- After unsuccessful attempts to contact the original owner of the wikiteam bot, we deployed our savewebbot on the wikiapiary.
- Preparing/promoting the first archive-thon event (theme is DokuWiki archives).
- Opened 10 threads for savewebbot, and the wikiapiary administrator said that the bot edited pages too frequently. So add a limit of one edit per minute.
- The original wikiteam bot uses a very hearse regular to modify wikitext. Deprecated, saveweb bot uses “more scientific” approach instead.
- Continue writing/repairing podcast archive tools. Fixed a BUG that the tool will blow up if an RSS enclosure reports an illegal size.
- Archived the DokuWiki site https://wiki.bash-hackers.org/ at the request of @JAA in wikiteam IRC. IA item
Discussion summary
Apart from IA, I haven’t found any other centralized storage service that simultaneously (compliantly) meets the requirements of unlimited capacity, free, long-term storage, and not-so-outrageous censorship.
All the recent upsurges have made me realize that human beings are highly inadequate in recording and collating various current development processes, and are disturbed and blocked by hot and popular events. For example, most of the recent AI-related products have not been recorded in places like Wikipedia.
What is the result of this? The few who were lucky enough to be recorded will become a representative symbol of the future pursuit of this history. Cause regret, distortion and misunderstanding.
Is there any tool for automatically archiving the designated up master of station b or even automatically archiving the videos in your own favorites?
Today I finally had some time to look at dokuwiki-dumper , and then I saw the last one of requirements, almost understood that only rich people can use it (x
STWP seems a bit similar to miHoYo’s slogan. (literally)
Grass, Archivist Save The World, right?
Saveweb bot for collecting nectar on wikiapiary
WikiTeam ‘s WikiTeam Bot on wikiapiary.com hasn’t worked since 2016, and none of the wikidumps ( MediaWiki site backups ) uploaded to IA since then link well into wikiapiary. Plus wikiapiary at The template reminds users that “wikidump related information_ is automatically maintained by wikiteam bot, and manual editing will be overwritten_”, which dispels the idea of many editors to manually maintain relevant information.
Since then, except for the wikidump information of fandom.com, there is a [[User:Shufflertoxin]]
bot on wikiapiary, which has basically stagnated except for a small amount of maintenance.
We wrote a new bot [[User:Savewebbot]]
and deployed it to wikiapiary, which is relinking 8 years of wikidump uploaded to IA to wikiapiary .
This bot helps us (and others) filter the wikiapiary for MediaWiki sites that haven’t been archived yet, and… keep it.
Source code: https://github.com/saveweb/wikiapiary-wikiteam-bot (modified from old code at wikiteam/wikiteam)
- Wikiapiary was down for a few months before, and its bot has a backlog of a large number of queued tasks. In addition, they have just upgraded to MediaWIki 1.39.2. They are still investigating database performance issues, so it is not very stable now. It takes a while to open and refresh few times.
This article is reproduced from: https://blog.save-web.org/blog/2023/04/16/stwp-2023-%E7%AC%AC-15-%E5%91%A8%E5%91%A8%E6 %8A%A5/
This site is only for collection, and the copyright belongs to the original author.