shakedown.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
A community for live music fans with roots in the jam scene. Shakedown Social is run by a team of volunteers (led by @clifff and @sethadam1) and funded by donations.

Administered by:

Server stats:

269
active users

#ArchiveBox

0 posts0 participants0 posts today
<p>a little tool I built to fight linkrot and save our sources from the memory hole → <a href="https://sij.law/deepciter" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="">sij.law/deepciter</span><span class="invisible"></span></a> </p><p><a href="https://earth.law/tags/digitalpreservation" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>digitalpreservation</span></a> <a href="https://earth.law/tags/selfhosting" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>selfhosting</span></a> <a href="https://earth.law/tags/archivebox" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>archivebox</span></a> <a href="https://earth.law/tags/opensource" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>opensource</span></a> <a href="https://earth.law/tags/foss" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>foss</span></a> <a href="https://earth.law/tags/textfragments" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>textfragments</span></a> <a href="https://earth.law/tags/waybackmachine" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>waybackmachine</span></a> <a href="https://earth.law/tags/linkrot" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>linkrot</span></a> <a href="https://earth.law/tags/memoryhole" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>memoryhole</span></a> <a href="https://earth.law/tags/legaltech" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>legaltech</span></a> <a href="https://earth.law/tags/permalink" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>permalink</span></a> <a href="https://earth.law/tags/deepcite" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>deepcite</span></a></p>
Tom MacWright<p>been trying to archive all outlinks from macwright.com with <a href="https://mastodon.social/tags/archivebox" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>archivebox</span></a> and results are decidedly mixed: tasks keep getting stuck in a 'pending' state with no feedback as to whether anything is working or not.</p>
Preston Maness ☭<p>I've mirrored a relatively simple website (redsails.org; it's mostly text, some images) for posterity via <a href="https://tenforward.social/tags/wget" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>wget</span></a>. However, I also wanted to grab snapshots of any outlinks (of which there are many, as citations/references). By default, I couldn't figure out a configuration where wget would do that out of the box, without endlessly, recursively spidering the whole internet. I ended up making a kind-of poor man's <a href="https://tenforward.social/tags/ArchiveBox" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>ArchiveBox</span></a> instead:</p><p>for i in $(cat others.txt) ; do dirname=$(echo "$i" | sha256sum | cut -d' ' -f 1) ; mkdir -p $dirname ; wget --span-hosts --page-requisites --convert-links --backup-converted --adjust-extension --tries=5 --warc-file="$dirname/$dirname" --execute robots=off --wait 1 --waitretry 5 --timeout 60 -o "$dirname/wget-$dirname.log" --directory-prefix="$dirname/" $i ; done</p><p>Basically, there's a list of bookmarks^W URLs in others.txt that I grabbed from the initial mirror of the website with some <a href="https://tenforward.social/tags/grep" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>grep</span></a> foo. I want to do as good of a mirror/snapshot of each specific URL as I can, without spidering/mirroring endlessly all over. So, I hash the URL, and kick off a specific wget job for it that will span hosts, but only for the purposes of making the specific URL as usable locally/offline as possible. I know from experience that this isn't perfect. But... it'll be good enough for my purposes. I'm also stashing a WARC file. Probably a bit overkill, but I figure it might be nice to have.</p><p><a href="https://tenforward.social/tags/RedSails" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>RedSails</span></a> <a href="https://tenforward.social/tags/archive" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>archive</span></a> <a href="https://tenforward.social/tags/archival" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>archival</span></a> <a href="https://tenforward.social/tags/archiving" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>archiving</span></a> <a href="https://tenforward.social/tags/warc" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>warc</span></a></p>
Enron Hubbard<p>of course, neofeudal lords are looking to <a href="https://tilde.zone/tags/wikipedia" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>wikipedia</span></a> and <a href="https://tilde.zone/tags/internetArchive" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>internetArchive</span></a> with arson in their hearts, as they always do with the great libraries</p><p>between this and the web continuing to enshittify with AI slop and critical mass of advertising, it's probably time to start thinking about things in terms of offline-first </p><p>make local copies of resources that are important, get your personal content off of cloud providers, and archive everything you can</p><p>old phones, random flash drives, unused laptops - all of that can be put to good use as self-sovereign libraries. and if you have the financial means, seriously consider building or investing in a NAS</p><p>we have plenty of tools to make this possible:</p><p>kiwix is an offline reader for Wikipedia, Project Gutenberg, and several other online sources - there's even a method to turn a raspi into a hotspot that serves the archived content: <a href="https://kiwix.org/en/how-to-set-up-kiwix-hotspot/" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">kiwix.org/en/how-to-set-up-kiw</span><span class="invisible">ix-hotspot/</span></a></p><p>youtube-dl is a program that you can use to download content from youtube, including full channels: <a href="https://ytdl-org.github.io/youtube-dl/" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="">ytdl-org.github.io/youtube-dl/</span><span class="invisible"></span></a></p><p>the Internet Archive also has a command line utility to bulk download content: <a href="https://archive.org/developers/internetarchive/cli.html" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">archive.org/developers/interne</span><span class="invisible">tarchive/cli.html</span></a></p><p>and take a look at <a href="https://tilde.zone/tags/archiveBox" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>archiveBox</span></a> - a self-hosted project that takes in urls and downloads relevant content while stripping out all of the extra shit you don't need: <a href="https://archivebox.io/" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="">archivebox.io/</span><span class="invisible"></span></a></p>
black lipstick on your flight controls<p>surviving and healthy bookmark archiver tools from last time i looked into replacing Pocket include <a href="https://www.linkace.org/" rel="nofollow noopener" target="_blank">LinkAce</a> and <a href="https://archivebox.io/" rel="nofollow noopener" target="_blank">ArchiveBox</a>. i don't see a clear winner based on two minutes of browsing docs except based on implementation language vibes: ArchiveBox is Python, while LinkAce is PHP.</p><p><a href="https://princess.industries/tags/mozilla" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Mozilla</span></a> <a href="https://princess.industries/tags/pocket" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Pocket</span></a> <a href="https://princess.industries/tags/linkace" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>LinkAce</span></a> <a href="https://princess.industries/tags/archivebox" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>ArchiveBox</span></a> <a href="https://princess.industries/tags/selfhost" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>SelfHost</span></a> <a href="https://princess.industries/tags/selfhosted" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>SelfHosted</span></a> <a href="https://princess.industries/tags/selfhosting" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>SelfHosting</span></a></p>