shakedown.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
A community for live music fans with roots in the jam scene. Shakedown Social is run by a team of volunteers (led by @clifff and @sethadam1) and funded by donations.

Administered by:

Server stats:

289
active users

#archival

1 post1 participant0 posts today

I've mirrored a relatively simple website (redsails.org; it's mostly text, some images) for posterity via #wget. However, I also wanted to grab snapshots of any outlinks (of which there are many, as citations/references). By default, I couldn't figure out a configuration where wget would do that out of the box, without endlessly, recursively spidering the whole internet. I ended up making a kind-of poor man's #ArchiveBox instead:

for i in $(cat others.txt) ; do dirname=$(echo "$i" | sha256sum | cut -d' ' -f 1) ; mkdir -p $dirname ; wget --span-hosts --page-requisites --convert-links --backup-converted --adjust-extension --tries=5 --warc-file="$dirname/$dirname" --execute robots=off --wait 1 --waitretry 5 --timeout 60 -o "$dirname/wget-$dirname.log" --directory-prefix="$dirname/" $i ; done

Basically, there's a list of bookmarks^W URLs in others.txt that I grabbed from the initial mirror of the website with some #grep foo. I want to do as good of a mirror/snapshot of each specific URL as I can, without spidering/mirroring endlessly all over. So, I hash the URL, and kick off a specific wget job for it that will span hosts, but only for the purposes of making the specific URL as usable locally/offline as possible. I know from experience that this isn't perfect. But... it'll be good enough for my purposes. I'm also stashing a WARC file. Probably a bit overkill, but I figure it might be nice to have.

Replied in thread

@FerdiMagellan @jwildeboer @soatok I use #XMPP+#OMEMO, more specifically, @monocles / #monclesChat & @gajim / #gajim!

docs.monocles.eu/apps/chat.app/

  • Another option that may be interesting if you are a business/org & need archival support (and have existing #eMail infrastructure) is @delta / #deltaChat, which is easy to bring into #compliance with #GoBD / #HGD and similar #Archival mandates for professional/business comms.
docs.monocles.eumonocles chat - monocles Documentation

I now have my own fully search-able mirror of #Kiwix hosted on my home server, including #Wikipedia. You can check it out at:

kiwix.marcusadams.me

Content may change as time goes on since I literally "just" got it set up and working, and you should definitely prioritize the original resources and donate to the folks hosting it. But IF some of these resources become unavailable from their original source, feel free to use my mirror as long as it's up.

kiwix.marcusadams.meWelcome to Kiwix Server

The Indie Archive is taken. There is a domain name

theindiearchive.com

for a different paid service.

My project is open source and free to anyone to use.

I'm not even in beta so I'm going to rename.

So far we came up with

Indie Archival Storage

which is an accurate descriptor but not nearly as cool as The Indie Archive.

So ...

Any ideas?

If you're interested here's the link to my podcast, What Is The Indie Archive.

hackerpublicradio.org/eps/hpr4

Thanks

hackerpublicradio.orgHacker Public Radio ~ The Technology Community PodcastHacker Public Radio is a podcast that releases shows every weekday Monday through Friday. Our shows are produced by the community (you) and can be on any topic that is of interest to hackers and hobbyists.

“We just launched a 16TB archive of every dataset that has been available on data.gov since November. This will be updated day by day as new datasets appear. It can be freely copied, and we're sharing the code behind it to help others make their own archives of data they depend on.” Harvard Library Innovation Lab (via BlueSky)

lil.law.harvard.edu/blog/2025/

bsky.app/profile/harvardlil.bs

lil.law.harvard.eduAnnouncing the Data.gov Archive | Library Innovation Lab
Replied in thread

@delta also the whole "BuT #mEtAdAtA?" Discussion is completely blown out of proportions by #Signal fanboys.

In fact, I'm convinced someone already made a #delta #chat #server as an #OnionService over @torproject / #Tor just for the lulz.

  • The biggest Advantage for Delta Chat is that it doesn't require yet another server but instead just uses #IMAP + #SMTP and can even be integrated in #corporate communications that require #archival and #indexing by merely feeding the private keys to said #eMail archival software [i.e. #benno #MailArchiv], which makes it possible to comply with regulations like #GoBD & #HGB where applicable.

Not that this is something the average user encounters, but it is a big bonus for larger organizations!

by pure chance I’ve stumbled upon a beautiful Indonesian web comic The Great Insula, 2024, which seems to be lost media :[

host was hardly crawled by Wayback, I can’t find a mention of its author. tried searching a few social media platforms. one chapter was shared before it disappeared.

DuckDuckGo still has some of the beautiful pages cached. I’ve saved a few of my favourites. just wanted to post here in the hopes it’s not lost forever. lossy archival </3

duckduckgo.com/?q=The+Great+In

I was squinting really hard at the files extracted from #Nintendo #Wii #VirtualConsole version of California Games.

I was like "Wait, why is one of the files called 'califo.games_rem'. That file name doesn't really make sense. ...Wait. is that 'rem' a group tag or something? Could this be the main program??"

Drop the file to C64 emulator.
BOOM. Enjoy the dulcet SIDchip tones of Cracktro.

Nintendo didn't have the cojones to make this the first thing that will be loaded by the Virtual Console emulator, obviously. Cowards.

Hey #library, #archives, #oralHistory & #memoryWorker friends!

I'm teaching a course with @LibraryJuiceAcademy

⭐Solidarity Memory Work⭐
February 3 - March 2

This 4-week class begins with an examination of emerging #solidarity practices in allied fields, such as #journalism, academic #history, and #law. We then explore the practice of #documenting #movements and #organizations, as well as activating #archival records for a more just society.

libraryjuiceacademy.com/shop/c

Library Juice AcademySolidarity Memory Work - Library Juice AcademyMemory work is a broad category of practices, including archival science, librarianship, oral history, journalism, community history, and others, that seek to document a place, community, or event. Solidarity memory work is the application of these practices in explicit solidarity with movements and causes. As our society confronts contemporary civil rights struggles, solidarity memory work is quickly solidifying as a central site of documenting, sharing and preserving the stories of those on the front line. In this class, participants will explore the unique position of memory workers within the broader landscape of political organizing work. This 4-week class begins with an examination of emerging solidarity practices in allied fields, such as journalism, academic history, and law. We then explore the practice of documenting movements and organizations, as well as activating archival records for a more just society.

calling all archivists, file hoarders and anyone else who is willing to help on fedi. i need a favour

i need books, like lots, i want to arhive as many queer books as i can get my hands on. fiction non fiction, scientific papers. i will take anything you have. just i need fiels that are reaady to use. iu dont have the skills to automate scraping or anything

#queer #lgbt #books #archival

#fluConf2025 will include a track on independent publishing and archival. We want to hear stories of what's being done to create non-corporate spaces on the web and preserve the media big companies so often erase.

Tell us about your motivations and experience moving from big platforms like Substack, Twitter, Instagram, and Wix to self-hosted or communally-operated alternatives.

Share your insights into the world of for-profit journals in academia, and efforts to establish better options not based on extraction.

How do you adapt to challenges like the falling adoption of established syndication protocols like RSS, the costs of AI scraping, and ever-changing search engine algorithms? How do you keep up with legal requirements for content moderation and age verification?

With so many corporate platforms shutting down, changing policies on media retention, or moving to monetize content for AI training: how have you gone about archiving your media? What tools and techniques have you used to ensure it isn't lost? How do we resist corporate capture of independent media and foster conditions for more long-lived infrastructure?

Apply up until midnight of January 19th, 2025 (anywhere on Earth)

fluconf.online/apply/

fluconf.onlineSubmit a proposalSubmit your proposal for FluConf 2025 until the end of January 19th, 2025

Update: thanks to @bwana for helping:
discordian.social/@bwana/11370

#FollowerPower: Does anyone have the dimensions for the #packaging of #Windows7 & #WindowsVista?

  • I'm talking about those rounded boxes for the retail version (not SystemBuilder/OEM/...)

#PleaseBoost :boost_ok: :boosted: :boost:

  • I'll also accept anonymous submissions... Contact details pinned in profile / my website...

If anyone has a calipher and said box, it would be cool if they could get the measurements.

  • I could propably "guesstimate" them based off the DVD-ROM inside but I'd ratzer have the correct dimensions just to be shure...

@coreysnipes

It's designed for Indie Producers, like myself, to store projects they really don't ever want to lose, like albums, videos, last year's blog posts, a novel or other writings, photos, etc.

Backups first, backup everything at least daily, automatically.

Archival storage, next, For more redundancy, 7 drives on 4 systems in 3 locations with file integrity checks for completed projects or milestones of works in progress. Files added to archival storage manually.