Replied to a post by Ton Zijlstra (zylstra.org)
Is it possible to annotate links in Hypothes.is that are in the Internet Archive? My browser bookmarklet for it doesn’t work on such archived pages. I can imagine that there are several javascript or iframe related technical reasons for it. An information related reason may be that bringing togeth...
The ability to annotate archived material on the Internet Archive with Hypothes.is is definitely possible, and I do it from time to time. I’m not sure which browser or annotation tool (via, browser extensions, other) you’re using, but it’s possible that one or more combinations may have issues allowing you to do it or not. The standard browser extension on Chrome has worked well for me in the past.

Hypothes.is has methods for establishing document equivalency which archive.org apparently conforms. I did an academic experiment a few years back with an NYT article about books where you’ll see equivalent annotations on the original, the archived version, and a copy on my own site that has a rel="canonical" link back to the original as well: 

  • https://www.nytimes.com/2017/01/16/books/obamas-secret-to-surviving-the-white-house-years-books.html
  • https://web.archive.org/web/20170119220705/https://www.nytimes.com/2017/01/16/books/obamas-secret-to-surviving-the-white-house-years-books.html
  • https://boffosocko.com/2017/01/19/obamas-secret-to-surviving-the-white-house-years-books-the-new-york-times/

I don’t recommend doing the rel-canonical trick on your own site frequently as I have noticed a bug, which I don’t think has been fixed.

The careful technologist with one tool or another, will see that I and a couple others have been occasionally delving into the archive and annotating Manfred Kuehn’s work. (I see at least one annotation from 2016, which was probably native on his original site before it was shut down in 2018.) I’ve found some great gems and leads into some historical work from his old site. In particular, he’s got some translations from German texts that I haven’t seen in other places.

Read captured conversation about Parler (Pastebin)

<Dash> I didn't see any mention of the header in the github repo so I feel it's helpful to mention this: in case anyone gets ratelimited, Parler will honor an arbitrary x-forwarded-for header with any IP and, well, not ratelimit you according to my unscientific test
[...]
<kallsyms> Dash: lmao really?
[...]
<andrew> Dash: okay, that's actually pretty huge cc kiska arkiver Fusl
<Kaz> heh, good to know Dash
<ave> this is amazing
[...]
<NotNite> yep
<Kaz> does that also apply to the API Dash?
<andrew> so, who wants to modify the thing to generate new IPs for X-Forwarded-For for each job and see what difference that makes?
[...]
<Dash> Not sure, but i'm credential stuffing them with 300 threads without getting ratelimited
[...]
<Dash> Someone else should probably test

Read 2020-11-04: New Twitter UI: Replaying Archived Twitter Pages That Never Existed by Himarsha Jayanetti (ws-dl.blogspot.com | The Web Science and Digital Libraries Research Group at Old Dominion University.)
When you visit web archives to go back in time and look at a web page, you naturally expect it to display the content exactly as it appeared on the live web at that particular datetime. That is, of course, with the assumption in mind that all of the resources on the page were captured at or near the time of the datetime displayed in the banner for the root HTML page. However, we noticed that it is not always the case and problems with archiving Twitter's new UI can result in replaying Twitter profile pages that never existed on the live web. In our previous blog post, we talked about how difficult it is to archive Twitter's new UI, and in this blog post, we uncover how the new Twitter UI mementos in the Internet Archive are vulnerable to temporal violations.
An interesting quirk of archiving pages on the modern internet.
Read New ways to learn about web archiving and Archive-It (Archive-It Blog)
We recently highlighted opportunities for partners and peers to learn more about web archiving technology and practices through the Archive-It Advanced Training webinar series–all recorded and available on-demand. As more organizations and communities find web archiving needs though, Internet Archive staff are also introducing new and extended training materials to get them crawling for the first time.
Bookmarked National Emergency Library : Free Texts : Free Download, Borrow and Streaming (Internet Archive)
A collection of books that supports emergency remote teaching, research activities, independent scholarship, and intellectual stimulation while universities, schools, training centers, and libraries are closed.
Read The Internet's Dark Ages (The Atlantic)
If a Pulitzer-nominated 34-part series of investigative journalism can vanish from the web, anything can.

[The Web] is a constantly changing patchwork of perpetual nowness.

Highlighted on January 07, 2020 at 11:58AM

Replied to IFTTT Recipes for PESOS by Charlotte AllenCharlotte Allen (charlotteallen.info)
So, I spend a long time trying to set up PESOS for individual silos on IFTTT, specifically Facebook and Instagram, because they are terrible. I’ve got it currently set up to publish my initial post, but no back feed support yet. Also, this is going to wordpress, but it shouldn’t matter (in theor...
This is some brilliant work. Thanks for puzzling it all out.

I do have a few questions/clarifications though so as not to be confused since there are a few pieces you’ve left out.

For the IndieAuth token, which is created at /wp-admin/users.php?page=indieauth_user_token one only needs to give it a title and the “create” scope?

For the “then” portion that uses IFTTT.com’s Webhooks service are the following correct?

  • The URL is (when used with WordPress) of the form: https://example.com/wp-json/micropub/1.0/endpoint
  • The Method is: POST
  • The Content Type I’m guessing based on the Body field you’ve included is: application/x-www-form-urlencoded

For your Pocket example, it looks like you’re using the Post Kinds Plugin, so I’m guessing that you could have gotten away without the {{Excerpt}} and {{Title}} portions and just have sent the URL which Post Kinds picks up and parses to give you your context portion with a title and an excerpt anyway?

It looks like part of the trouble of this PESOS set up is that you’re too reliant in the long run of relying on Pocket (or other services) being around in the long term. If Pocket disappears, then really, so does most of your bookmark, which ideally should point to the canonical URL of the content you’re bookmarking. Of course perhaps IFTTT may not give you that URL in many cases. It looks to me like the URL you’re bookmarking would make a more appropriate syndication link URL.

For most of my bookmarks, likes, reads, etc. I use a plugin that scrapes my post and saves a copy of the contents of all the URLs on my page to the Internet Archive so that even in the event of a site death, a copy of the content is saved for me for a later date.

In any case, I do like this method if one can get it working. For some PESOS sources, I’ve used IFTTT before, though typically with RSS feeds if the silo provides them. Even then I’m often saving them directly to WordPress as drafts for later modification if the data that IFTTT is providing is less than ideal. Sometimes worse, using RSS doesn’t allow one to use Post Kinds URL field and parsing functionality the way your webhook method does.

Replied to LA Roadshow Recap by Jim GroomJim Groom (bavatuesdays)

10 days ago I was sitting in a room in Los Angeles with 12 other folks listening to Marie SelvanadinSundi Richard, and Adam Croom talk about work they’re doing with Domains, and it was good! That session was followed by Peter Sentz providing insight on how BYU Domains provides and supports top-level domains and hosting for over 10,000 users on their campus. And first thing that Friday morning Lauren and I kicked the day off by highlighting Tim Clarke’s awesome work with the Berg Builds community directory as well as Coventry Domains‘s full-blown frame for a curriculum around Domains with Coventry Learn. In fact, the first 3 hours of Day 2 were a powerful reminder of just how much amazing work is happening at the various schools that are providing the good old world wide web as platform to their academic communities. 

https://roadshow.reclaimhosting.com/LA/

I’m still bummed I couldn’t make it to this event…

One of the questions that came up during the SPLOT workshop is if there’s a SPLOT for podcasting, which reminded me of this post Adam Croom wrote a while back about his podcasting workflow: “My Podcasting Workflow with Amazon S3.” . We’re always on the look-out for new SPLOTs to bring to the Reclaim masses, and it would be cool to have an example that moves beyond WordPress just to make the point a SPLOT is not limited to WordPress (as much as we love it) —so maybe Adam and I can get the band back together.

I just outlined a tiny and relatively minimal/free way to host and create a podcast feed last night: https://boffosocko.com/2019/12/17/55761877/

I wonder if this could be used to create a SPLOT that isn’t WordPress based potentially using APIs from the Internet Archive and Huffduffer? WordPress-based infrastructure could be used to create it certainly and aggregation could be done around tags. It looks like the Huffduffer username SPLOT is available.
–annotated December 17, 2019 at 10:46AM

I’ve been going through a number of broken links on my website and slowly, but surely, bringing many of them back to life. Thanks Broken Link Checker! Apparently there were 429 broken links, but I’m able to quickly fix many of them because as I made my posts, I backed up the original links automatically using Post Archival in the Internet Archive. (Somehow this plugin has violated one of WordPress’ guidelines and can’t be downloaded, though I haven’t seen any details about why or been notified about it.)

I’ve only come across one or two which archive.org didn’t crawl or didn’t have. Many of the broken links I’m able to link directly to archive copies on the same day I made them and my archive snapshots were the only ones ever made.

Replied to #oext374 #oextend Send Someone a Message from the Internet Archive’s Great 78 Collection | The Daily Extend (extend-daily.ecampusontario.ca)

Not only is it digitized analog, it’s an amazing open resource of music history. The Internet Archive’s Great 78 Project has over 150,000 digitized 78rpm discs

Browse the collection, and look for a title the represents how you feel about Ontario Extend, or a colleague’s work. Tweet it out so we can all listen to the digital record spin.

Yes, Extend, You Are My Sunshine.