Long term I see it as potential revenue stream. I don’t believe in decentralized web. I believe in hyper local web. The newspapers or local libraries can be backbone. Provide dead simple publishing software on subscription, coupon newsletters, take back the marketplace from facebook.
The web is local news. Local news belongs to the people. thanks for what you do.
BTW here is library project I am working on https://checkoutmydomain.glitch.me
Greg, the outlet you’re thinking of is ColoradoBlvd.net, a local paper here in Pasadena, CA, which does support webmentions including backfeed of interactions with Twitter using Brid.gy. (Sadly Facebook’s API turned off their access to this sort of feature on August 1st.)
As for Ben Keith’s concern about spam, yes, Webmention can be a potential vector like trackbacks and pingbacks, but it does learn from their mistakes with better mitigation and verification. Work on the Vouch protocol/extension of Webmention continues to mitigate against these issues. I’ll also note that Akismet for WordPress works relatively well for Webmentions too, though there have still yet to be examples of Webmention spam in the wild.
For publishers using WordPress, there are some excellent plugins including Webmention (which has some experimental Vouch plumbing included already) and Symantic Linkbacks which work with WordPress’s native comments. I’ll note that they’re developed and actively maintained by several, including the core maintainer for pingbacks and trackbacks in WordPress.
The researcher’s post can webmention an aggregating website similar to the way they would pre-print their research on a server like arXiv.org. The aggregating website can then parse the original and display the title, author(s), publication date, revision date(s), abstract, and even the full paper itself. This aggregator can act as a subscription hub (with WebSub technology) to which other researchers can use to find, discover, and read the original research.
Readers of the original research can then write about, highlight, annotate, and even reply to it on their own websites to effectuate peer-review which then gets sent to the original by way of Webmention technology as well. The work of the peer-reviewers stands in the public as potential work which could be used for possible evaluation for promotion and tenure.
Readers of original research can post metadata relating to it on their own website including bookmarks, reads, likes, replies, annotations, etc. and send webmentions not only to the original but to the aggregation sites which could aggregate these responses which could also be given point values based on interaction/engagement levels (i.e. bookmarking something as “want to read” is 1 point where as indicating one has read something is 2 points, or that one has replied to something is 4 points and other publications which officially cite it provide 5 points. Such a scoring system could be used to provide a better citation measure of the overall value of of a research article in a networked world. In general, Webmention could be used to provide a two way audit-able trail for citations in general and the citation trail can be used in combination with something like the Vouch protocol to prevent gaming the system with spam.
Government institutions (like Library of Congress), universities, academic institutions, libraries, and non-profits (like the Internet Archive) can also create and maintain an archival copy of digital and/or printed copies of research for future generations. This would be necessary to guard against the death of researchers and their sites disappearing from the internet so as to provide better longevity.
I’m hitting a fork in the road with this site and the experiment of using a blog as a directory of blogs. The problem here is me: I’m running out of time. I’m duplicating a lot …
Source: Announcement: The Future of Blog Snoop – Blog Snoop Weblog Directory
We’ll see what happens. It...
If I recall, programming wasn’t necessarily your strong suit, but like many in the IndieWeb will say: “Manual until it hurts!” By doing things manually, you’ll more easily figure out what might work and what might not, and then when you’ve found the thing that does, then you spend some time programming it to automate the whole thing to make it easier. It’s quite similar to designing a college campus: let the students walk around naturally for a bit then pave the natural walkways that they’ve created. This means you won’t have both the nicely grided and unused sidewalks in addition to the ugly grass-less beaten paths. It’s also the broader generalization of paving the cow paths.
In addition to my Following page I’ve also been doing some experimenting with following posts using the Post Kinds Plugin. It is definitely a lot more manual than I’d like it to be. It does help to have made a bookmarklet to more quickly create follow posts, but until I’ve got it to a place that I really want it, it’s not (yet) worth automating taking the data from those follow posts to dump them into my Follow page for output there as well. Of course the fact that my follow posts have h-entry and h-feed mark up means that someone might also decide to build a parser that will extract my posts into a feed which could then be plugged into something else like a microsub-based reader so that I could make a follow post on my own site and the source is automatically added to my subscription list in my reader automatically.
In addition to Kicks Condor, I’me seeing others start to kick the tires of these things as well. David Shanske recently wrote Brainstorming on Implementing Vouch, Following, and Blogrolls, but I think he’s got a lot more going on in his thinking than he’s indicated in his post which barely scratches the surface.
I also still often think back to a post from Dave Winer in 2016: Are you ready to share your OPML?This too has some experimental discovery features that only scratch the surface of the adjacent possible.
And of course just yesterday, Kevin Marks (previously of Technorati) reminded us about rel=”directory” which could have some interesting implications for discovery and following. Think for a bit of how one might build a decentralized Technorati or something along the lines of Ryan Barrett’sindie map.
As things continue to grow, I’m seeing some of all of our decisions and experiments begin to effect others as these are all functionality and discovery mechanisms that we’ll all need in the very near future. I hope you’ll continue to experiment and make cow paths that can eventually be paved.
Vouch is an extension to the webmention protocol. Webmentions usually have two parameters…source and target. Target is the URL on your website that the Source URL is linking to.
The vouch parameter is a third URL to help the target determine whether or not they should accept the webmention. This...
I like the sound of where this is going already! All these small little pieces loosely joined to build a much larger edifice is certainly interesting.
I’ve got a somewhat reasonable bookmarklet for quickly following people, though it’s not marked up with XFN data (yet) — perhaps another data field for Post Kinds? I do wish that there was either a mechanism for adding those to my Following page via the WordPress Link Manager or someone had a means of parsing lots of follow posts so I could quickly have data for both Vouch as well as for microsub readers either via my follow feed list or via OPML export and/or OPML subscription. WordPress obviously has some of the infrastructure built already, but there’s certainly a more IndieWeb way of doing it that wouldn’t require side-files like OPML.