Replies
From what I can tell, they’re creating the content to help users realize how to set up a separate front page and a separate blog page which isn’t always intuitive to newcomers to the platform. Fortunately they haven’t made too much useless content and new users can simply either edit the front page the theme creates to something that suits them or just delete that page and create something of their own.
The only other content I can see that one might want to modify are the two footer widgets that you can either edit or simply delete or replace.
In all I think they’ve taken this route simply to give new users an idea about how they could set up their sites and give them an idea about the way the theme uses the Gutenberg editor. (It’s not too different from the long standing “Hello World” introductory post or the “Hi, this is a comment.” first comment on a fresh install.) For people new to WordPress this is probably pretty helpful, though for older hands it may be annoying. Fortunately the content the customizer creates is pretty minimal and easy to get rid of.
I ran across an example yesterday of someone using a private local TiddlyWiki as a static site web generator, which is quite different from people hosting them directly on web servers.
I’m interested in off-label use cases for wikis (particularly in the vein of commonplace books), so do let us know when your article comes out.
While most of its ecosystem revolves around methods for running the program locally (and often privately) or in Google or Dropbox storage, I’ve come across a growing number of people hosting their instances on their own servers and using them publicly as a melange of personal websites, blogs, and wikis.
Has anyone tried hosting one (particularly the newer TW5) through Reclaim before? Of the many methods, I’m curious which may be the easiest/simplest from a set up perspective?
Here are some interesting examples I’ve come across:
* “A Thesis Notebook” by Alberto Molina
* PESpot Lesson Planner by Patrick Detzner (this one seems to be heavily modified)
* sphygm.us
I am aware that Jon Udell has built some separate UI that may help out people trying this as a commenting system. Here’s a URL to it with an illustrative example: https://jonudell.info/h/facet/?wildcard_uri=https%3A%2F%2Fblog.cjeller.site%2F*&max=100
My favorite solution to this problem and a few related others (like updating my bio and where you can find me on social media) is the meta data route using something like Microformats. Since I provide an h-card
on my website’s homepage, it should be relatively easy for any service to take my URL as my identity (rather than one of my thousands of email addresses) parse my page and find my name, photo, bio, etc. and display them.
Nearly every social silo on the planet wants all of these details, so why should I need to incessantly have to input them manually much less keep them up to date? And I’ve yet to see a social service in the wild that hasn’t asked for my URL, so it’s obviously pretty universal.
Jeremy Keith‘s Huffduffer is a great example of something that already uses this data nicely. It doesn’t pull in my photo (though I think at one time he did have a set up that would poll Flickr avatars?) or my bio, but the “Elsewhere” section of my Huffduffer account lists where you can find me on dozens of social media accounts as well as my own websites. Huffduffer can do this because I gave it my domain name and the service parses my page looking for the rel="me"
tags on my homepage. It could easily pull in my other provided data.
Incidentally Kevin Marks has also proposed a distributed verification system (remember the problem that Twitter had of attempting this?) that uses the rel="me"
idea.
I’ll note that my own website will parse yours to pull in the author name, URL, and avatar to display a reply context for this response on my website! So hooray for microformats! (Though I’ll note that I did modify them a tad for my own idiosyncrasies.) My site does this with David Shanske‘s excellent Post Kinds plugin uses Parse This, which parses for microformats, JSON-LD, and then, if nothing is found it falls back to Open Graph Protocol. He’s been extending it lately to cover a handful of the bigger snowflake services like YouTube, IMDb, etc. to cover some additional edge cases that don’t have good mark up. Incidentally Aaron Parecki has a version of something like this called X-ray, which he uses for various things including microsub readers, not to mention the variety of other parsers available.
I’m sure there may be other versions of this in the wild, but it would be cool to see more social services provide functionality like this.
Clicking through to the photo, there is no mention of this image appearing on this important announcement. Perhaps the author privately contact the photographer about using his image. Since Ken Doctor is so incredible with his media experience (i’m being serious), I’m fairly certain someone from his team would have contacted the photographer to give him a heads up. ❧
I’m sure I’ve said it before, but I maintain that if the source of the article and the target both supported the Webmention spec, then when a piece used an image (or really any other type of media, including text) with a link, then the original source (any website, or Flickr in this case) would get a notification and could show—if they chose—the use of that media so that others in the future could see how popular (or not) these types of media are.
Has anyone in the IndieWeb community got examples of this type of attribution showing on media on their own websites? Perhaps Jeremy Keith or Kevin Marks who are photographers and long time Flickr users?
Incidentally I’ve also mentioned using this notification method in the past as a means of decentralizing the journal publishing industry as part of a peer-review, citation, and preprint server set up. It also could be used as part of a citation workflow in the sense of Maria Popova and Tina Roth Eisenberg‘s Curator’s Code[1]set up, which could also benefit greatly now with Webmention support.
Annotated on March 09, 2020 at 12:18PM
Charlotte Allen wrote up a method for using Micropub and Webhooks to PESOS to a WordPress website recently.
I’ll note, hopefully for ease of implementation, that a Micropub solution will already allow you to post to WordPress, Drupal, WithKnown, Craft, Jekyll, Kirby, Hugo, Blot, Micro.blog and others.
There is also an open source project called Silo.pub that provides a micropub endpoint for services like Tumblr, WordPress.com, Blogger, and Twitter (among others). Aaron Parecki has a public version I’m sure he wouldn’t mind if you tried.
Other platforms could quickly allow the functionality and so much more by building their own micropub servers, which would be a major boon to the social media space and the open web.
If you have questions about implementation while building, feel free to pop into the IndieWeb #dev chat (where prior implementers and others) are available for help. (Alternate chat modalities including Slack and IRC are available if you prefer.)
Another great option is Kevin Marks‘ NoterLive which lets you Tweetstorm away and cut/paste (PESOS) the stream to your website (with appropriate mark up when done). Other ideas at: https://indieweb.org/tweetstorm