Bookmarked CSS Stats (cssstats.com)

Writing CSS is hard, especially at scale.

CSS Stats provides analytics and visualizations for your stylesheets. This information can be used to improve consistency in your design, track performance of your app, and diagnose complex areas before it snowballs out of control.

Hat tip:

Some sketch thoughts about OER to come back and revisit

Does anyone/organization maintain a wiki or centralized repository of OER textbooks? (Especially a consortium of institutions which provide financial support).

It should contain a list of people/departments who’ve adopted (an indicator of quality).

It could maintain lists of people with technical expertise that can help to reshuffle pieces or allow customization? Maybe create easier methods for customization with related UI.

How to best curate resources and put them into a searchable repository for easy later use?

Can we create an organization that somewhat models the instutionalization of traditional textbook publishers that organize and track their assets? This institution should be supported by a broad array of colleges and universities as a means of supporting the otherwise invisible labor that is otherwise going on.

How can we flip the script to allow students to choose their own materials instead of allowing professors to do this? Their economic pressure alone will dramatically help the system. (Especially the hidden labor issues.)

 

 

Read Building a personal knowledge base by Kirill Maltsev (kirillmaltsev.net)
I try to unload all information that has any meaning to me from my brain to external storage because I don’t like to rely on my memory nor do I trust it. In this blog post, I’m going to describe my current approach of working with a personal knowledge base.
He’s got a solid listing of tools here, many I’ve either tried or am currently using myself. Looks like he hasn’t come across Roam yet which I like for it’s ability to change tags across many linked pages. I wish other tools had that one killer feature.

Microsub reader clients and unread entries

I’ve become inextricably ensnared by the wealth of awesome Microsub feed readers out there. However, one small piece of UI keeps rearing its ugly head as I move variably from one to another either to move from mobile to desktop or just to enjoy the variety of user interfaces available.

While they all do a fantastic job of keeping track of what I’ve read or left unread, many of them are missing the ability to explicitly ask for just the unread items in particular channels. Invariably, I’ll find one or two pieces that I want to leave unread to revisit later, but then finding them in a stream of hundreds later becomes an impossible task.

Aaron Parecki‘s model version in Monocle has a handy menu item to request just the unread items in a channel.

Screencapture of the Monocle UI
Monocle has a simple dropdown to allow me to see just the unread entries.

 

Either it’s missing or I’m not able to easily find the same functionality in Kristof De Jaeger‘s Indigenous for Android or Grant Richmond‘s Together. (I’ve yet to have time to try out some of the others.)

I suppose I should simply start bookmarking those pieces I still want to read later and rely on my site for the memory. Of course this also then makes me itch for having private feeds in these readers to find my unpublished bookmarks for reading via my favorite Microsub clients on a future date.

Read settings page in Indigenous

I’ve noticed that Indigenous for Android does have the ability to create an additional channel for all unread items. This seems useful while I’ve only got a few dozen feeds and a handful of channels, but I don’t expect it to be quite as useful when I’ve moved over several dozen channels with hundreds of feeds. The benefit is that it does replicate the sort of functionality that most social silos like Facebook and Twitter have of an unending stream of unread posts. 

Indigenous also allows one to either manually mark items as read individually or automatically mark them read a page at a time. The page at a time seems to clear out the entire channel rather than marking things read as they’re scrolled, so it’s a bit too broad for my taste. Monocle does a much better job at this marking read while scrolling functionality. Indigenous also says it has a “Mark all read” button per channel, but somehow I’m not seeing it in the UI despite the many ways I toggle the options.

Indigenous also has the ability to set a Read later channel, which seems useful. There is another setting for “Move items” that indicates one can move posts from one channel to another, but when choosing individual posts to move, the UI reads “Select channel to add the feed to”. I was leery at first because I didn’t want to move my entire feed to the new channel, but after  trying it there’s a pop up that said “Post moved to channel X”. Perhaps Kristof might change the word “feed” to “post” in that part of the interface? Sadly though, I have to report that looking at my Unread items channel doesn’t actually show the things that were to have been moved.

 

 

Read Twitter Archiving Google Spreadsheet TAGS v5 by Martin Hawksey (MASHe)
For a couple of years I’ve been sharing a Google Sheet template for archiving searches from Twitter. In September 2012 Twitter announced the release of a new version of their API (the spreads…
This looks like a cool little tool for archiving content from Twitter. There’s apparently a newer version out too.
Bookmarked Blogroll by Dan MacKinlay (danmackinlay.name)

Make your own automatic blogroll

This is the script I use to generate a blogroll from my OPML:

#! /usr/bin/env python3
"""
Parse OPML into markdown.
"""
import sys
import re
from xml.etree import ElementTree


def main(fname):
    with open(fname, 'r', encoding='utf8') as fp:
        tree = ElementTree.parse(fp)
    for cat_node in tree.find('body').findall('outline'):
        print("\n## {}\n".format(cat_node.get('title')))
        for node in cat_node.findall('outline'):
            name = node.attrib.get('text')
            feedurl = node.attrib.get('xmlUrl')
            url = node.attrib.get('htmlUrl')
            print("* [{}]({}) ([feed]({}))".format(name, url, feedurl))


if __name__ == "__main__":
    main(*sys.argv[1:])
Read Get your RSS feeds and podcasts in one place with this open source tool (Opensource.com)
Last year, I brought you 19 days of new (to you) productivity tools for 2019. This year, I'm taking a different approach: building an environment that will allow you to be more productive in the new year, using tools you may or may not already be using.
Read Finding phrases that match the syllable stress pattern of Teenage Mutant Ninja Turtles by Matt Maldre (Spudart)
Have you seen this tool that makes any phrase into the Teenage Mutant Ninja Turtles logo? GEEEEEEEE! Just type any words in, and MAGICALLY your words transform into the TMNT logo. It’s incredible. I love it love it love it. Here’s Super Duper Fun Potatoes. Why does this incredible tool exist? This tool came about …
Why is the ancient art of meter seemingly so lost in our language now? Apparently we need to bring back the trivium and quadrivium into our educational system.

Can’t wait to play with this generator…

How to follow the complete output of journalists and other writers?

In a digital era with a seemingly ever-decreasing number of larger news outlets paying journalists and other writers for their work, the number of working writers who find themselves working for one or more outlets is rapidly increasing. 

This is sure to leave journalists wondering how to better serve their own personal brand either when they leave a major publication for which they’ve long held an association (examples: Walt Mossberg leaving The New York Times or Leon Wieseltier leaving The New Republic)  or alternatively when they’re just starting out and writing for fifty publications and attempting to build a bigger personal following for their work which appears in many locations (examples include nearly everyone out there).

Increasingly I find myself doing insane things to try to follow the content of writers I love. The required gymnastics are increasingly complex to try to track writers across hundreds of different outlets and dozens of social media sites and other platforms (filtering out unwanted results is particularly irksome). One might think that in our current digital media society, it would be easy to find all the writing output of a professional writer like Ta-nehisi Coates, for example, in one centralized place.

I’m also far from the only one. In fact, I recently came across this note by Kevin:

I wish there was a way to subscribe to writers the same way you can use RSS. Obviously twitter gets you the closest, but usually a whole lot more than just the articles they’ve written. It would be awesome if every time Danny Chau or Wesley Morris published a piece I’d know.

The subsequent conversation in his comments or  on Micro.blog (a fairly digital savvy crowd) was less than heartening for further ideas.

As Kevin intimates, most writers and journalists are on Twitter because that’s where a lot of the attention is. But sadly Twitter can be a caustic and toxic place for many. It also means sifting through a lot of intermediary tweets to get to the few a week that are the actual work product articles that one wants to read. This also presumes that one’s favorite writer is on Twitter, still using Twitter, or hasn’t left because they feel it’s a time suck or because of abuse, threats, or other issues (examples: Ta-Nehisi Coates, Lindy West, Sherman Alexie). 

What does the universe of potential solutions for this problem currently look like?

Potential Solutions

Aggregators

One might think that an aggregation platform like Muck Rack which is trying to get journalists to use their service and touts itself as “The easiest, unlimited way to build your portfolio, grow your following and quantify your impact—for free” might provide journalists the ability to easily import their content via RSS feeds and then provide those same feeds back out so that their readers/fans could subscribe to them easily. How exactly are they delivering on that promise to writers to “grow your following”?!

An illustrative example I’ve found on Muck Rack is Ryan O’Hanlon, a Los Angeles-based writer, who writes for  a variety of outlets including The Guardian, The New York Times, ESPN, BuzzFeed, ESPN Deportes, Salon, ESPN Brasil, FiveThirtyEight, The Ringer, and others. As of today they’ve got 410 of his articles archived and linked there. Sadly, there’s no way for a fan of his work to follow him there. Even if the site provided an RSS feed of titles and synopses that forced one to read his work on the original outlet, that would be a big win for readers, for Ryan, and for the outlets he’s writing for–not to mention a big win for Muck Rack and their promise.

I’m sure there have to be a dozen or so other aggregation sites like Muck Rack hiding out there doing something similar, but I’ve yet to find the real tool for which I’m looking. And if that tool exists, it’s poorly distributed and unlikely to help me for 80% of the writers I’m interested in following much less 5%.

Author Controlled Websites

Possibly the best choice for everyone involved would be for writers to have their own websites where they archive their own written work and provide a centralized portfolio for their fans and readers to follow them regardless of where they go or which outlet they’re writing for. They could keep their full pieces privately on the back end, but give titles, names of outlets, photos, and synopses on their sites with links back to the original as traditional blog posts. This pushes the eyeballs towards the outlets that are paying their bills while still allowing their fans to easily follow everything they’re writing. Best of all the writer could own and control it all from soup to nuts.

If I were a journalist doing this on the cheap and didn’t want it to become a timesuck, I’d probably spin up a simple WordPress website and use the excellent and well-documented PressForward project/plugin to completely archive and aggregate my published work, but use their awesome forwarding functionality so that those visiting the URLs of the individual pieces would be automatically redirected to the original outlet. This is a great benefit for writers many of whom know the pain of having written for outlets that have gone out of business, been bought out, or even completely disappeared from the web. 

Of course, from a website, it’s relatively easy to automatically cross-post your work to any number of other social platforms to notify the masses if necessary, but at least there is one canonical and centralized place to find a writer’s proverbial “meat and potatoes”. If you’re not doing something like this at a minimum, you’re just making it hard for your fans and failing at the very basics of building your own brand, which in part is to get even more readers. (Hint, the more readers and fans you’ve got, the more eyeballs you bring to the outlets you’re writing for, and in a market economy built on clicks, more eyeballs means more traffic, which means more money in the writer’s pocket. Since a portion of the web traffic would be going through an author’s website, they’ll have at least a proportional idea of how many eyeballs they’re pushing.)

I can’t help but point out that even some who have set up their own websites aren’t quite doing any of this right or even well. We can look back at Ryan O’Hanlon above with a website at https://www.ryanwohanlon.com/. Sadly he’s obviously let the domain registration lapse, and it has been taken over by a company selling shoes. We can compare this with the slight step up that Mssr. Coates has made by not only owning his own domain and having an informative website featuring his books, but alas there’s not even a link to his work for The Atlantic or any other writing anywhere else. Devastatingly his RSS feed isn’t linked, but if you manage to find it on his website, you’ll be less-than-enthralled by three posts of Lorem ipsum from 2017. Ugh! What has the world devolved to? (I can only suspect that his website is run by his publisher who cares about the book revenue and can’t be bothered to update his homepage with events that are now long past.)

Examples of some journalists/writers who are doing some interesting work, experimentation, or making an effort in this area include: Richard MacManus,  Marina Gerner, Dan Gillmor, Jay RosenBill Bennett, Jeff JarvisAram Zucker-Scharff, and Tim Harford

One of my favorite examples is John Naughton who writes a regular column for the Guardian. He has his own site where he posts links, quotes, what he’s reading, his commentary, and quotes of his long form writing elsewhere along with links to full pieces on those sites. I have no problem following some or all of his output there since his (WordPress-based) site has individual feeds for either small portions or all of it. (I’ve also written a short case study on Ms. Gerner’s site in the past as well.)

Newsletters

Before anyone says, “What about their newsletters?” I’ll admit that both O’Hanlon and Coates both have newsletters, but what’s to guarantee that they’re doing a better job of pushing all of their content though those outlets? Most of my experience with newsletters would indicate that’s definitely not the case with most writers, and again, not all writers are going to have newsletters, which seem to be the flavor-of-the month in terms of media distribution. What are we to do when newsletters are passé in 6 months? (If you don’t believe me, just recall the parable of all the magazines and writers that moved from their own websites or Tumblr to Medium.com.)

Tangential projects

I’m aware of some one-off tools that come close to the sort of notifications of writers’ work that might be leveraged or modified into a bigger tool or stand alone platform. Still, most of these are simple uni-taskers and only fix small portions of the overall problem.

Extra Extra

Savemy.News

Ben Walsh of the Los Angeles Times Data Desk has created a simple web interface at www.SaveMy.News that journalists can use to quickly archive their stories to the Internet Archive and WebCite. One can log into the service via Twitter and later download a .csv file with a running list of all their works with links to the archived copies. Adding on some functionality to add feeds and make them discoverable to a tool like this could be a boon.

Granary

Ryan Barrett has a fantastic open source tool called Granary that “Fetches and converts data between social networks, HTML and JSON with microformats2, ActivityStreams 1 and 2, AtomRSSJSON Feed, and more.” This could be a solid piece of a bigger process that pulls from multiple sources, converts them into a common format, and outputs them in a single subscribe-able location.

Splash page image and social logos from Granary.io

SubToMe

A big problem that has pushed us away from RSS and other formatted feed readers is providing an easy method of subscribing to content. Want to follow someone on Twitter? Just click a button and go. Wishing it were similar for a variety of feed types, Julien Genestoux‘s SubToMe has created a universal follow button that allows a one-click subscription option (with lots of flexibility and even bookmarklets) for following content feeds on the open web.

Splash image on SubToMe's home page

Others?

Have you seen any other writers/technologists who have solved this problem? Are there aggregation platforms that solve the problem in reverse? Small pieces that could be loosely joined into a better solution? What else am I missing?

How can we encourage more writers to take this work into their own hands to provide a cleaner solution for their audiences? Isn’t it in their own best interest to help their readers find their work?

I’ve curated portions of a journalism page on  IndieWeb wiki to include some useful examples, pointers, and resources that may help in solving portions of this problem. Other ideas and solutions are most welcome!

Read How to Clear Chrome Cache for Specific Website Only (3 Steps) (PIT Designs)
When browsing a website, Google Chrome (or any other browser) stores some data for speeding up the website pages load, in what's called Cache. This by default i
This is a clever method for quickly clearing cache for a particular page.