I’ve found in the past that highlighting on Chrome for Android was nearly impossible. I’ve switched to using Firefox when I need to use hypothes.is on mobile.
In this episode, we talk about Audrey's decision to block annotations from her websites.
A great little episode that talks about annotations, but impinges on a lot of issues relating to ownership and running of a (personal) website. It discusses a lot of things that folks on the web often take for granted, but which should really be done with a lot more civility and care for others.
While I have been seeing some really great and thoughtful conversations, particularly in the last 6 months or more, as a result of people posting on their own websites and using open standards like webmentions to carry on conversations, the rest of the internet still needs to take great strides to improve itself. A lot of these issues are ones of bias, and particularly of the white male sort, but I think that the ideas encapsulated in this short podcast will help to open people’s eyes. While it’s sad that Audrey had these experiences on Twitter, I’m glad that she and Kin took the time to discuss them here in hopes of improving the space in the future for others.
There are potential solutions to the recent News Genius-gate incident, and simple notifications can go a long way toward helping prevent online bullying behavior.
There has been a recent brouhaha on the Internet (see related stories below) because of bad actors using News Genius (and potentially other web-based annotation tools like Hypothes.is) to comment on websites without their owner’s knowledge, consent, or permission. It’s essentially the internet version of talking behind someone’s back, but doing it while standing on their head and shouting with your fingers in their ears. Because of platform and network effects, such rude and potentially inappropriate commentary can have much greater reach than even the initial website could give it. Naturally in polite society, such bullying behavior should be curtailed.
This type of behavior is also not too different from more subtle concepts like subtweets or the broader issues platforms like Twitter are facing in which they don’t have proper tools to prevent abuse and bullying online.
A creator receives no notification if someone has annotated their content.–Ella Dawson
Towards a Solution: Basic Awareness
I think that a major part of improving the issue of abuse and providing consent is building in notifications so that website owners will at least be aware that their site is being marked up, highlighted, annotated, and commented on in other locations or by other platforms. Then the site owner at least has the knowledge of what’s happening and can then be potentially provided with information and tools to allow/disallow such interactions, particularly if they can block individual bad actors, but still support positive additions, thought, and communication. Ideally this blocking wouldn’t occur site-wide, which many may be tempted to do now as a knee-jerk reaction to recent events, but would be fine grained enough to filter out the worst offenders.
Toward the end of notifications to site owners, it would be great if any annotating activity would trigger trackbacks, pingbacks, or the relatively newer and better webmention protocol of the W3C which comes out of the IndieWeb movement. Then site owners would at least have notifications about what is happening on their site that might otherwise be invisible to them. (And for the record, how awesome would it be if social media silos like Facebook, Twitter, Instagram, Google+, Medium, Tumblr, et al would support webmentions too!?!)
Perhaps there’s a way to further implement filters or tools (a la Akismet on platforms like WordPress) that allow site users to mark materials as spam, abusive, or “other” so that they are then potentially moved from “public” facing to “private” so that the original highlighter can still see their notes, but that the platform isn’t allowing the person’s own website to act as a platform to give safe harbor (or reach) to bad actors.
Further some site owners might appreciate gradable filters (G, PG, PG-13, R, X) so that either they or their users (or even parents of younger children) can filter what they’re willing to show on their site (or that their users can choose to see).
Consider also annotations on narrative forms that might be posted as spoilers–how can these be guarded against? For what happens when a even a well-meaning actor posts an annotation on page two which foreshadows that the butler did it thereby ruining the surprise on the last page? Certainly there’s some value in having such a comment from an academic/literary perspective, but it doesn’t mean that future readers will necessarily appreciate the spoiler. (Some CSS and a spoiler tag might easily and unobtrusively remedy the situation here?)
Certainly options can be built into the annotating platform itself as well as allowing server-side options for personal websites attempting to deal with flagrant violators and truly hard-to-eradicate cases.
Do you have a solution for helping to harden the Internet against bullies? Share it in the comments below.
- Genius Wants To Let Readers Annotate Any News Article. What Could Possibly Go Wrong? by Jessica Goldstein, ThinkProgress 2016-03-30
- Genius responds to Congresswoman Katherine Clark’s letter on preventing abuse by Noah Kulwin, Re/code 2016-03-29
- Misguided Genius by Chelsea Hassler, Slate 2016-03-28
- The Genius Problem by Chuq Von Rospach 2016-03-28
- Genius Web Annotator vs. One Young Woman With a Blog by Brady Dale, The Observer 2016-03-28