Create videos enriched with interactions
Project Naptha automatically applies state-of-the-art computer vision algorithms on every image you see while browsing the web. The result is a seamless and intuitive experience, where you can highlight as well as copy and paste and even edit and translate the text formerly trapped within an image.
What the hand dare seize the ﬁre? ❧
I find it so heartening that one can use Project Naptha to highlight, copy and paste, and even edit and translate text formerly trapped within an image.
I’m further impressed that it also works with Hypothes.is!
–December 01, 2019 at 09:40AM
Though upon revisiting, it seems like the text is temporarily highlighted on Hypothesis (which probably only works with Naptha installed), then disappears, and the annotation is shown as an orphan.
Apparently Naptha only acts as a middle layer to allow the OCR of the image and that without it, the fingerprinting process Hypothes.is uses can’t find it after the fact.
Perhaps Hypothes.is could recognize that the highlighted text is being supplied by a third-party layer and instead of orphaning the highlighted text, it could anchor the highlight to the associated image instead? ❧
–December 01, 2019 at 09:44AM
Naptha, its current name, is drawn from an even more tenuous association. See, it comes from the fact that “highlighter” kind of sounds like “lighter”, and that naptha is a type of fuel often used for lighters. It was in fact one of the earliest codenames of the project, and brought rise to a rather fun little easter egg which you can play with by quickly clicking about a dozen times over some block of text inside a picture. ❧
Now if only I could do this with my Hypothes.is annotations! Talk about highlighting!
–December 01, 2019 at 10:06AM
There is a class of algorithms for something called “Inpainting”, which is about reconstructing pictures or videos in spite of missing pieces. This is widely used for film restoration, and commonly found in Adobe Photoshop as the “Content-Aware Fill” feature. ❧
This reminds me of a tool called asciinema that allows highlighting text within a video.
–December 01, 2019 at 10:13AM
Live coding screencasts are way more awesome when the text can be copied. asciinema makes this possible.
This could be an awesome teaching tool, particularly for programming.
Foldable Phone, Online Civility
- The Samsung Developers Conference Keynote features a foldable phone, SmartThings IoT, and Bixby innovations.
- Android will support foldable phones.
- Google employees stage a walkout over sexual harassment
- Tim Berners-Lee's Contract for the Web
- How to encourage civility online
- YouTube Content ID
- Facebook and "White Genocide"
- Young people are deleting Facebook in droves
- Facebook's holiday pop-up store
- Everybody gets free Amazon shipping
- Amazon's new HQ2(s)
- 8 new Chromebook features
- Google Home Hub teams up with Sephora
- Ajit Pai's FCC is hopping mad about robocalls
Picks of the Week
- Jeff's Number: Black Friday home tech deals
- Stacey's Thing: Extinct cables, Alexa Christmas Lights
While most people are forced to rely on Google as their silo of choice for video and specifically live streaming video, he points out a painful single point of failure in their system with regard to copyright rules and Google’s automatic filters that could get a user/content creator permanently banned. Worse, as Leo indicates, this ban could also extend to related Google accounts (YouTube, Gmail, etc.) One is thus open to potential chilling effects of intimidation, censorship, and deplatforming.
Leo discusses the fact that he’s not as beholden to YouTube because he streams and hosts all of his content on his own website and only utilizes silos like YouTube as ancillary distribution. In IndieWeb parlance what he does is known as POSSE or Post to your Own Site, Syndicate Elsewhere and this prevents his journalism, commentary, and even his business from being ravaged by the whims of corporate entities whose rules he can’t control directly.
The discussion starts at 1:05:11 into the episode and goes for about 10 minutes for those who are interested in this particular sub-topic.
As an interesting aside, I’ll note that just a few months ago that YouTube allowed people to do embeds with several options, but they’re recently removed the option to prevent their player from recommending additional videos once you’re done. Thus the embedding site is still co-opted to some extent by YouTube and their vexing algorithmic recommendations.
In a similar vein audio is also an issue, but at least an easier and much lower bandwidth one. I’ve been running some experiments lately on my own website by posting what I’m listening to on a regular basis as a “faux-cast” and embedding the original audio. I’ve also been doing it pointedly as a means of helping others discover good content, because in some sense I can say I love the most recent NPR podcast or click like on it somewhere, but I’m definitely sure that doesn’t have as much weight or value as my tacitly saying, “I’ve actually put my time and attention on the line and actually listened to this particular episode.” I think having and indicating skin-in-the-game can make a tremendous difference in these areas. In a similar vein, sites like Twitter don’t really have a good bookmarking feature, so readers don’t know if the sharing user actually read any of an article or if it was just the headline. Posting these things separately on my own site as either reads or bookmarks allows me to differentiate between the two specifically and semantically, both for others’ benefit as well as, and possibly most importantly, for my own (future self).
Shindig is a turnkey solution for online video chat events. Its unique technology offers the dynamics of an in-person event at internet scale. Shindig enables a host to give a video conference, lecture, seminar, interview or media event in front of an online audience of thousands. Hosts can share the stage for face-to-face interactions with audience members before the entire gathering or sidebar with participants privately. Meanwhile, unlike other video conference or webinar meeting technologies, audience members themselves are also able to network, discuss and socialize with one another in their own self initiated private video chats just as they would naturally at in person events. Prominent organizations and individuals ranging from CNN, The Economist, Forbes and Conde Nast Publications to Bill Gates, Sheryl Sandberg and Jim Cramer are among Shindig's early adopters.
Create interactive video stories on Timelinely. Timelinely empowers people to go beyond just video.
Highlight interesting parts of a video on a timeline with interactive comments, pictures, links, maps, other videos, and more.
We create almost everything on the internet, but we control almost none of it.
As time passes, I fear that more and more of what happened in those days will live only in memory. The internet has slowly unraveled since 2011: Image-hosting sites went out of business, link shorteners shut down, tweets got deleted, and YouTube accounts were shuttered. One broken link at a time, one of the most heavily documented historical events of the social media era could fade away before our eyes.
Text, as the Hypothesis annotation client understands it, is HTML, or PDF transformed to HTML. In either case, it’s what you read in a browser, and what you select when you make an annotation. What’s the equivalent for audio and video? It’s complicated because although browsers enable us to select passages of text, the standard media players built into browsers don’t enable us to select segments of audio and video. It’s trivial to isolate a quote in a written document. Click to set your cursor to the beginning, then sweep to the end. Now annotation can happen. The browser fires a selection event; the annotation client springs into action; the user attaches stuff to the selection; the annotation server saves that stuff; the annotation client later recalls it and anchors it to the selection. But selection in audio and video isn’t like selection in text. Nor is it like selection in images, which we easily and naturally crop. Selection of audio and video happens in the temporal domain. If you’ve ever edited audio or video you’ll appreciate what that means. Setting a cursor and sweeping a selection isn’t enough. You can’t know that you got the right intro and outro by looking at the selection. You have to play the selection to make sure it captures what you intended. And since it probably isn’t exactly right, you’ll need to make adjustments that you’ll then want to check, ideally without replaying the whole clip.
I suspect that media fragments experimenters like Aaron Parecki, Marty McGuire, Kevin Marks, and Tantek Çelik will appreciate what he’s doing and will want to play as well as possibly extend it. I’ve already added some of the outline to the IndieWeb wiki page for media fragments (and a link to fragmentions) which has some of their prior work.
I too look forward to a day where web browsers have some of this standardized and built in as core functionality.
Highlights, Quotes, & Marginalia
This selection tool has nothing intrinsically to do with annotation. It’s job is to make your job easier when you are constructing a link to an audio or video segment.
(If I were Virginia Eubanks I might want to capture the pull quote myself, and display it on my book page for visitors who aren’t seeing it through the Hypothesis lens.)
Of course, how would she know that the annotation exists? Here’s another example of where adding webmentions to Hypothesis for notifications could be useful, particularly when they’re more widely supported. I’ve outlined some of the details here in the past: http://boffosocko.com/2016/04/07/webmentions-for-improving-annotation-and-preventing-bullying-on-the-web/
What concerns me is that this is just one aspect of a kind of infrastructural violence being done to all of us, all of the time, and we’re still struggling to find a way to even talk about it, to describe its mechanisms and its actions and its effects.
I think this fits the definition of a Weapon of Math Destruction.
It's been nearly four months since the W3C held the most controversial vote in its decades-long history of standards-setting: a vote where accessibility groups, security experts, browser startups, public interest groups, human rights groups, archivists, research institutions and other worthies went up against trillions of dollars' worth of corporate muscle: the world's largest electronics, web, and content companies in a battle for the soul of the open web.
The underlying mathematics of what he’s discussing are fantastic (though he doesn’t go into them in his book), but the overarching implications of his ideas with relation to the future of humankind as a function of our economic system and society could have some significant impact.
“César visits the RSA to present a new view of the relationship between individual and collective knowledge, linking information theory, economics and biology to explain the deep evolution of social and economic systems.
In a radical rethink of what an economy is, one of WIRED magazine’s 50 People Who Could Change the World, César Hidalgo argues that it is the measure of a nation’s cultural complexity – the nexus of people, ideas and invention – rather than its GDP or per-capita income, that explains the success or failure of its economic performance. To understand the growth of economies, Hidalgo argues, we first need to understand the growth of order itself.”