tl;dr: use this script to build a GitHub Repo like this one where you store all the highlights from your kindle books in an organized way. Kindle sucks, kindle is great We all love reading in our Kindle. You can travel with more than one book at the time, you can search for words you don’t underst...
Python For Data Analysis
Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow is by far the best book to get started with machine learning.
Introduction to Statistical Learning.
Recently, my eBook on Python programming, How To Code in Python 3, was made available as a Manifold publication. I would like to offer my perspective to the Manifold community to give some background on the work and how I believe the Manifold platform provides additional layers of value to the text through providing a place for learning and idea exchange in both university communities and broader publics.
We had our largest RSVP list to date, though some had last minute issues pop up and one sadly had trouble finding the location (likely due to a Google map glitch).
Angelo and Chris met before the quiet writing hour to discuss some general planning for future meetings as well as the upcoming IndieWebCamp in LA in November. Details and help for arrangements for out of town attendees should be posted shortly.
Notes from the “broadcast” portion of the meetup
Chris Aldrich (co-organizer)
- Still working on a workflow for owning all of his reading related data, particularly with respect to POSSE to GoodReads.com
- Registered for the upcoming Dodging the Memory Hole 2016: Saving Online News on October 13/14 at UCLA which has the flavor of IndieWeb as well as the Decentralized Web movements.
Angelo Gladding (co-organizer)
- Work is proceeding nicely on the overall build of Canopy
- Discussed an issue with expanding data for social network in relation to events and potentially expanding contacts based on event attendees
Srikanth Bangalore (our host at Yahoo!)
- Discussed some of his background in coding and work with Drupal and WordPress.
- His personal site is https://srib.us/
Notes from the “working” portion of the meetup
We sketched out a way to help Srikanth IndieWeb-ify not only his own site, but to potentially help do so for Katie Couric’s Yahoo! based news site along with the pros/cons of workflows for journalists in general. We also considered some potential pathways for potentially bolting on webmentions for websites (like Tumblr/WordPress) which utilize Disqus for their commenting system. We worked through the details of webmentions and a bit of micropub for his benefit.
Srikanth discussed some of the history and philosophy behind why Tumblr didn’t have a more “traditional” native commenting system. The point was generally to socially discourage negativity, spamming, and abuse by forcing people to post their comments front and center on their own site (and not just in the “comments” of the receiving site) thereby making the negativity be front and center and redound to their own reputation rather than just the receiving page of the target. Most social media related sites hide (or make hard to search/find) the abusive nature of most users, while allowing them to appear better/nicer on their easier-to-find public facing persona.
Before closing out the meeting officially, we stopped by the front lobby where two wonderful and personable security guards (one a budding photographer) not only helped us with a group photo, but managed to help us escape the parking lot!
I think it’s agreed we all had a great time and look forward to more progress on projects, more good discussion, and more interested folks at the next meeting. Srikanth was so amazed at some of the concepts, it’s possible that all of Yahoo! may be IndieWeb-ified by the end of the week. 🙂
We hope you’ll join us next month on 10/05! (Details forthcoming…)
Live Tweets Archive
Ever with grand aspirations to do as good a job as the illustrious Kevin Marks, we tried some livetweeting with Noterlive. Alas the discussion quickly became so consuming that the effort was abandoned in lieu of both passion and fun. Hopefully some of the salient points were captured above in better form anyway.
(This CMS comment may have been the biggest laugh of the night, though the tone captured here (and the lack of context), doesn’t do the comment any justice at all.)
We met at Charlie’s Coffee House, 266 Monterey Road, South Pasadena, CA, where we stayed until closing at 8:00. Deciding that we hadn’t had enough, we moved the party (South Pasadena rolls up their sidewalks early) over to the local Starbucks, 454 Fair Oaks Ave, South Pasadena, CA where we stayed until they closed at 11:00pm.
Quiet Writing Hour
Angelo manned the fort alone with aplomb while building intently. If I’m not mistaken, he did use my h-card to track down my phone number to see what was holding me up, so as they say in IRC: h-card++!
Introductions and Demonstrations
Needing no introductions this week, Angelo launched us off with a relatively thorough demo of his Canopy platform which he’s built from the ground up in python! Starting from an empty folder on a host with a domain name, he downloaded and installed his code directly from Github and spun up a completely new version of his site in under 2 minutes. In under 20 minutes of some simple additional downloads and configuration of a few files, he also had locations, events, people and about modules up and running. Despite the currently facile appearance of his website, there’s really a lot of untapped power in what he’s built so far. It’s all available on Github for those interested in playing around; I’m sure he’d appreciate pull requests.
Along the way, I briefly demoed some of the functionality of Kevin Marks’ deceptively powerful Noterlive web app for not only live tweeting, but also owning those tweets on one’s own site in a simple way after the fact (while also automatically including proper markup and microformats)! I also ran through some of the overall functionality of my Known install with a large number of additional plugins to compare and contrast UX/UI with respect to Canopy.
We also discussed a bit of Angelo’s recentnetwork crawling project, and I took the opportunity to fix a bit of the representative h-card on my site. (Angelo, does a new crawl appear properly on lahacker.net now?)
Before leaving Charlie’s we did manage to remember to take a group photo this time around. Not having spent enough time chatting over the past few weeks, we decamped to a local Starbucks and continued our conversation along with some addition brief demos and discussion of other itches for future building.
We also spent a few minutes discussing the upcoming IndieWebCamp LA logistics for November as well as outreach to the broader Los Angeles area dev communities. If you’re interested in attending, please RSVP. If you’d like to volunteer or help sponsor the camp, please don’t hesitate to contact either of us. I’m personally hoping to attend DrupalCamp LA this weekend while wearing a stylish IndieWebCamp t-shirt that’s already on its way to me.
In keeping with the schedule of the broader Homebrew movement, so we’re already committed to our next meeting on September 7. It’s tentatively at the same location unless a more suitable one comes along prior to then. Details will be posted to the wiki in the next few days.
Thanks for coming everyone! We’ll see you next time.
Live Tweets Archive
Though not as great as the notes that Kevin Marks manages to put together, we did manage to make good use of noterlive for a few supplementary thoughts:
On my way to Homebrew Website Club Los Angeles in moments. http://stream.boffosocko.com/2016/homebrew-website-club-la-2016-08-24 #
I’ve torn some things down, but slowly rebuilding. I’m just minutes away from rel-me to be able to log into wiki #
My github was receiving some autodumps from a short-lived indieweb experiment. #
is describing his canopy system used to build his site #
Canopy builds in a minute and 52 secs… inside are folders roots and trunk w/ internals #
Describing how he builds in locations to Canopy #
Apparently @t has a broken certificate for https, so my parser gracefully falls back to http instead. #