Model Blog: Chris Moody of Stitch Fix

30 Oct 2017 . category: Blogging . Comments

http://multithreaded.stitchfix.com/blog/

As I find resources online to learn more and more about word embeddings1, I keep coming back to one resource in particular.

It started in September with the discovery of “A Word is Worth a Thousand Vectors” (and the corresponding Text By the Bay 2015 talk + slides). Then a week later, I found a link to “Introducing our Hybrid lda2vec Algorithm” through a forums.fast.ai search2 for more on LDA3.

The author for both incredibly detailed yet accessible blog posts is Chris Moody (“Caltech - Astrostats - PhD supercomputing. Now Applied AI”).

Despite his deeply technical background in things like whatever astrostatistics is, I admire Moody’s commitment to explaining things well. He provides both the bird’s eye view and the implementation details (the latter of which I happen to be very interested in at the moment). He illustrates his ideas with examples and lots of appealing, brightly-colored visualizations. His references sections at the end of posts are extraordinarily helpful unto themselves4. All the while, he never seems to dumb his content down, and so the posts stand as complete works on their own.

The biggest point I’d like to take away from Moody’s blogging is the importance of visual language. The more I blog, the more I feel like I use too many words per post. I want to push myself to stitch together a more visual portfolio of work5.

Footnotes

  1. Word embeddings are a neat way of encoding semantic relationships between words into numeric vectors, which is important because these vectors are what machine learning models accept as input. At the moment, word embeddings appear to be the #1 preferred building block for natural language processing (NLP) in machine learning. 

  2. LDA stands for Latent Dirichlet Allocation, a technique for topic modeling. lda2vec is a word embeddings + topic modeling extension… invented by Chris Moody! 

  3. For example, “A Word is Worth a Thousand Vectors” had a reference to “Dependency-Based Word Embeddings” by Omer Levy - which I wasn’t looking for but in the end found very useful while thinking through the work behind my own NLP for Task Classification post. 

  4. Visual work like the Shiny app and other charts that I replicated earlier this month from David Robinson of StackOverflow, as an example. 


Me

Nadja does not particularly enjoy writing about herself.