In the time of COVID-19: Take your time

This is a deeply uncertain time. Many people are reaching out through whatever tools they have at their disposal to make a human to human connection, to share their thoughts, to raise their voice, to feel less alone, less scared.

I’ve been struggling with my own uncertainties, and my own perceptions of how I “should” be in this time. While I’m not an academic, this article by Aisha S. Ahmad in particular helped me name and tame some of what I’ve been going through - I highly recommend reading the whole thing, as I reference some of it below.

What I want to say here, though, is something a little different. I want to share my hope for the world, especially the world of business and technology that I inhabit as a worker.

Which boils down to this: Take this time.

Over the past 10 - 20 years, it’s felt like our world just kept going faster and faster, with demands for productivity to always increase, always be on, working on a never-resting global schedule. Gogogogogogogo.

Until now.

Now many of us are at least on pause, if not stopped completely. Or part of our lives are waiting for resolution while others feel like they’re trying to push forward at the pre-pandemic pace.

What if, instead of trying to soldier through as if everything is normal, or as if everything is going to return to what normal used to be, we take this time to feel, be, think? So that when a new normal arrives, we can be intentional about where we go, what we do, and how we do it.

Right now, we should focus on helping our human systems, because those are the foundation of everything else. And those are the ones that are hurting the most. Start with yourself, and your family, blood or created. Are you safe? Are they safe? Do what you can to get to that sense of safety. Then, it’s your work family. Same thing - how safe are they feeling right now? And remember, safety manifests in a lot of different areas of our lives; there’s physical safety, psychological safety, food safety, housing safety … Then extend that to customers and clients - how can you help them feel safe? (Again, Aisha S. Ahmad says this way better than I do … )

This process takes time. We’re humans - we cling to fear, we magnify fear, and it takes time for us to crawl out of that hole. We all do it at different rates. Some of us are forced out of it too soon because our work calls us to put others before ourselves. The physical and psychological costs for those on the front line will be immense, and they will need our support now and into the future. Respect that. But also try to understand that some people won’t be able to move as quickly, and respect that too.

This virus moves at its own pace and right now we cannot affect it. We can, and are, changing the pace at which it spreads, but we are not changing the pace at which it incubates, develops, and causes illness. It does not respect our need for quick resolution, our desire to move out of this space of not-knowing.

Let’s use that as a forceful reminder of all the different clocks and schedules that we do control. Let’s take this time as an opportunity to figure out what we need to be our best human selves, both in isolation and together as a network, as a community, and then start taking purposeful, intentional, mindful steps towards building those bonds.

Take. Your. Time.

(For the musical theatre inclined, I leave you with the brilliance of Jason Robert Brown’s “The Schmuel Song”.)

How UX Makes Us More Human: ConveyUX 2017

(This post was originally published on L4Digital.com in March of 2017)

My ConveyUX adventure began as I elbowed my way into a crowded hotel meeting room, listening to Marianne Sweeny present on “The IA of AI,” a deep dive into both the history of AI and its future, rife with suggestions for future reading. The next three days were jam-packed with sessions, networking, and great conversations with designers, researchers, and others affiliated with or interested in UX. Topics leaned heavily toward new(er) technologies and interfaces: augmented reality (AR)/virtual reality (VR), conversational user interfaces (like Apple’s Siri), and also included data/AI, healthcare, and tactical explorations of design artifacts and processes.

As I moved from session to session, I noticed certain concepts kept popping up; presenters repeated words like “play” and “context;” there were discussions of the intermingling of the physical and the digital; and everywhere speakers rallied for deeper collaboration between technologists/developers.

UI & Our Future

While listening to a researcher talk about designing conversational UI for Amazon’s Alexa, it struck me — as we’re discovering how to make machines behave in ways that are more human, we’re starting to uncover and explore more deeply for ourselves what it means to BE human.

When talking about artificial intelligence, and what it means even to be intelligent as a human, much less artificially, Marianne Sweeney asked the question: “Is physical embodiment necessary to our experience?” As with any really good question, there’s no one, true answer, but I would say, “Yes.”

Sitting at a computer screen, moving only our fingers and our eyes as we type and click, it can be easy to separate the “mind” from the “body” and assume that one has no bearing on the other. However, studies have shown that physical movement actually helps the brain function, hinting at a deeper interconnection between thought and experience. How, then, will we be able to “teach” our algorithms to be humanly intelligent without that type of physical input/stimuli?

Where Digital & Physical Worlds Connect

The most visible interconnection of the digital and physical worlds happens with augmented reality (AR) and virtual reality (VR). Dr. Annemarie Lesage introduced an alternative to these types of spaces: mixed reality (MR). She landed on the phrase because of her experience working at Hybridlab (Universite de Montreal) with the Hyve-3D , a mixed-reality design tool that allows designers to collaborate on design concepts using 3D projection and tablets.

MR differs from AR in that there is direct physical control of the digital portion of the experience (albeit in this case through a tablet interface); and it differs from VR in that the people using the technology are still aware of their own bodies in the space.

What’s fascinating is that this technology, developed at a university and purchased by corporations, is also used in Montreal for group experiences like concerts and dance parties. This highlights one of the biggest insights from utilizing MR: immersion in an experience is not about being physically immersed or surrounded, but is rather a consequence of the quality of engagement you have with the experience. And the experience is much more engaging when it is connective (social and physical) rather than separate or isolative.

And finally, conversational UI is driving us to think more deeply about how we as humans use and understand language, because the capability to create realistic dialogue is crucial in building a conversational UI that is effective, delightful, and not creepy. Phillip Hunter, UX Design Manager for Alexa Skills at Amazon, actively encouraged designers and developers working on these types of tools to learn playwriting/scriptwriting tools and techniques to hone their ability to make better voice products (that roar of approval you heard was all those theatre and film nerds like me who can finally tell our parents, “Look! I’m using my theatre training AND making money at the same time!”).

Building a Better Future Together

Amber Case is a cyborg anthropologist, and also my new BFF. Seriously, I have proof. We first met when I asked her to sign my yearbook a copy of her newest book, “Calm Technology,” partially because I thought she shared some great insights in her talk “Designing for Privacy in Mobile and Web Apps,” and partially because because “Cyborg Anthropologist” ties with “Special Agent in Charge” as the second coolest title ever (the first, obviously, being Supreme Allied Commander).

We’re at a crucial place in time: we’re developing technologies that previously only existed in science fiction, technologies that start to meld what is human with what is machine. And we’re developing them fast, perhaps faster than our limited brains can actually process them. As Case says, “Humans have a metabolism rate for features;” swap out “anything new” for “features” and this still rings true.

The insights and observations shared at conferences like ConveyUX remind us that we can’t let our excitement for the machines blind us to the capabilities and amazing qualities of humans, and that we need all voices at the table to build a future that brings out the best in both.

On Speaking Terms with NLP

(This post was originally published on L4digital.com in November of 2017.)

TALKING LIKE HUMANS ABOUT HOW COMPUTERS “TALK”

Language is beautiful. Consider the word ‘palimpsest’—the faint ghosting that remains after a previously-written thing has been erased. ‘Palimpsest’ gives me one word to describe an idea that would otherwise take thirteen.

Language is also hard, especially when it comes to describing abstract linear concepts. This is where math is really useful. In the field of Natural Language Processing (NLP), in which computers are taught to understand and generate words in order to communicate in more human ways, math is the lingua franca.

I recently attended Global Data Geeks’ Data Day Seattle specifically to learn more about NLP, but the one-day conference covered that and more across the spectrum of data science, artificial intelligence (AI), and machine learning (ML).  (For a more thorough introduction to AI and ML, check out my colleagues’ “AI for Dummies”.)

Discussions of NLP, AI, and ML go hand in hand because a computer has to learn in order to process language. That happens when we provide a training set of data to the models or algorithms in use and then iteratively refine them until the model or algorithm generally does what you want it to do.

Given such highly technical subject matter, I wasn’t surprised that the presentations at Data Day Seattle were more mathematical than I’m used to. Words popped up repeatedly in talks and conversations that everyone around me seemed to understand, but which threw me for a bit of a loop. By using context clues—something I’m currently better at than a computer, but won’t be for long—I walked out of each session both more knowledgeable and more excited.

My experience at Data Day Seattle prompted me to create a glossary for others like me who are just starting to dive into NLP. This glossary introduces key concepts and terms in a way that (hopefully) even their grandmothers could understand.

As is the case any time you simplify something, there are nuances—many nuances—that have necessarily been left out in favor of accessibility.

COMMONLY USED NLP TERMS DEFINED IN A NON-MATHEMATICAL BUT HELPFUL WAY

Since most of the terms listed here are borrowed from general ML or linguistics, they have general meanings that you can use to help tease out their specialized meanings in the context of NLP. Rather than list the terms alphabetically, as most glossaries do, I’ve listed them by term frequency (TF), with the most frequently used words, based on the corpus of Data Day Seattle talks, defined first.

Machine learning: When given enough information (data), a computer uses algorithms and models to ‘learn’ to make predictions, rather than being told specifically what to do via a program.

Neural networkInspired by what we know about how neuron structures work in the human brain, these specific types of machine learning networks use different layers: Nodes of data, or ‘neurons’, serve as inputs, and layers of mathematical computations applied to these nodes create the output layer. The output, or ‘prediction’, is based on the strength of the connections between different nodes. Neural networks are capable of deep learning, and are what often makes the news. AlphaGo Zero, for example, recently made headlines for learning enough about the game ‘Go’, without any specific training, to beat all competitors in just three days. AlphaGo Zero even beat its older sibling, AlphaGo, 100% of the time.

A specialized form of neural network used most often in NLP is the recurrent neural network (RNN). A RNN is a network that can process sequences. What does that mean? Well, traditional neural networks see a sentence simply as a collection of words because they don’t consider the order in which the words appear. But in RNNs, the sequence of the words is retained, allowing the network to have more context and therefore a greater chance at “understanding” the sentence as a whole.

(I have put “understanding” in quotes as a nod to a point that Jonathan Mugan made during Data Day Seattle. While these amazingly powerful networks can identify information and execute actions based on data inputs, Mugan says, there is no indication that they actually understand what they’re doing. Computers can’t pun.)

Embedded/embedding: Embedding is the process of taking words or phrases from a body of work and mapping them to vectors (see next entry for a description of vectors).

Vector: Simply put, word vectors are a shortcut to providing context or additional information about one particular word or phrase. When given context or additional information, a model is able to learn more precisely and more efficiently. What that additional information is can vary.

Here’s one example.

The idea that words used in the same context often mean the same thing was popularized sixty years ago by linguist J. R. Firth in his pithy statement, “You shall know a word by the company it keeps.”

So if our training set contains both “The quick brown fox jumped over the lazy dogs” and “The quick brown vulpine jumped over the lazy dogs”, the model might determine that “fox” and “vulpine” probably mean the same thing. “Fox” and “vulpine” might then be embedded in the same vector in order to help the model more quickly process a new data set that contains either of these words.

More technically, a word vector is a mathematical space where words or phrases are mapped in relation to semantically similar words or, alternately, “are embedded nearby each other”.

Feature: As a product manager, I use the word ‘feature’ quite regularly, but ‘feature’ in the context of NLP requires a slight reframing.

Rather than being a stand-alone piece of functionality, a ‘feature’ in NLP is more closely related to the idea of a distinctive attribute; it’s a value of some parameter within the text that is determined and then used later. An example of a NLP-specific feature would be the number of times a particular word is used in a body of text, or how many proper nouns a text contains. ‘Feature extraction’ is the process by which that feature is determined and made available to the model for use.

Bag of words: This is pretty much what you’d think it is: a model wherein all the words in the document or corpus being examined are considered without context, similar to the bag of letters in a Scrabble game. It’s also fun to say.

Term Frequency – Inverse Document Frequency (TF-IDF): Term Frequency – Inverse Document Frequency (TF-IDF) is a method of determining the probable subject of a document by comparing how often any one word is used to how unique that word is across an entire body of work that includes the original document. If a word is used often but only in one document, this indicates a high likelihood that it is meaningful to that particular document (as opposed to words which occur often across all documents, like ‘the’ and ‘and’).

My favorite example of TF-IDF comes from Julia Silge. Silge examined the entire corpus of Jane Austen’s work and determined that the most frequent words in each of her books were proper names, indicating, as most Jane Austen fans would agree, that the core meaning of her books comes from her characters.

Attention: In the context of NLP, ‘attention’ refers to a model looking at previous steps or bits of information, and using that information to refine the outcome for the current step. The idea is based on human attention mechanisms, whereby a person can focus strictly on one part of an image and more loosely on the parts of the image surrounding that focal point, providing additional context about what they’re looking at. Attention is another key attribute of neural networks that helps them learn faster, and thereby shortens iteration and testing time.

N-gram: ‘N-gram’ refers to the sequence of some number (n) of items from a given document or corpus. Items can be things like letters, phonemes, or words. At the word level, for example, trigram sequences contained in this very sentence could be “sequences, contained, in”, “trigram, sequences, contained”, “contained, in, this”. By storing text in this way, the model can then compare the n-grams and better predict the sequences that it’s parsing.

SO, NOW WHAT?

Now that you’re on speaking terms with NLP, let’s consider these glossary words and concepts in context.

Almost all of these concepts focus on how to better train the model (“better” can mean making it more accurate, or reducing its training time), which is obviously a key issue for those in the field. Humans use language as a code. (Remember “palimpsest”? One word covers an idea that takes thirteen words to describe.)  With NLP, however, language is code. And it’s an exciting field because it’s in its infancy; its capabilities and influence are only going to grow.

What we shouldn’t forget is that our language, the human code, is more than just words. Our language represents underlying images and ideas, and those images and ideas are steeped in our own psychological biases, whether we choose to recognize them or not.

Computers can’t understand that, but they can parrot and reflect it. Microsoft’s Tay.ai made a splash last year when it was removed from Twitter twenty-four hours after it first appeared because it had so quickly picked up racist, sexist, and otherwise offensive language.

Many data scientists and technologists are working hard to recognize and remove inherent bias in the training data we feed these models, and are advocating that the rest of the field does the same. In Silge’s “She Giggles, He Gallops.”, she shines a light on how movie scripts reflect our assumptions on gender roles.

All of this makes me wonder: Since NLP reflects our own language back to us so that we can actually see what we’re saying, how might we learn from the very machines we’re trying to teach?

FURTHER LEARNING

For a truer, more detailed understanding of the field, I highly recommend digging into the resources available on the Internet. The NLP community is very open and collaborative, so there’s a lot of information that is freely available. Below are a few that I’ve used, but I’m sure that there are countless more. If there are any resources that you’d like to share, please do so in the comments below.

MY DATA DAY AGENDA

Below is the list of presentations that I attended at Data Day Seattle, accompanied by brief summaries of each. You can find many of the speakers on Twitter (and I encourage you to do so).

  1. Jonathon Morgan (New Knowledge / Data For Democracy): This is Our Fight: Technology for Defending Public Discourse

  2. Stefan Krawczyk (Stitch Fix): Scaling Data Science at Stitch Fix

  3. Zornitsa Kozareva (Amazon): Conversational Assistants with Deep Learning

  4. Rob McDaniel (Lingistic): Detecting Bias in News Articles

  5. Sanhgamitra Deb (Chegg): Evolution of Natural Language Comprehension with Human Machine Collaboration:

  6. Julia Silge (Stack Overflow): Text Mining Using Tidy Data Principles:

  7. Jonathan Mugan (Deep Grammar): From Natural Language Processing to Artificial Intelligence

  8. Jonathan Mugan (Data Day Seattle) https://www.slideshare.net/jmugan/data-day-seattle-from-nlp-to-ai

  9. Garrett Eastham (Data Exhaust): Bootstrapping Knowledge-Bases from Text