How the Washington Post used data and natural language processing to get people to read more news

In April, Washington Post announced that it had set a new single-month traffic record, with more than 52 million unique visitors. The figure represented not only a new record, but also a 65 percent year-over-year gain that led other big-name publishers, according to the Post.

Publisher Frederick J. Ryan praised the addition of new editorial staffers and awards, and then called special attention to engagement: While unique visitors were up 65 percent, pageviews were up 109 percent year-over-year.

The note on engagement struck me. I’d recently found myself clicking through story after story on the Washington Post site, primarily through the Post Recommends section at the bottom of each article. The accuracy of the recommendations was uncanny and so the experience stuck with me.

Then, at NICAR, I heard about a new technology that helps power Post Recommends, Clavis, and set out to learn how it works.

In a nutshell, Clavis is technology that figures out what stories are about, categorizes them by topic, and assigns each a series of keywords. It runs that same process on the Post’s readers and identifies their presumed interests based on stories they’ve read. Clavis then pairs readers with stories that match their reading history.

A list of keywords and topics Clavis generated on a recent story about the perfume industry.

Clavis is just one part of the larger Post Recommends system and has helped the Post improve its click through rate on Post Recommends 95 percent since last June. Clavis also offers a good look at the process behind a recommender build (including natural language processing) and the challenges and opportunities news organizations face when building one.

Building Clavis

Work on Clavis began about a year ago with a brainstorming session between developer Gary Fenstamaker and digital product analyst Angela Wong. At the time the two only sought to build a story classifier that would help automate the process of adding SEO keywords to stories, but the technology they developed eventually became a full recommender.

News organizations are finally able to do something with all the words that they’ve written and published over the years.

Despite the tech savvy of the builders, Fenstamaker and Wong started the project with the rather low-tech strategy of talking to section editors about topics the Washington Post covers.

The challenge was to create a list of topics that would be thorough enough that every Post story could easily be assigned a variety of topics, but flexible enough that the system would be able to adapt to new coverage areas over many years.

“[We] wanted to build something foundational that wouldn’t shift around a lot,” Wong said.

That meant that some topics that seemed like obvious candidates for inclusion — the Iraq War for example — would be dropped for something more general.

“We didn’t want ‘Al Qaeda,’” Fenstamaker said. “We wanted ‘terrorism groups.’”

The logic was that while Al Qaeda would eventually fade from the news, the general topics of terrorism and terrorism groups are likely to remain.

The pair eventually settled on a list of about 130 topics that could be assigned to each one of the more than 700 stories the Post publishes each day — a task that is easier said than done.

“It’s hard to teach a machine to understand words,” Wong said. “How are you going to teach a computer about why this (particular) is so significant?”

While it’s simple enough to tell a computer make a list of the all the words in a story and perhaps provide a sum of each, how do you teach it to ignore “the,” “this,” and “good,” which are very common, but aren’t very telling about a story’s main ideas?

The answer for the Washington Post team was in a set of common machine learning algorithms.

Mapping new articles

Like other content-based recommenders, Clavis maps articles by running them through a term-frequency, inverse document frequency (tf-idf) algorithm to figure out which words in a story are important, and thus a likely indicator of what the story is about.

For the non-nerds among us, tf-idf basically counts how many times a word appears in an article and compares that number to the number of times it appears in other stories in a given set. The fewer articles a word appears in, the more importance it is assigned in the current article.

In addition to the standard tf-idf algorithm, Clavis also adds extra weight proper nouns and more weight still if those nouns they appear earlier in a story, Fenstamaker said.

Every time the Post publishes a new story, Clavis is notified and then processes the story. To lighten the load on systems, it starts recommending a new batch of stories once an hour.

Mapping Users

Assigning topics to stories is only half of what Clavis does. The technology also classifies readers in a similar fashion, assigning each reader topics and a group of keywords based on the stories they read.

The site’s registered users are obviously easier to assign topics to, but the Washington Post can recommend relevant content to readers who only visit occasionally by using cookies.

“It’s definitely harder with anonymous users,” Fenstamaker said, acknowledging that relying on cookies is “highly volatile” and that the Post’s picture of unregistered users decays over time as people’s interests change.

A hybrid approach

Clavis is what is known as a content-based recommendation engine, which means it looks at the content of a piece of media (or metadata associated with it), figures out what the content is about and then recommends similar pieces of media. By way of example, Pandora uses a content-based recommendation engine and plays songs for uses based on their similarity to songs it knows they like.

The other primary type of recommendation engine is a collaborative filter, which assumes that two users who have read similar stories in the past will read similar stories in the future and forms a recommendation based on that history. A collaborative filter requires a fair amount of data, but doesn’t require the algorithm to “understand” anything about the thing it’s recommending. The most well known example of a collaborative filter is likely Amazon’s “People who viewed this item also viewed this item” recommendation.

Each approach has specific strengths and weaknesses. A collaborative filter, for example, faces what’s known as the “cold start” problem — meaning that it requires a history and data to learn from. In the news business, that’s a problem because new stories are published constantly and a machine takes too long to collect data on a new story.

On the other hand, collaborative filters don’t have to know anything about what a piece of media is about to make decent recommendations — an attribute that is particularly helpful when it comes to things that are difficult for a machine to understand like a video, photo, or piece of audio.

The Washington Post ended up relying on a hybrid system to power Post Recommends, said Sam Han, the Post’s director of big data and personalization. That system takes into account both personalization (Clavis) and popularity (collaborative filtering).

Custom recommender vs. vendor

While the Post and a few other publications have had some success analyzing data and recommending content, it’s a difficult thing for most organizations to do.

The Post has an entire data and personalization team and content recommendation is one of its primary projects, said Wong.

“It’s very expensive to do all this,” Wong said. Naturally many traditional publications won’t have the resources to devote to personalization and recommendation.

“There is so much information to collect, that I don’t know that a lot of news organizations are ready to collect, analyze, and act on it,” she said.

Most publishers turn to a third-party solution like Outbrain or Taboola to keep users clicking.

But for those publishers with the ability to collect and understand data and then make smart recommenders, it can add significant value.

“News organizations are finally able to do something with all the words that they’ve written and published over the years,” Wong said.

About the author

Ryan Graff

Communications and Outreach Manager, 2011-2016

Journalism, revenue, whitewater, former carny. Recently loving some quality time @KelloggSchool.

Latest Posts

  • Prototyping Augmented Reality

    Something that really frustrates me is that, while I’m excited about the potential AR has for storytelling, I don’t feel like I have really great AR experiences that I can point people to. We know that AR is great for taking a selfie with a Pikachu and it’s pretty good at measuring spaces (as long as your room is really well lit and your phone is fully charged) but beyond that, we’re really still figuring...

    Continue Reading

  • Capturing the Soundfield: Recording Ambisonics for VR

    When building experiences in virtual reality we’re confronted with the challenge of mimicking how sounds hit us in the real world from all directions. One useful tool for us to attempt this mimicry is called a soundfield microphone. We tested one of these microphones to explore how audio plays into building immersive experiences for virtual reality. Approaching ambisonics with the soundfield microphone has become popular in development for VR particularly for 360 videos. With it,...

    Continue Reading

  • How to translate live-spoken human words into computer “truth”

    Our Knight Lab team spent three months in Winter 2018 exploring how to combine various technologies to capture, interpret, and fact check live broadcasts from television news stations, using Amazon’s Alexa personal assistant device as a low-friction way to initiate the process. The ultimate goal was to build an Alexa skill that could be its own form of live, automated fact-checking: cross-referencing a statement from a politician or otherwise newsworthy figure against previously fact-checked statements......

    Continue Reading

  • Northwestern is hiring a CS + Journalism professor

    Work with us at the intersection of media, technology and design.

    Are you interested in working with journalism and computer science students to build innovative media tools, products and apps? Would you like to teach the next generation of media innovators? Do you have a track record building technologies for journalists, publishers, storytellers or media consumers? Northwestern University is recruiting for an assistant or associate professor for computer science AND journalism, who will share an appointment in the Medill School of Journalism and the McCormick School...

    Continue Reading

  • Introducing StorylineJS

    Today we're excited to release a new tool for storytellers.

    StorylineJS makes it easy to tell the story behind a dataset, without the need for programming or data visualization expertise. Just upload your data to Google Sheets, add two columns, and fill in the story on the rows you want to highlight. Set a few configuration options and you have an annotated chart, ready to embed on your website. (And did we mention, it looks great on phones?) As with all of our tools, simplicity...

    Continue Reading

  • Join us in October: NU hosts the Computation + Journalism 2017 symposium

    An exciting lineup of researchers, technologists and journalists will convene in October for Computation + Journalism Symposium 2017 at Northwestern University. Register now and book your hotel rooms for the event, which will take place on Friday, Oct. 13, and Saturday, Oct. 14 in Evanston, IL. Hotel room blocks near campus are filling up fast! Speakers will include: Ashwin Ram, who heads research and development for Amazon’s Alexa artificial intelligence (AI) agent, which powers the...

    Continue Reading

Storytelling Tools

We build easy-to-use tools that can help you tell better stories.

View More