Feeds, tweets and APIs are the beginning. Will news orgs step up to augment reality?

In her TED talk, Unveiling the “Sixth Sense,” game-changing wearable tech, Pattie Maes demos a system that creates interactive visual layers over the real world. The actual implementation, a tiny projector tied to a wearable computer that watches your fingers for input (using colored marker caps to identify fingertips!) is cheap, but not something you’d likely want to wear to the store.

But imagine for a moment a similar system, one that detects more subtle gestures and does not physically project light onto the objects you’re manipulating. A device that annotates the real world and presents information about the person in front of you, the product you’re considering purchasing or the comparitive likelihood of catching a cab at this corner or the next block over.

Map for driving by eszter
Map for driving by eszter, based on MacGyver Tip: Heads up display with a reversed paper map from LifeHacker.

I’ve blogged about this before, but Maes’ talk reminded me how important this technology will be. It *will* happen, and although there’s much work left to do in the end user interface (Rainbow’s End by Vernor Vinge, and Counting Heads by David Marusek present brilliant visions of how they might work) the inputs to these systems are coming online today.

Feeds, tweets and APIs aren’t just for the web

Twitter, when paired with TweetDeck gives me an always-on, ambient awareness of events worldwide. Its like a tiny, quiet news radio, feeding me timely information on events I care about. When I’m at my desk, I can hear the Internet hum. Soon, that spatial restriction will be lifted.

I already use Amazon from my phone’s web browser when I’m shopping, but the APIs are there to build new, better interfaces, that, as the Maes’ demos in her talk, can port Amazon ratings and everything else into the real world.

The NYT’s and The Guardian’s new APIs are similarly useful, but present even richer information. Detailed, expert analysis of not just products, but news and events. (And surely Bittman’s recipe for Roast Chicken With Cumin, Honey and Orange would be handy to have on a heads-up display, at the grocery store, when cooking, and when you’re regaling friends with the elegant simplicity of roasting a whole bird.)

Who’s building the future?

Of the 1180 APIs cataloged at ProgrammableWeb, only 18 are categorized as “news”. If news orgs want to hang on to their last shred of credibility as the essential information providers of the last century, they’d best get on it.

APIs are the future of information, and the content creators who adopt them will augment our reality.


ChangeTracker: Tracking changes at White House web sites so you don’t have to

ChangeTracker is a little tool that updates an RSS feed, emails, and tweets when pages change at whitehouse.gov, recovery.gov and financialstability.gov. It’s also the first project of my internship at ProPublica! Woot!

@changetracker on Twitter
@changetracker on Twitter

We put to together a bit of tricky internet plumbing and massaging of the tubes. ChangeTracker tweets, emails and feeds an RSS that links to pages at Versionista (a totally awesome tool, previously blogged about) that show left-right diffs of the page changed.

Versionista co-founder Peter Bray used Versionista to highlight position changes on Barack Obama’s campaign web site.

And we’re giving away the code! So it’s dead-simple for anyone to copy our work and set up change trackers for any web site.

Needless to say, it fills me with joy that I work at such a badass non-profit newsroom.

News Mixer roundup: links and thoughts on what comes next

It’s been a while since I’ve posted about News Mixer, and since then the intertubes seem to have taken a liking to our little project. I’m delighted that our code might live on in other projects.

Steal this code!

Like I explained in my interview with Kristen Taylor at Knight Pulse, the code is free.

So take it, and make cool stuff. Please!


Using News Mixer right now

I was delighted to learn that the Populous Project decided to adopt News Mixer. Anthony Pesce’s post at Media Shift Idea Lab,Populous Is Adopting News Mixer (And More), covers the details.

Initially we were planning on building similar features into Populous, but our original vision was to create a whole separate network on our own site to handle it. That plan had a few problems, but two in particular were too large to ignore: Facebook is ubiquitous on college campuses and it does social networking better than we ever could, and new readers would have to join a whole new network which is an unacceptable impediment.

We realized that using Facebook Connect as a way of authentication for the site, and as a way of giving our readers a robust social networking experience, would almost work better than making the whole thing on our own from scratch. Facebook, we think, will also help drive additional traffic to our site because people who aren’t already on our network will still be exposed to content when their friends interact with it.

Patrick Beeson wrote a very thoughtful post about News Mixer in December (and I’m dying to know what he’s got up his sleeve…): Medill’s News Mixer remixes story comments

Although News Mixer [doesn’t] change the traditional story format — stories are still stories that don’t work as well online as they do in print — I think their radical take on user participation is a great step forward for news sites.

And because News Mixer is built in Django, I plan on using their open-sourced code for my own project very soon, in fact.

And be sure to check out Rich Gordon’s comprehensive post about how news organizations might use News Mixer: News Mixer Options: Launch a Site, Use the Code or Be Inspired

This past week, e-Me Ventures (a Chicago-based technology firm affiliated with Gazette Communications, which sponsored the class that developed News Mixer) announced it had deployed a portion of the News Mixer code as an add-in to a test site, powered by WordPress.

“The News Mixer idea was huge. I was really blown away by the work that [the students] did,” said Abe Abreu, CEO of e-Me. “We wanted to be the first to do something with it.”

With these new developments, it seems like a good time to lay out some of the ways News Mixer — and/or its functionality — might be implemented on a production Web site.

News Mixer in the news

Finally, if you’re interested in reading more about the press we’ve received, check out Rich’s excellent roundup, keep an eye on my newsmixer tag on delicious, and follow along on Twitter.

Three reasons the new Tribune tabloid should be free, a twitter serial (republished in blog form)

The Tribune announced this week that they will begin printing the fat Chicago daily as a tabloid! Huzzah!

The Tribune’s move, replacing its broadsheet edition with the tabloid version at the retail level, is an aggressive bet that a switch in size will improve sales. There are no plans to make the tabloid-sized edition available for home delivery

I thought today’s tweets on the topic might be good to republish.

Brian Boyer
brianboyer Three reasons the new Tribune tabloid (http://bit.ly/qLN3) should be free, a twitter serial. Bring on the haters
Brian Boyer
brianboyer 1 If the paper is free, I pick it up. (Even the Red Eye, which runs 1/3 celebritrash, not including sports, neither of which I care for.)
Brian Boyer
brianboyer 2 On a tabloid page, I notice (and occasionally read!) the ads. On a broadsheet, they’re just annoying blocks to reading.
Brian Boyer
brianboyer 3 TribCo’s current gratis daily is crap. A proper paper would bring a new age of enlightenment in Chicago. Embiggen, Obama, Olympics.
Brian Boyer
brianboyer = $$$ from more effective ads shown to a larger audience, plus an improved, vital brand, loved by a better-informed populace. You dig?

Am I nuts? Riding the train this morning, I saw a lot of glossy-eyed folks reading the Red Eye, and a few diligent readers struggling with the crowd and the fatty broadsheet.

From concept to sketch to software: Building a new way to visualize votes… mmm, environminty!

Ryan Mark and I built enviroVOTE to help people visualize the environmental impact of the 2008 elections. We designed it in two evenings and made it real in a three-and-a-half-day long bender of data crunching and code.

This is the story of that time.

Sketch of enviroVOTE
+ coffee = enviroVOTE is real, live software

Sunday evening, 26 October: the concept

The idea struck us when Ryan and I discovered we had a common problem: homework. Ryan was on the hook to produce a story about the environment for News 21‘s election night coverage, and I needed to build an example presenting news data in some interesting way using charts and graphs. So we decided to combine our efforts and make something that would visualize environmental information about the election.

We searched for data to present, and found that it came in many shapes; like a candidate’s track record of support on environmental issues, or statistics on national parks, nuclear power and everything in-between. But the most compelling data set we found was not stats- or issue-based: endorsements made by environmental groups.

Statistics were cut because they’re only peripherally related to the races being run. It’s not particularly interesting to say something like “in states with more than five hydroelectric power sources, the democratic candidate prevailed 18% of the time.”

Only sportscasters can get away with that crap.

Why not issues, then? They’re hard to quantify. Candidate websites are frequently slippery, ambiguous things, and we found that few politicians responded to efforts that would make their positions crystal clear like Project Vote Smart’s Political Courage Test and Candid Answers’ Voters Guide to the Environment. The best data we could find were candidates’ voting records, but without understanding the nuance of each piece of legislation, it’s nearly impossible to determine if a vote was for or against the goodness of the earth. (Also, only incumbents have voting records.)

An endorsement is a true-false, unambiguous, easy to count thing. Environmental groups like the Sierra Club and the League of Conservation Voters publish their support for candidates online. Even better, the aforementioned Project Vote Smart — a volunteer group dedicated to strengthening our democracy through access to information — aggregates endorsements, and makes them readily available for current and historic races. And Vote Smart makes them available via an API, so others can mash up their data, just like we were itching to do.

Wednesday evening, 29 October: the design

A second fury of inspiration led to the design of the site. Marcel Pacatte, my instructor and head of Medill’s Chicago newsroom, was our source of journalistic wisdom. He and I identified our audience and discussed the angles and presentation methods that would best serve them. Obvious ideas like red/blue states and a map of the nation’s greenness were tossed — maps aren’t all that good at showing off numbers. (Notable exceptions include cartograms and the famous diagram of Napoleon’s march to Moscow, neither of which seemed sensible metaphors to adopt.)

Working out the enviroVOTE concepts on a whiteboard
Scope creep, be damned!

We decided to not make a voter’s guide, since there was little time before the election for folks to find the site, and to instead make something that’s interesting the day of the elections, and useful in the days following. So we looked for numbers to support that mission.

Counting environmentally-friendly victories would be both timely on election night, and purposeful later. We could calculate a win for the earth by counting endorsements: if the winning candidate had more endorsements, it was a green race. This was easy to aggregate nationally as well as by state.

And by running the same numbers on the previous races (two years ago for the House, six for the Senate, etc.) we could calculate the change in the environmental-friendliness of the nation’s elected officials, a figure that became known as “environmintiness.”

In addition, some races potentially held more impact for the environment than others — because of their location or the candidates running — so we decided it was necessary to highlight these key races alongside the numbers.

The sketch that served as the primary design document for enviroVOTE
The sketch that served as the primary design document for enviroVOTE

In a whirlwind sketch-a-thon, the design for the site flew together. We would show off the two big numbers in the simplest possible way. No maps, pies or (praise the lord!) Flash necessary. They’re both just percentages. To set off one from the other, we decided on a percentage for the percent change, and a one-bar chart for the victory counts, in aggregate and for individual states.

Users would be interested in seeing results from their home state, so we made the states our primary navigation, and listed them, along with their bar chart, down the left side of the page. (We explicitly decided to not use a map for navigation, like most sites do. If I lived in Rhode Island, I’d effing hate those sites.)

Putting the big numbers front and center and listing the incoming race results down the right gave users an up-to-the-minute snapshot of the evening. The writeups about key races, though important, were our least timely information, so we made them big and bold, but placed them mostly below the fold.

We produced a simple design, just three pages — home, a state and a race — each presenting more detail as you drilled down.

Saturday and Sunday morning, 1-2 November: the development

Development began Saturday morning. We decided to build the site on Django, the free and open source web development framework that we were concurrently using to build News Mixer, the big final project of our master’s degree program. (If you’re interested in our reasons why, and how it all works, check out my post that explains the same stuff re: News Mixer.)

We brainstormed names for our new baby, and immediately checked to see if the urls were available. envirovote.us was the first one we really liked, so we bought it and started running. Ryan designed a logo and whipped up a color scheme, and thus a brand was born.

Improvising the details, we built the site very closely to as it was designed. (The initial sketches were mine, but Ryan gets the props for making it look so damn sexy.) Coding the site took about a day and a half, minus time for Ryan to go home and sleep, and for me to cook soup.

We used the awesome, free tools at Google Code to list tasks and ideas, manage our source code, and track defects. The simple concept and excellent tools helped make this a relatively issue-free development cycle. Django, FTW!

Sunday afternoon and Monday, 2-3 November: the gathering, massaging, and jamming in of data

Pretty much finished with the code, minus subsequent bug fixes and tweaks, we started on the data.

Ryan used the Project Vote Smart API to gather information on current and historical races: the states, districts, and candidates that form the backbone of our system. He wrote Python scripts to repeatedly call the API, munge the response, and aggregate all of the races, candidates, wacky political parties, and the rest into files we could then pump into the database.

I attacked from the other side and scoured environmental groups’ websites, as well as the endorsement data provided by Project Vote Smart, to collect the endorsements we use to calculate the big numbers.

Once all the data was collected into text files, we then wrote more scripts to read those files, scrub the data of inconsistencies, poor spelling, and other weirdness, and finally fill the database.

All of this took a day and a half, far longer than we had hoped, and as much time as was necessary to build the website. I did not cook soup. We ordered in.

enviroVOTE is real, live software
Coffee, nerd sweat… smells like software. Yet, curiously minty-fresh.

Tuesday, 4 November

After attending class all day in Evanston, Ryan and I headed downtown for an evening of data input and cursing at screens.

Julia Dilday and Alexander Reed watched the AP wire all night, tracking races and gathering results and entering them into the system. I cannot express how much more difficult this was than we anticipated. Julia and Alex: thank you thank you thank you thank you.

Ryan kept the system humming through the night. He tamed the beast: keeping the site online, fixing bugs, and updating the administrative interface in an effort to improve the poor working conditions of Julia and Alex.

I ran the public relations effort: taking interviews, helping input incoming races, and getting the word out about our little project. I also gave enviroVOTE a voice. We set up a Twitter account to tell the nation about environmintiness as the results came in. (For a time, the site automatically twittered with each race result, until we realized that it was sending far more tweets than anyone would ever want to read, and turned it off.)

The aftermath

I’m told the presidential race was noteworthy, though I can’t recall who won — it was just one of nearly 500 races we recorded that night, and we weren’t watching the TV.

Since the 2nd, we’ve fixed a few bugs and we’ve slowly added the final race results as they’ve trickled in. The site is not nearly as dynamic as is was election night, but maybe we’ll have another few days free next year.

How we built News Mixer, part 3: our agile process

This post is last in a three-part series on News Mixer — the final project of my masters program for hacker-journalists at the Medill School of Journalism. It’s adapted (more or less verbatim) from my part of our final presentation. Visit our team blog at crunchberry.org to read the story of the project from its conception to birth, and to (soon) read our report and watch a video of our final presentation.

When you made software back in the day, first you spent the better part of year or so filling a fatty 3-ring binder with detailed specifications. Then you threw that binder over the cubicle wall to the awkward guys on the programming team.

They’d spend a year building software to the specification, and after two years, you’d have a product that no one wanted, because while you were working, the world changed. Maybe Microsoft beat you to market, or maybe Google took over. Either way, you can’t dodge the iceberg.

IMG_3605 by nautical2k
IMG_3605 by nautical2k

Agile software development is different. With agile, we plan less up front, and correct our course along the way. We design a little, we build a little, we test a little, and then we look up to see if we’re still on course. In practice, we worked in one-week cycles, called “iterations,” and kept a strict schedule.

How we met
Every morning, we scrum. A scrum is a five-minute meeting where everyone stands up, and tells the team what they did yesterday and what they’re going to do today.

And at the end of the work week, we all met for an hour to review the work done during the iteration, and to present it to our stakeholders, in this case, Annette Schulte at the Gazette and our instructors Rich Gordon and Jeremy Gilbert.

Design, develop, test, repeat!
In the following iteration, our consumer insights team tested what we built, our panel in Cedar Rapids. And their input contributed to upcoming designs and development.

And we managed this process with free and open-source tools. With a couple hundred bucks (hosting costs) and some elbow grease, we had version control for our code, a blog to promote ourselves (using WordPress), a task tracking system with a wiki for knowledge management, and a suite of collaboration tools – all of which are open source, or in the case of the Google tool suite, based heavily on open source software, and all free like speech and free like beer.

That’s all for now! Hungry for more on agile? Check out my posts about our agile process on the Crunchberry blog, and read Agile Software Development, Principles, Patterns, and Practices by Robert C. Martin, and The Pragmatic Programmer, by Andy Hunt and Dave Thomas, and Getting Real by the folks at 37signals.

How we built News Mixer, part 2: the trouble with Facebook Connect

This post is second in a three-part series on News Mixer — the final project of my masters program for hacker-journalists at the Medill School of Journalism. It’s adapted (more or less verbatim) from my part of our final presentation. Visit our team blog at crunchberry.org to read the story of the project from its conception to birth, and to (soon) read our report and watch a video of our final presentation.

Facebook Connect launched last week to much fanfare. Put simply, it’s a tool that gives Facebook users a way to log in to News Mixer, or any web site, without having to first set up a username and password and all the usual crap you’re forced to do when you when you want to use a web site. Just click the blue button, and you’re in.

Log in with Facebook Connect on News Mixer

Besides reducing barriers to entry, Connect lets us do some pretty neat stuff. Comments used to happen in the darkness — they were buried at the bottom of an article, where only the trolls dwell. But when you make a comment on News Mixer, your Facebook friends can see — and it’s our hope that this will bring your friends into the conversation.

More identity => less jackassery
In addition, by putting your name and face next to what you say, and showing your words to all your friends, we also hope that you’ll put a little more thought into that comment you’re writing.

But at what cost?
The thing is, we can find out a lot more about you than just your friends list. Connect reveals to us all sorts of information about our users. We could wish you happy birthday or tell you when one of your favorite bands is coming to town. Or we could help you find an apartment when you change from married to “it’s complicated.”

You see, whenever you use a Facebook Connect site, or any Facebook application for that matter, you effectively make the application your “friend.” Anything you reveal to your loved ones, we know too.

Facebook’s terms of service do tell us that we’re not allowed to store this data, but this is almost impossible for them to police. Facebook does allow users to restrict the information revealed to applications, but the reality is most people have no idea how much privacy they’re giving up with each witty remark.

But I promise, we’re being good! Other sites might be creepy, but we’re not. The only data News Mixer looks at is your name and your friend list, and we don’t store anything.

That’s it for part two! Can’t wait and hungry for more? Check out Facebook Connect in action at News Mixer!