rc3.org

Strong opinions, weakly held

Month: September 2012 (page 1 of 2)

One quick analytics lesson

Yesterday I saw an interesting link on Daring Fireball to a study that reported the results of searching for 2028 cities and towns in Ontario in the new iOS 6 Maps app for which Apple has apologized. Unsurprisingly, the results of the searches were not very good.

The first question that sprang to my mind when I read the piece, though, was, “How good are the Google Maps results for these searches?” Not because I thought Google’s results would be just as bad, but because looking at statistics in isolation is not particularly helpful when it comes to doing useful analysis. Obviously you can look at the results of the searches and rate the Apple Maps versus reality, but rating them against their competitors is also important. What should our expectations be, really?

Marco Tabini dug into that question, running the same searches under iOS 5.1 (running the Maps app that uses Google’s data). He found that the old Maps app does not outperform the new Maps app by a wide margin, and some interesting differences in how Apple and Google handle location searches.

This isn’t an argument that people shouldn’t be mad about the iOS 6 Maps search capabilities or lack of data, but rather that useful comparisons are critical when it comes to data analysis. That’s why experiments have control groups. Analysis that lacks baseline data is particularly pernicious in cases when people are operating under the assumption that they already know what the baseline is. In these cases, statistics are more likely to actually make people less informed.

Columnist Jon Carroll explains the Well

The Well does not own your words

San Francisco Chronicle columnist and long-time Well member Jon Carroll attempts to explain The Well. If you’ve ever been curious about it, now is a great time to join.

Why Web developers should care about analytics

I’m pretty sure the universe is trying to teach me something. For as long as I can remember, I’ve been dismissive of Web analytics. I’ve always felt that they’re for marketing people and that, at least in the realm of personal publishing, paying attention to analytics makes you some kind of sellout. Analytics is a discipline rife with unfounded claims and terrible, terrible products, as well as people engaging in cargo cultism that they pretend is analysis. Even the terminology is annoying. When people start talking about “key performance indicators” and “bounce rate” my flight instinct kicks in immediately.

In a strange turn of events, I’ve spent most of this year working in the field of Web analytics. I am a huge believer in making decisions based on quantitative analysis but I never connected that to Web analytics. As I’ve learned, Web analytics is just quantitative analysis of user behavior on Web sites. The problem is that it’s often misunderstood and usually practiced rather poorly.

The point behind this post is to make the argument that if you’re like me, a developer who has passively or actively rejected Web analytics, you might want to change your point of view. Most importantly, an understanding of analytics gives the team building for the Web a data-based framework within which they can discuss their goals, how to achieve those goals, and how to measure progress toward achieving those goals.

It’s really important as a developer to be able to participate in discussions on these terms. If you want to spend a couple of weeks making performance improvements to your database access layer, it helps to be able to explain the value in terms of increased conversion rate that results from lower page load time. Understanding what makes your project successful and how that success is measured enables you to make an argument for your priorities and, just as importantly, to be able to understand the arguments that other people are making for their priorities as well. Will a project contribute to achieving the overall goals? Can its effect be measured? Developers should be asking these questions if nobody else is.

It’s also important to be able to contribute to the evaluation of metrics themselves. If someone tells you that increasing the number of pages seen per visit to the site will increase the overall conversion rate on the site, it’s important to be able to evaluate whether they’re right or wrong. This is what half of the arguments in sports statistics are about. Does batting average or on base percentage better predict whether a hitter helps his team win? What better predicts the success of a quarterback in football, yards per attempt or yards per completion? Choosing the right metrics is no less important than monitoring the metrics that have been selected.

Finally, it often falls on the developer to instrument the application to collect the metrics needed for analytics, or at least to figure out whether the instrumentation that’s provided by a third party is actually working. Again, understanding analytics makes this part of the job much easier. It’s not uncommon for non-developers to ask for metrics based on data that is extremely difficult or costly to collect. Understanding analytics can help developers recommend alternatives that are just as useful and less burdensome.

The most important thing I’ve learned this year is that the analytics discussion is one that developers can’t really afford to sit out. As it turns out, analytics is also an extremely interesting problem as well, but I’ll talk more about that in another post. I’m also going to revisit the analytics for this site, which I ordinarily never look at, and write about that as well.

Git has taken over where Linux left off separating the geeks into know-nothings and know-it-alls. I didn’t really expect anyone to use it because it’s so hard to use, but that turns out to be its big appeal. No technology can ever be too arcane or complicated for the black t-shirt crowd.

Linus Torvalds, by way of Typical Programmer. Satire, of course.

The New York Times on power consumption at data centers

Data Centers Waste Vast Amounts of Energy, Belying Industry Image

This is an area where the Internet industry could use improvement. Here’s the damning statistic:

But at the request of The Times, the consulting firm McKinsey & Company analyzed energy use by data centers and found that, on average, they were using only 6 percent to 12 percent of the electricity powering their servers to perform computations. The rest was essentially used to keep servers idling and ready in case of a surge in activity that could slow or crash their operations.

We can do better than that.

Update: You should read this rebuttal of the article as well. (Chris posted a pointer to it in the comments but it should get top billing.)

Ask Patents Beta

Ask Patents

A new site for digging into prior art from Stack Exchange. Will it result in more patents being overturned? I doubt it. Certainly prior art exists for this type of effort. Back in 2000, O’Reilly launched a site called BountyQuest with similar goals. Tim O’Reilly explained why the site never worked out in 2003.

Generally speaking, I’m not convinced that the patent problem can be solved by digging up prior art like this. Most of the time when I see a patent claim, I can think of products that covered similar territory before the patent in question was filed. I’m sure that defense attorneys and their consultants can do the same.

David Simon on Mitt Romney

David Simon: Mitt Romney paid taxes at a rate of 13 percent and he’s proud to say so. Redux.

David Simon is a values voter.

RIP, Steve Sabol

New York Times: Steve Sabol, 69, Creative Force Behind NFL Films, Dies

Farewell to a man who was the master of his craft.

The increasingly compromised position of journalists

In today’s New York Times, David Carr writes about the subjects of articles demanding to review their quotations before they are published. Needless to say, this practice is corrosive to the practice of legitimate journalism, which is about reporting on things that public figures would not openly tell you on their own.

I think it’s interesting to look at why this is taking place. Not only is the democratization of the media hurting news reporting as a business, but it’s also reducing the value of the media as an interface between public figures and the public. At one time, if public figures wanted to disseminate a message, they had to talk to reporters with the hope that the reporter would convey their message in the way they intended.

That’s no longer necessary. A celebrity, politician, or business leader can publish a tweet, or a tweet that links to a blog post, or a video on YouTube. Those, in turn, will be shared by everyone who cares about them. The idea of making an announcement by giving a reporter an exclusive interview is almost completely dead.

The power to reach the public directly gives public figures the power to dictate the terms of their relationships with reporters. The other side of this story is the increased reliance on anonymous sources in reporting. In the modern age, if you’re going to speak publicly, you may as well just deliver the news yourself. Talking to journalists is what you do if you want to deliver news without having to give your name.

I don’t see this relationship between the media and sources changing anytime soon, so it’s up to us to take this changing relationship into account as media consumers. One thing’s for sure — it’s not a great time to be a journalist.

Matthew Yglesias on Scott Sumner

The Scott Sumner Rally

Matthew Yglesias explains how a determined economics blogger that nobody had heard of a few years ago led the charge on promoting the policies that underlie the stimulus policy announced by the Fed today. It’s an amazing story.

Older posts

© 2024 rc3.org

Theme by Anders NorenUp ↑