Skip to content →

Category: mobile

The Big Screen, Social Applications and The Battleground of the Living Room

If “devices” are the entry point to the Network, and the market is settling in to the three screens and a cloud model, all eyes are now on the big screen. The one that occupies the room in your house that used to be called the “living room.” The pattern for the little screen, the telephone, has been set by the iPhone. The middle sized screen is the process of being split between the laptop and the iPad. But the big screen has resisted the Network, the technology industry is already filled with notable failures in this area. Even so, the battle for the living room is billed as the next big convergence event on the Network.

A screen is connected to the Network when being connected is more valuable than remaining separate. We saw this with personal computers and mobile telephones. Cable television, DVD players and DVRs have increased the amount of possible video content an individual can consume through the big screen to practically infinite proportions. If the Network adds more, infinity + infinity, does it really add value? The proposition behind GoogleTV seems to be the insertion of a search-based navigation scheme over the video/audio content accessible through the big screen.

As with the world wide web, findability through a Yahoo-style index gives way to a search-based random access model. Clearly the tools for browsing the program schedule are in need of improvement. The remote channel changer is a crippled and distorted input device, but adding a QWERTY keyboard and mouse will just make the problem worse. Google has shown that getting in between the user and any possible content she wants to access is a pretty good business model. The injection of search into the living room as a gateway to the user’s video experience creates a new advertising surface at the border of the content that traditionally garners our attention. The whole audience is collected prior to releasing it into any particular show.

Before we continue, it might be worth taking a moment to figure out what’s being fought over. There was a time when television dominated the news and entertainment landscape. Huge amounts of attention were concentrated into the prime time hours. But as Horace Deidu of Asymco points out, the living room isn’t about the devices in the physical space of the living room — it’s about the “…time and attention of the audience. The time spent consuming televised content is what’s at stake.” He further points out that the old monolithic audiences have been thoroughly disrupted and splintered by both cable and the Network. The business model of the living room has always been selling sponsored or subscription video content. But that business has been largely hollowed out, there’s really nothing worth fighting for. If there’s something there, it’ll have to be something new.

Steve Jobs, in a recent presentation, said that Apple had made some changes to AppleTV based on user feedback. Apple’s perspective on the living room is noticeably different from the accepted wisdom. They say that users want Hollywood movies and television shows in HD — and they’d like to pay for them. Users don’t want their television turned into a computer, and they don’t want to manage and sync data on hard drives. In essence it’s the new Apple Channel, the linear television programming schedule of cable television splintered into a random access model at the cost of .99¢ per high-definition show. A solid vote in favor of the stream over the enclosure/download model. And when live real-time streams can be routed through this channel, it’ll represent another fundamental change to the environment.

When we say there are three screens and a cloud, there’s an assumption that the interaction model for all three screens will be very similar. The cloud will stream the same essential computing experience to all three venues. However, Jobs and Apple are saying that the big screen is different than the other two. Sometimes this is described as the difference between “lean in” and “lean back” interactions. But it’s a little more than that: the big screen is big so that it can be social— so that family, friends or business associates can gather around it. The interaction environment encourages social applications rather than personal application software. The big screen isn’t a personal computer, it’s a social computer. This is probably what Marshall McLuhan was thinking about when he called the television experience “tribal.” Rather than changing the character of the big screen experience, Apple is attempting to operate within its established interaction modes.

Switching from one channel to another used to be the basic mode of navigation on the television. The advent of the VCR/DVD player changed that. Suddenly there was a higher level of switching involved in operating a television, from the broadcast/cable input to the playback device input. The cable industry has successfully reabsorbed some aspects of the other devices with DVRs and onDemand viewing. But to watch a DVD from your personal collection, or from Netflix, you’ll still need to change the channel to a new input device. AppleTV also requires the user to change the input channel. And it’s at this level, changing the input channel, that the contours of the battleground come in to focus. The viewer will enable and select the Comcast Channel, the Apple Channel, the Google Channel, the Game Console Channel or the locally attached-device channel. Netflix has an interesting position in all of this, their service is distributed through a number of the non-cable input channels. Netflix collects its subscription directly from the customer, whereas HBO and Showtime bundle their subscriptions into the cable company’s monthly bill. This small difference exposes an interesting asymmetry and may provide a catalyst for change in the market.

Because we’ve carried a lot of assumptions along with us into the big screen network computing space, there hasn’t been a lot of new thought about interaction or what kind of software applications make sense. Perhaps we’re too close to it; old technologies tend to become invisible. In general the software solutions aim to solve the problem of what happens in the time between watching slideshows, videos, television shows and movies (both live stream and onDemand). How does the viewer find things, save things, determine whether something is any good or not. A firm like Apple, one that makes all three of the screen devices, can think about distributing the problem among the devices with a technology like AirPlay. Just as a screen connects to the Network when it’s more valuable to be connected than to be separate, each of the three screens will begin to connect to the others when the value of connection exceeds that of remaining separate.

It should be noted that just as the evolution of the big screen is playing out in living rooms around the world, the same thing will happen in the conference rooms of the enterprise. One can easily see the projected Powerpoint presentation replaced with a presentation streamed directly from an iPad/iPhone via AirPlay to an AppleTV-connected big screen.

One Comment

Electronic Yellow Sticky Routing Slips: Tweets As Pointers

After all this time, it’s still difficult to say what a tweet is. The generic form of the word has been expressed as microblogging, but this is the wrong metaphor. Blogging and RSS advocates see Twitter as a short-form quick publishing platform. What blogging tools made easy, Twitter, and other similar systems, make even easier. Given this definition, the 140 character limit on tweets seems to be an unnecessary constraint— microblogging could simply be expanded to miniblogging and a 500 character limit for individual posts. Blog posts can be any length, they are as small or large as they need to be.

“All my plays are full length, some are just longer than others.”
– Samuel Beckett

But Twitter didn’t start with blogging or blogging tools as its central metaphor, it began with the message streams that flow through dispatching systems. The tweet isn’t a small blog post, it’s a message in a communications and logistics system. There’s a tendency to say that the tweet is a “micro” something— a very small version of some normally larger thing. But tweets are full sized, complete and lack nothing. Their size allows them to flourish in multiple communications environments, particularly the SMS system and the form factor of the mobile network device (iPhone).

The best metaphor I’ve found for a tweet is the yellow sticky. The optimal post-it note is 3 inches square and canary yellow in color. It’s not a small version of something else, its size is perfect for its purpose. There are no limitations on what can be written on a yellow sticky, but its size places constraints on the form of communication. Generally, one expects a single thought per yellow sticky. And much like Twitter, explaining what a yellow sticky is to someone who’s never used one is a difficult task. Initial market tests for the post-it note showed mixed reactions. However after extensive sampling, 90% of consumers who tried the product wanted to buy it. Like the tweet, the post-it note doesn’t have a specific purpose. Arthur Fry, one of the inventors of the post-it note, wanted a bookmark with a light adhesive to keep his place in his hymnal during church choir. The rapid acceptance of the yellow sticky, in part, had to do with not defining what it should be used for. It’s hard to imagine someone saying that you’re not using a post-it note correctly, although people say that about Twitter all the time.

One thing people use yellow stickies for is as a transmittal. I find a magazine article that I like and I pass it on to you with a short message on a yellow sticky that marks the page. I might send this package to you through the mail, use inter-office mail at work, or I might just leave it on your desk. More formal routing slips might request specific actions be taken on the attached item. Fax cover sheets are another example of this kind of communication. And Twitter is often used in a similar way. The hyperlink is the adhesive that binds the message to article I’d like to pass on to you. With Twitter, and other directed social graph services, the you I pass things on to includes followers, potentially followers of followers and users who track keywords contained in my message. At any given time, the who of the you will describe a different group. The message is passed on without obligation, the listeners may simply let it pass through, or they may take up the citation and peruse its contents.

Just as the special low-tack adhesive on the back of a yellow sticky allows you to attach it to anything without leaving marks or residue, the hyperlink allows the user of Twitter to easily point at something. Hey, look at this! Rather than a long explanation or justification, it’s just my finger pointing at something of interest. That’s interesting to me. It’s the way we talk to each other when the words aren’t the most important part of the communication.

This model of passing along items of interest is fundamentally different from web syndication. Syndication extends the distribution of published content to additional authorized contexts. Some may argue that the mostly defunct form of the ‘link blog‘, or an aggregation of link blogs, offers exactly the same value. The difference is that the tweet, as electronic routing slip, exists in a real-time social media communications system. It operates like the messages in a dispatching system. There’s an item at 3rd and Webster about cute kittens, here’s the hyperlink for interested parties. Syndication implies that I think what I’ve published is valuable, I’ve extended my distribution area and you should have a look at it. With a tweeted electronic routing slip, the value is assigned by the reader who decides to pass something along and the readers who choose to take it up within a real-time (instant) messaging system. Value is external to the thing being evaluated.

As we start to look at new applications like Flipboard, an app that collects routing slips from your social network and lays them out into a magazine format, it’s important to understand the basic unit from which the experience is built. We’re used to a newspaper filled with a combination of syndicated wire stories and proprietary ones. We know about magazines where all the stories are proprietary. A few of us are familiar with web syndication aggregators that allow us to pull in, organize and read feeds from thousands of publication sources. Building an electronic publication from sets of real-time routing slips is a fundamentally different editorial process than we’ve seen before. Of course, it could be that you don’t find the stories that your friends pass on to be very interesting. In the end, this method of  assembling a real-time publication will be judged based on the value it provides. A magazine with a thousand stories isn’t really very useful, just as a Google search result with a million answers doesn’t help you find something. Can you imagine a real-time magazine that captures the ten stories that are worth reading right now? Can you imagine a time when such a thing didn’t exist?

63 Comments

Vanilla Flavored: The Corporate Web Presence

The corporate web site used to have a brilliant excuse for its plain and simple execution. It needed the broadest possible distribution across browsers and operating systems. All customers, regardless of the technical specs of their rig, needed to be served. Some basic HTML, a few images, a conservative dollop of CSS and javascript. Transactions and data are all handled on the back end with a round trip to the server for each and every update of the display. And the display? Order up a screen resolution that serves 90%+ of the installed base as reported by server logs. Make that 800 x 600, just to be sure. This down level, conservative approach has been baked into enterprise content management systems and a boundary has been drawn around what’s possible with a corporate web presence. Mobile web was even simpler, a down level version of a down level experience. Rich internet applications (RIAs) were put into the same category as custom desktop apps, generally not worth the effort.

Back in 1998, Jakob Nielsen reported on the general conservatism of web users:

The usability tests we have conducted during the last year have shown an increasing reluctance among users to accept innovations in Web design. The prevailing attitude is to request designs that are similar to everything else people see on the Web.

When we tested advanced home page concepts we got our fingers slapped hard by the users: I don’t have time to learn special conventions for your site as one user said. Other users said, Just give it to us plain and simple, using interaction techniques we already know from other sites.

The Web is establishing expectations for narrative flow and user options and users want pages to fit within these expectations. A major reason for this evolving genre is that users frequently move back and forth between pages on different sites and that the entire corpus of the Web constitutes a single interwoven user experience rather than a set of separate publications that are accessed one at a time the way traditional books and newspapers are. The Web as a whole is the foundation of the user interface and any individual site is nothing but a speck in the Web universe.

Adoption of modern browsers was thought to be a very slow process. In 1999, Jakob Nielsen insists that we would be stuck with old browsers for a minimum of three years. Here was another reason to keep things plain and simple.

The slow uptake speeds and the bugs and inconsistencies in advanced browser features constitute a cloud with a distinct silver lining: Recognizing that we are stuck with old technology for some time frees sites from being consumed by technology considerations and focuses them on content, customer service, and usability. Back to basics indeed: that’s what sells since that’s what users want.

Over time, a couple things changed. The web standards movement gained traction with the people who build web sites. That meant figuring out what CSS could really do and working through the transition from table-based layouts to div-based layouts. Libraries like Jquery erased the differences between browser implementations of javascript. XMLhttpRequest, originally created for the web version of Microsoft’s Outlook, emerged as AJAX and turned into a defacto browser standard. The page reload could be eliminated as a requirement for a data refresh. The Webkit HTML engine was open sourced by Apple, and Google, along with a number of other mobile device makers, began to release Webkit-based browsers. With Apple, Google, Microsoft and Mozilla all jumping on the HTML5 band wagon, there’s a real motivation to move users off of pre-standards era browsers. Even Microsoft has joined the Kill IE6 movement.

The computing power of the cloud combined with the transition from a web of documents to a web of applications has changed the equation. Throw in the rise of real-time and the emergence of social media: and you’ve got an entirely different ballgame. With the massive user embrace of the iPhone, and an iPad being sold every three seconds, we might want to re-ask the question: what do users want?

Jakob Nielsen, jumps back to 1993 in an effort to preserve his business model of plain and simple:

The first crop of iPad apps revived memories of Web designs from 1993, when Mosaic first introduced the image map that made it possible for any part of any picture to become a UI element. As a result, graphic designers went wild: anything they could draw could be a UI, whether it made sense or not.

It’s the same with iPad apps: anything you can show and touch can be a UI on this device. There are no standards and no expectations.

Worse, there are often no perceived affordances for how various screen elements respond when touched. The prevailing aesthetic is very much that of flat images that fill the screen as if they were etched. There’s no lighting model or pseudo-dimensionality to indicate raised or lowered visual elements that call out to be activated.

Don Norman throws cold water on gestures and natural user interfaces by saying they aren’t new and they aren’t natural:

More important, gestures lack critical clues deemed essential for successful human-computer interaction. Because gestures are ephemeral, they do not leave behind any record of their path, which means that if one makes a gesture and either gets no response or the wrong response, there is little information available to help understand why. The requisite feedback is lacking. Moreover, a pure gestural system makes it difficult to discover the set of possibilities and the precise dynamics of execution. These problems can be overcome, of course, but only by adding conventional interface elements, such as menus, help systems, traces, tutorials, undo operations, and other forms of feedback and guides.

Touch-based interfaces built around natural interaction metaphors have only made a life for themselves outside of the research laboratory for a few years now. However I tend to think that if these interfaces were as baffling for users as Norman and Nielsen make them out to be the iPhone and iPad would have crashed and burned. Instead they can barely make them fast enough to keep up with the orders.

The classic vanilla flavored corporate web site assumes that users have old browsers and don’t want anything that doesn’t look like everything else. All new flavors are inconceivable without years and years of work by standards bodies, research labs, and the odd de facto behavior blessed by extensive usability testing. There’s a big transition ahead for the corporate web presence. Users are way ahead and already enjoying all kinds of exotic flavors.

Comments closed

Apple’s UX Strategy: I Want To Hold Your Hand

A few thoughts about the iPhone 4 and why technology does or doesn’t catch on. I’ve yet to hold one in my hand, but like everyone else I’ve got opinions. The typical gadget review takes the device’s feature list and compares using technical measures to other devices deemed competitive. Using this methodology, it would be fairly simple to dismiss the iPhone as introducing no new features. The other lines of attack involve dropped calls on the AT&T network and the App Store approval process. For some people these two items trump any feature or user experience.

Google talks about their mission as organizing the world’s information. When I think of Apple’s mission, at least their mission for the last five years or so, it revolves around getting closer to the user in real time. The technology they build flows from that principle.

I’d like to focus on just two new iPhone 4 features. The first is the new display, here’s John Gruber’s description:

It’s mentioned briefly in Apple’s promotional video about the design of the iPhone 4, but they’re using a new production process that effectively fuses the LCD and touchscreen — there is no longer any air between the two. One result of this is that the iPhone 4 should be impervious to this dust-under-the-glass issue. More importantly, though, is that it looks better. The effect is that the pixels appear to be painted on the surface of the phone; instead of looking at pixels under glass, it’s like looking at pixels on glass. Combined with the incredibly high pixel density, the overall effect is like “live print�.

The phrase that jumped out at me was “the pixels appear to be painted on the surface of the phone; instead of looking at pixels under glass.” While it seems like a small distance, a minor detail, it’s of the utmost importance. It’s the difference between touching something and touching the glass that stands in front of something. Putting the user physically in touch with the interaction surface is a major breakthrough in the emotional value of the user experience. Of course the engineering that made this kind of display is important, but it’s the design decision to get the device ever closer to the user that drove the creation of the technology. Touch creates an emotional relationship with the device, and that makes it more than just a telephone.

In a 2007 interview at the D5 conference, Steve Jobs said:

And, you know, I think of most things in life as either a Bob Dylan or a Beatles song, but there’s that one line in that one Beatles song, “you and I have memories longer than the road that stretches out ahead.

You could say that Apple’s strategy is encapsulated in the Beatles song: I Want To Hold Your Hand.

The lines that describe the feeling Jobs wants the iPhone and iPad to create are:

And when I touch you i feel happy, inside
It´s such a feeling
That my love
I can’t hide
I can’t hide
I can’t hide

The other new feature is FaceTime. Since the launch of the iPhone 3GS it’s been possible to shoot a video of something and then email it to someone, or post it to a network location that friends and family could access. Other phones had this same capability. That’s a real nice feature in an asynchronous sort of way. One of the problems with it is it has too many steps and it doesn’t work the way telephones work. Except when things are highly dysfunctional, we don’t send each other recorded audio messages to be retrieved later at a convenient time. We want to talk in real time. FaceTime allows talk + visuals in real time.

FaceTime uses phone numbers as the identity layer and works over WiFi with iPhone 4 devices only. That makes it perfectly clear under what circumstances these kind of video calls will work. Device model and kind of connectivity are only things a user needs to know. These constraints sound very limiting, but they dispel any ambiguity around the question of whether the user will be able to get video calls to work or not.

We often look to the network effect to explain the success of a product or a new platform. Has the product reached critical mass, where by virtue of its size and connectedness it continues to expand because new users gain immediate value from its scale. The network must absolutely be in place, but as we look at this window into our new virtual world, the question is: does the product put us in touch, in high definition, in real time? The more FaceTime calls that are made, the more FaceTime calls will be made. But the system will provide full value at the point when a few family members can talk to each other. Critical mass occurs at two.

Comments closed