Within the bounds of our brief transit on this earth, we attempt to make our mark. Leaving a permanent trace of one’s life, in some quarters, is a large part of the purpose of our lives. In our digital lives, we leave traces wherever we go. We generate clouds of data as we surf along the surfaces of the Network. In the name of data portability, we claim the data we generate and assert personal ownership over it. We even leave instructions for how the data should be handled in the event of our death. What were footprints in the sand are now captured in digital amber.
While our most everyday communications have migrated to the Network, some of our most secret communications take a different path. It’s believed that governments have been sending secret messages using Numbers Stations since World War I. Here’s Wikipedia’s definition:
Numbers stations (or number stations) are shortwave radio stations of uncertain origin. They generally broadcast artificially generated voices reading streams of numbers, words, letters (sometimes using a spelling alphabet), tunes or Morse code. They are in a wide variety of languages and the voices are usually female, though sometimes male or children’s voices are used.
“Because [a message] can be broadcast over such an enormous area, you can be transmitting to an agent who may be thousands of miles away,” he says. And, he adds, computer communications almost always leave traces.
“It’s really hard to erase data out of your hard drive or off a memory stick,” he says. “But all you need here is a shortwave radio and pencil and paper.”
By using what’s called a one-time pad, these messages can’t be cracked. Again, here’s Mark Stout:
…because the transmissions use an unbreakable encryption system called a one-time pad: encryption key is completely random and changes with every message.
“You really truly cryptanalytically have no traction getting into a one-time pad system,â€? Stout says. “None at all.”
The use of short wave radio combines the capacity to send messages over great distances with the ability to obscure the origin of the broadcast. By taking down the message using a pencil and paper, the coded message stays off the information grid of the digital Network. Tools that pre-date the digital Network route around the media that makes permanent copies as a part of the process of transmission. While these messages are out there for anyone to listen to, and even record, the endpoints of the communication and the content of the messages remain opaque.
Historically, we’ve always had a medium that would allow us to communicate without leaving a trace. Now a whisper in the ear becomes an SMS message for your eyes only. While there’s much to be gained from our new modes of permanent public social messaging, I wonder if there’s a case to be made for the message without a paper trail, without a digital imprint, without any trace at all. Can we ever embrace the impermanence of a moment that can only be imperfectly replayed in human memory? The Numbers Station is reminder of another mode of speaking in a temporary medium.
As we consider machines that may think, we turn toward our own desires. We’d like a machine that understands what we mean, even what we intend, rather than what we strictly say. We don’t want to have to spell everything out. We’d like the machine to take a vague suggestion, figure out how to carry on, and then return to us with the best set of options to choose from. Or even better, the machine should carry out our orders and not bother us with little ambiguities or inconsistencies along the way. It should work all those things out by itself.
We might look to Shakespeare and The Tempest for a model of this type of relationship. Prospero commands the spirit Ariel to fulfill his wishes; and the sprite cheerfully complies:
ARIEL
Before you can say ‘come’ and ‘go,’
And breathe twice and cry ‘so, so,’
Each one, tripping on his toe,
Will be here with mop and mow.
Do you love me, master? no?
But The Tempest also supplies us with a counter-example in the character Caliban, who curses his servitude and his very existence:
CALIBAN
You taught me language; and my profit on’t
Is, I know how to curse. The red plague rid you
For learning me your language!
For the most part we no longer look to the spirit world for entities to do our bidding. We now place our hopes for a perfect servant in the realm of the machine. Of course, machines already do a lot for us. But frankly, for a long time now, we’ve thought that they could be a little more intelligent. Artificial intelligence, machines that think, the global brain: we’re clearly under the impression that our lot could be improved by such an advancement in technology. Here we aren’t merely thinking of an augmentation of human capability in the mode of Doug Engelbart, but rather something that stands on its own two feet.
In 2002, David Gelernter wrote a book called The Muse in the Machine: Computerizing the Poetry of Human Thought. Gelernter explored the spectrum of human thought from tightly-focused task-driven thought to poetic and dream thoughts. He makes the case that we need both modes, the whole spectrum, to think like a human does. Recently, Gelernter updated his theme in an essay for Edge.org called Dream-Logic, The Internet and Artificial Thought. He returns to the theme that most of the advocates for artificial intelligence have a defective understanding of what makes up human thought:
Many people believe that the thinker and the thought are separate. For many people, “thinking” means (in effect) viewing a stream of thoughts as if it were a PowerPoint presentation: the thinker watches the stream of his thoughts. This idea is important to artificial intelligence and the computationalist view of the mind. If the thinker and his thought-stream are separate, we can replace the human thinker by a computer thinker without stopping the show. The man tiptoes out of the theater. The computer slips into the empty seat. The PowerPoint presentation continues.
But when a person is dreaming, hallucinating — when he is inside a mind-made fantasy landscape — the thinker and his thought-stream
are not separate. They are blended together. The thinker inhabits his thoughts. No computer will be able to think like a man unless it, too, can inhabit its thoughts; can disappear into its own mind.
Gelernter makes the case that thinking must include the whole spectrum of the thought. He extends this idea of the thinker inhabiting his thoughts by saying that when we make memories, we create alternate realities:
Each remembered experience is, potentially, an alternate reality. Remembering such experiences in the ordinary sense — remembering “the beach last summer” — means, in effect, to inspect the memory from outside.  But there is another kind of remembering too: sometimes remembering “the beach last summer” means re-entering the experience, re-experiencing the beach last summer: seeing the water, hearing the waves, feeling the sunlight and sand; making real the potential reality trapped in the memory.
(An analogy: we store potential energy in an object by moving it upwards against gravity. We store potential reality in our minds by creating a memory.)
Just as thinking works differently at the top and bottom of the cognitive spectrum, remembering works differently too. At the high-focus end, remembering means ordinary remembering; “recalling” the beach. At the low-focus end, remembering means re-experiencing the beach. (We can re-experience a memory on purpose, in a limited way: you can imagine the look and fragrance of a red rose. But when focus is low, you have no choice. When you remember something, you must re-experience it.)
On the other side of the ledger, you have the arguments for a technological singularity via recursive self-improvement. One day, a machine is created that is more adept at creating machines than we are. And more importantly, it’s a machine who’s children will exceed the capabilities of the parent. Press fast forward and there’s an exponential growth in machine capability that eventually far outstrips a human’s ability to evolve.
In 2007, Gelernter and Kurzweil debated the point:
When Gelernter brings up the issue of emotions, poetic thought and the re-experiencing of memory as fundamental constituents of human thought, I can’t help but think of the body of the machine. Experience needs a location, a there for its being. Artificial intelligence needs an artificial body. To advance even a step in the direction of artificial intelligence, you have to endorse the mind/body split and think of these elements as replaceable, extensible, and to some extent, arbitrary components. This move begs a number of questions. Would a single artificial intelligence be created or would many versions emerge? Would natural selection cull the herd? Would an artificial intelligence be contained by the body of the machine in which it existed? Would each machine body contain a unique artificial intelligence with memories and emotions that were solely its own? The robot and the android are the machines we think of as having bodies. In Forbidden Planet, the science fiction update of Shakespeare’s The Tempest, we see the sprite Ariel replaced with Robby the Robot.
In Stanley Kubrick’s film 2001: A Space Odyssey, the HAL 9000 was an artificial intelligence who’s body was an entire space ship. HAL was programmed to put the mission above all else, which violated Asimov’s three laws of robotics. HAL is a classic example of an artificial intelligence that we believe has gone a step too far. A machine who has crossed a line.
When we desire to create machines that think; we want to create humans who are not fully human. Thoughts that don’t entirely think. Intelligence that isn’t fully intelligent. We want to use certain words to describe our desires, but the words express so much more than we intend. We need to hold some meaning back, the spark that makes humans, thought and intelligence what they are.
Philosophy is a battle against the bewitchment of our intelligence by means of language.
– Ludwig Wittgenstein
Clearly some filters, algorithms and agents will be better than others, but none of them will think, none will have intelligence. If part of thinking is the ability to make new analogies, then we need to think about what we do when we create and use these software machines. It becomes an easier task when we start our thinking with augmentation rather than a separate individual intelligence.
It’s interesting to think of banks as walled gardens. For example, on the Network, we might call Facebook, or aspects of Apple or Microsoft, a walled garden. The original America Online was the classic example. While most of us prefer to have walls, of some sort, around our gardens; the term is generally used to criticize a company for denying users open access, a lack of data portability and for censorship (pulling weeds). However when we consider our finances, we prefer there be a secure wall and a strong hand in the cultivation and tending of the garden. Context is everything.
More generally, a walled garden refers to a closed or exclusive set of information services provided for users. This is in contrast to providing consumers open access to the applications and content.
The recent financial crisis has presented what appears to be an opportunity to attack the market share of the big banks. Trust in these institutions is lower than normal and the very thing that made them appealing, their size, is now a questionable asset. The bigness of a bank in some ways describes the size of their private Network. On the consumer side, it’s their physical footprint with branches, or stores as some like to call them, and the extension of that footprint through their proprietary ATM network plus affiliated ATM networks. On the institutional side, there’s a matching infrastructure that represents the arteries, veins and capillaries that circulate money and abstractions of money around the country. Network is the medium of distribution. Once the platform of a big bank’s private network is in place, they endeavor to deliver the widest possible variety of product and services through these pipes. Citibank led the way in the financial supermarket space, now all the major players describe themselves as diversified financial services firms.
Every so often, in the life of the Network, the question of centralized versus distributed financial services comes up. Rather than buying a bundle of services from a single financial services supermarket, we wonder whether it’s possible to assemble best of breed services through a single online front-end. This envisions financial services firms providing complete APIs to aggregators so they can provide more friendly user interfaces and better analytics. Intuit/Mint has been the most successful with this model. It’s interesting to note that since the financial supermarkets are generally built through acquisition, under the covers, their infrastructures and systems of record are completely incompatible. So while the sales materials tout synergy, the funds to actually integrate systems go begging. The financial services supermarket in practice is aggregated, not integrated.
We’re starting to see the community banks and credit unions get more aggressive in their advertising— using a variation on the “small is beautiful” theme. For consumers, the difference in products, services and reach has started to narrow. By leveraging the Network, the small financial institution can be both small and big at the same time. In pre-Network history, being simultaneously small and big violated the laws of physics. In the era of the Network, any two points on the planet can be connected in near real time as long as Network infrastructure is present. An individual can have an international footprint. Of course, being both big and big allows a financial institution to take larger risks because, theoretically at least, it can absorb larger loses. We may see legislation from Congress that collars risk and puts limitations on the unlimited relationship between size and risk.
The Network seems to continually present opportunities for disintermediation of the dominant players in the financial services industry. Ten years ago, account aggregation via the Network seemed to be on the verge. But the model was never able to overcome its usability problems, which at bottom are really internet identity problems. We’re beginning to see a new wave of companies sprouting up to test whether a virtual distribution network through the internet can supplant the private physical networks of the established players. SmartyPig, Square and BankSimple present different takes on disintermediating the standard way we route and hold the bits that represent our money.
Once any Network endpoint can be transformed into a secure transaction environment, the advantage of the private network will have been largely neutralized. And while it hasn’t solved account aggregation’s internet identity problem yet, the mobile network device (some call it a telephone) has significantly changed the identity and network landscape. The walls around the garden represent security and engender trust. The traditional architecture of bank buildings reflect this concept. But the walled garden metaphor is built on top of the idea of carving out a private enclave from physical space. The latest round of disintermediation posits the idea that there’s a business in creating ad hoc secure transaction connections between any two Network endpoints. In this model, security and trust are earned by guaranteeing the transaction wherever it occurs.
There have always been alternative economies, transactions that occur outside of the walled gardens. In the world of leading-edge technology, we tend to look for disruption to break out in the rarefied enclaves of the early adopter. But when the margins of the urban environment grow larger than the traditional center, there’s a good chance that it’s in the improvisational economies of the favelas, shanty towns and slums that these new disruptive financial services will take root.
The corporate web site used to have a brilliant excuse for its plain and simple execution. It needed the broadest possible distribution across browsers and operating systems. All customers, regardless of the technical specs of their rig, needed to be served. Some basic HTML, a few images, a conservative dollop of CSS and javascript. Transactions and data are all handled on the back end with a round trip to the server for each and every update of the display. And the display? Order up a screen resolution that serves 90%+ of the installed base as reported by server logs. Make that 800 x 600, just to be sure. This down level, conservative approach has been baked into enterprise content management systems and a boundary has been drawn around what’s possible with a corporate web presence. Mobile web was even simpler, a down level version of a down level experience. Rich internet applications (RIAs) were put into the same category as custom desktop apps, generally not worth the effort.
The usability tests we have conducted during the last year have shown an increasing reluctance among users to accept innovations in Web design. The prevailing attitude is to request designs that are similar to everything else people see on the Web.
When we tested advanced home page concepts we got our fingers slapped hard by the users: I don’t have time to learn special conventions for your site as one user said. Other users said, Just give it to us plain and simple, using interaction techniques we already know from other sites.
The Web is establishing expectations for narrative flow and user options and users want pages to fit within these expectations. A major reason for this evolving genre is that users frequently move back and forth between pages on different sites and that the entire corpus of the Web constitutes a single interwoven user experience rather than a set of separate publications that are accessed one at a time the way traditional books and newspapers are. The Web as a whole is the foundation of the user interface and any individual site is nothing but a speck in the Web universe.
Adoption of modern browsers was thought to be a very slow process. In 1999, Jakob Nielsen insists that we would be stuck with old browsers for a minimum of three years. Here was another reason to keep things plain and simple.
The slow uptake speeds and the bugs and inconsistencies in advanced browser features constitute a cloud with a distinct silver lining: Recognizing that we are stuck with old technology for some time frees sites from being consumed by technology considerations and focuses them on content, customer service, and usability. Back to basics indeed: that’s what sells since that’s what users want.
Over time, a couple things changed. The web standards movement gained traction with the people who build web sites. That meant figuring out what CSS could really do and working through the transition from table-based layouts to div-based layouts. Libraries like Jquery erased the differences between browser implementations of javascript. XMLhttpRequest, originally created for the web version of Microsoft’s Outlook, emerged as AJAX and turned into a defacto browser standard. The page reload could be eliminated as a requirement for a data refresh. The Webkit HTML engine was open sourced by Apple, and Google, along with a number of other mobile device makers, began to release Webkit-based browsers. With Apple, Google, Microsoft and Mozilla all jumping on the HTML5 band wagon, there’s a real motivation to move users off of pre-standards era browsers. Even Microsoft has joined the Kill IE6 movement.
The computing power of the cloud combined with the transition from a web of documents to a web of applications has changed the equation. Throw in the rise of real-time and the emergence of social media: and you’ve got an entirely different ballgame. With the massive user embrace of the iPhone, and an iPad being sold every three seconds, we might want to re-ask the question: what do users want?
Jakob Nielsen, jumps back to 1993 in an effort to preserve his business model of plain and simple:
The first crop of iPad apps revived memories of Web designs from 1993, when Mosaic first introduced the image map that made it possible for any part of any picture to become a UI element. As a result, graphic designers went wild: anything they could draw could be a UI, whether it made sense or not.
It’s the same with iPad apps: anything you can show and touch can be a UI on this device. There are no standards and no expectations.
Worse, there are often no perceived affordances for how various screen elements respond when touched. The prevailing aesthetic is very much that of flat images that fill the screen as if they were etched. There’s no lighting model or pseudo-dimensionality to indicate raised or lowered visual elements that call out to be activated.
More important, gestures lack critical clues deemed essential for successful human-computer interaction. Because gestures are ephemeral, they do not leave behind any record of their path, which means that if one makes a gesture and either gets no response or the wrong response, there is little information available to help understand why. The requisite feedback is lacking. Moreover, a pure gestural system makes it difficult to discover the set of possibilities and the precise dynamics of execution. These problems can be overcome, of course, but only by adding conventional interface elements, such as menus, help systems, traces, tutorials, undo operations, and other forms of feedback and guides.
Touch-based interfaces built around natural interaction metaphors have only made a life for themselves outside of the research laboratory for a few years now. However I tend to think that if these interfaces were as baffling for users as Norman and Nielsen make them out to be the iPhone and iPad would have crashed and burned. Instead they can barely make them fast enough to keep up with the orders.
The classic vanilla flavored corporate web site assumes that users have old browsers and don’t want anything that doesn’t look like everything else. All new flavors are inconceivable without years and years of work by standards bodies, research labs, and the odd de facto behavior blessed by extensive usability testing. There’s a big transition ahead for the corporate web presence. Users are way ahead and already enjoying all kinds of exotic flavors.