Category: Tools


User Experience: About To Be Commoditized?

October 2nd, 2008 — 12:00am

Reading about the recent release of SocialText 3 I was struck by the strong parallels between the defining characteristics of enterprise environments in 2003/2004, and the emerging public Web 2.0 landscape. The essential characteristics of many enterprise environments are:

  • Syndication: streams of modular content and functionality broadcast widely to subscribers within the firewall, such as enterprise data feeds, ERP, BI capabilities, CRM, custom capabilities shared via SOA
  • Services (e.g. environmental, like the bees we used to have for pollination): identity, security, publication, data management, cloud storage, imap email, etc.
  • Social Structures: tangible networks & communities of like-minded people, oriented around a common practice, purpose, process, or pain; think of all the matrixed, horizontal org structures and ad-hoc networks encoded via internal email lists, IM, sprawling intranets, corporate directories, etc.

These same attributes are emerging as the hallmarks of the public Web 2.0 landscape. This is how the three S’s manifest for Web 2.0:

  • Syndication: A literal and figurative torrent of content in the form of blogs, RSS, feeds, streams, APIs, for social objects of all types, as well as catalogs of rentable content
  • Services: This layer is growing rapidly for the public internet, with OpenID / OAuth, mapping, visualization, backup, calendaring – the list is nearly infinite, and still expanding
  • Social Structures: The Web (and soon the mobile universe) is profoundly social now, and will continue to become ever more so.

I think you can easily see the strong parallels. It’s this similarity between the older enterprise environments and the emerging Web 2.0 environment that user experience practitioners, — and especially anyone practicing information architecture — should note.
Why? As I’ve written before, modularity is everywhere in this new environment, it’s apparent at all layers of the information world, from utilities like processing power, to services, to the elements that make up the user experience. The effects of modularity in syndication, services, and social structures on developers and IT have been profound; practices, processes, organizational structures, and business models have all shifted in response.

This wave of change first affected the developers who build and work directly with code and systems. But inevitably, disciplines further up the stack are feeling the impact of this shift, though many of us (and I’m putting user experience in this class) may not know it yet.

How will we feel that impact? One obvious way is in the pressure to adopt agile and other modular product construction practices created by and for developers as the preferred way to structure user experience and design efforts. This is a mistake that confuses the different stages of software / digital product creation (as Alan Cooper explained well at Agile2008). Design is not construction, and shouldn’t be treated as if it is. And one size fits all does not work when choosing the process and toolkit used for creating complex digital products, services, or experiences.

One result of this modularity rules all approach to user experience is the erosion of bounded or well-structured design processes that balance risk effectively for the various stages of design, and were meant to ensure the quality and relevance of the resulting products and experiences. Erosion is visible the trends toward compression or elimination of recognizable design concept exploration and usability verification activities in many design methods.

More immediately – in fact staring us right in the face, though I haven’t seen mention of it yet in m/any user experience forums – is the growing number of situations wherein there’s “No designer required”.

Examples of this abound, but just consider this feature list for the Social Text 3 Dashboard:

  • You decide what matters
  • Create your dashboard in minutes
  • Include 3rd party information and applications
  • Track & attend to what’s most important to you
  • Status updates flow automatically, as you work

If that’s not specific enough, here’s what comes out of the box, in the form of pre-built widgets:

  • My Conversations – changes others have made to any Socialtext workspace page you authored, edited, or commented on
  • My Colleagues – recent updates made by people you are subscribed to
  • Workspaces – workspaces you have access to and their activity metrics
  • Workspace Page – any page from any of your Socialtext workspaces
  • RSS Viewer – results of an RSS feed you configure
  • Workspace Tags – a tag cloud of all tags in a particular workspace
  • All People Tags – a tag cloud of all tags on people in Socialtext People

No architect required for most people here… and this trend is everywhere.

And then there’s the awesome spectre ofcommoditization. Listening to a friend describe the confusing experience of trying to select a short list of design firms for inclusion in an RFP made the linkage clear to me. I’ll quote Weil’s definition of commoditization from the paper referenced above, to make the point explicit.

Please recall that commoditization denotes the development of a competitive environment where:

  • Product differentiation is very difficult;
  • Customer loyalty and brand values are low;
  • Competition is based primarily on price; and
  • Sustainable advantage comes from cost (and sometimes quality) leadership.
  • Commoditization is driven by excess capacity.

Please note that I’m not implying user experience practitioners face overnight obsoletion.

But I am saying that I doubt our current disciplinary worldview and toolkit adequately prepare us for the realities of the new environment emerging so rapidly. Code, by contrast, is and always will be modular. (After all, that is the defining attribute of our alphabets.)

But user experience is holistic, and has to learn to build in its own way from these smaller pieces like a writer combining words and phrases. Eventually, you can create works of tremendous depth, richness, and sophistication; think of Ulysses by James Joyce, or the Mahabharata. These are richly nuanced experiences that are the result of working with modular elements.

My suggestion for one response to the oncoming wave of modularity and commoditization is to focus our value proposition in the creation of tools that other people use to define their individual experiences. In other words, shift our professional focus to higher layers of abstraction, and get into the business of defining and designing frameworks, networks, and systems of experience components. Practically, this will mean things like observing and defining the most valuable patterns arising in the use of systems of modular elements we design, and then advising on their use to solve problems. This is the direction common within enterprise environments, and in light of the appearance of public pattern libraries (Yahoo’s UI), I think I see it happening within parts of the user experience community. I’m not sure it’s happening fast enough, though.

I hoped to communicate some of these ideas in my talk on why frameworks are the future (at least for anyone practicing Experience Architecture) for the 2008 EuroIA Summit that just took place here in lovely Amsterdam. I’ll post the slides shortly. In the meantime, what do you think? Is user experience ready for the modularized, enterprise-like environment of Web 2.0? How are you responding to these changes? Is commoditization even on your radar?

Comment » | Information Architecture, Tools, User Experience (UX)

Discount Code For Rosenfeld Media

March 17th, 2008 — 12:00am

Use the discount code FOJOEL10 to receive 10% off Rosenfeld Media books purchased online. Everyone loves a bargain!

Comment » | Information Architecture, Reading Room, Tools

Al Gore Wins The Nobel *Presentation* Prize…?

October 12th, 2007 — 12:00am

Former U.S. Vice President Al Gore and the United Nations Intergovernmental Panel on Climate Change were awarded the 2007 Nobel Peace Presentation Prize today.
Al Gore
Though I’m sad to say it, this latest round of Celebrity Information Design Death Match, pitting Information Visualization Guru Dr. Edward Tufte vs. presentation tools and their legions of droning slide shufflers goes too –
Presentation software (at least it’s Keynote)…
<announcer voice>
Gore’s Nobel Prize must truly be a bitter pill for the esteemed Dr. Tufte, whose extensive declamations on the evils of PowerPoint remain insightful and even amusing, but have been outflanked by Gore’s combination of savvy presentation techniques, and repeated use of the famous “Earth’s Environment Is About to Perish” flying scissorkick move.
</announcer voice>

Seriously: Aside from the environment (we fervently hope), the real winner of this year’s Nobel Peace prize is effective storytelling that blends qualitative and quantitative messages to create a compelling visually supported narrative experience that clearly communicates complex ideas in an emotionally compelling package.

The scientists and Mr. Gore take quite different approaches to the climate changes. The committee has been a measured, peer-reviewed, government-approved statement focused on the most non-controversial findings, whereas Mr. Gore rails against a “planetary emergency.”

Both messages — however imperfect — play their part, scientists said on Friday. The Nobel Prize “is honoring the science and the publicity, and they’re necessarily different,” said Spencer A. Weart, a historian at the American Institute of Physics and author of The Discovery of Global Warming, a recent book.
From Gore and U.N. Panel Win Peace Prize for Climate Work

Dr. Tufte says, “PowerPoint presentations too often resemble a school play – very loud, very slow, and very simple.” Too often, Dr. Tufte is right: think about how many times in the last five years you’ve considered feigning a seizure or gastro-intestinal distress to escape a truly awful presentation.
book_pp_cover.gif
Yet for some ideas – and perhaps the very biggest of audiences – ‘the [school] play’s the thing’. Loud, slow, and simple might be just the right rhetorical style for complex messages that require the broadest kinds of consensus. (If Gore had figured this out during the campaign in 2000, the world would certainly be a very different place today…)

And yet, despite Gore’s pivotal role in shaping the Internet, a search for “al gore inconvenient truth” on the Slideshare website turns up – well – nothing that seems relevant in the first 10 results. There’s likewise no slideware to be had at the official site for the movie. But rest assured Mr. Gore, we know the humble origins of your Nobel Prize and Oscar winning documentary An Inconvenient Truth lie in a mere slide show.

Comment » | People, Tools

Jumpchart Sitemap Service: 3 Months Free

October 11th, 2007 — 12:00am

Jumpchart – the online sitemap service – is about to move from beta to subscription pricing.
Anyone who like to try it out, or who wants 3 free months of service should drop me a line to get an invite code.
Good luck to the Jumpchart team!

Related posts:

Comment » | Information Architecture, Tools

Photoshop And Knowledge War in Iraq

May 7th, 2007 — 12:00am

Direct connections between the war in Iraq and the realm of user experience are rare, so I was surprised when one popped up today in an article by the New York Times, titled 2 Car Bombings in Iraq Kill 25.

The article quotes an Iraqi, reacting to the destruction of a house containing a cache of munitions by American soldiers.  “The Americans are lying,” said Ali Jabbar, 28, one of several men digging through the rubble, where bicycle handlebars could be seen poking out. “If there were weapons there, they should have taken pictures to prove it.” But in a sign of the challenge Americans face here, Mr. Jabbar said that even if he saw such pictures, he would not be convinced that the destruction was justified. “The Americans can make it up with Photoshop,” he said.

It’s simultaneously terrible and fascinating that a tool I use regularly would appear in this sort of context. And yet it’s not unreasonable, given the ways that many futurists envision and describe warfare centered on information.

Here’s Alvin Toffler, from How will future wars be fought?

Above all, the full implications of what we termed Third Wave “knowledge warfare” have not yet been digested – even in the United States. The wars of the future will increasingly be prevented, won or lost based on information superiority and dominance. And that isn’t just a matter of taking out the other guy’s radar. It means waging the kind of full-scale cyber-war we described in War and Anti-War. Cyber-war involves everything from strategic deception and perception management down to tactical disruption of an adversary’s information systems. It also means understanding the role played by the global media in any conflict today. It means enhancing all your knowledge assets from intelligence, to research and development, training, and communication.

Comment » | The Media Environment, Tools, User Experience (UX)

Intranet Review Toolkit Version 1.1

April 1st, 2006 — 12:00am

Congratulations to James Robertson and StepTwo Designs for releasing an updated version of the Intranet Review Toolkit, just before this year’s IA summit in lovely Vancouver (obligatory flickr link).

Version 1.1 of the Intranet Review Toolkit includes a heuristics summary designed for quick use; it’s based on a condensed version of the complete set of heuristics you may remember I offered a while back. StepTwo was kind enough to credit my modest contribution to the overall effort.

Other additions include a collaboration / community of use destination site http://www.intranetreviewtoolkit.org.

Comment » | Intranets, Tools

Intranet Review Toolkit: Quick Heuristics Spreadsheet

December 2nd, 2005 — 12:00am

Update: Version 1.1 of the Intranet Review Toolkit is available as of 03/20/2006, and now includes a summary spreadsheet.
Thanks go to James Robertson for very gently reminding me that the licensing arrangements for the Intranet Review Toolkit preclude republishing it as a summarized form, such as the spreadsheet I posted earlier today. In my enthusiasm to share a tool with the rest of the community, I didn’t work through the full licensing implications…
Accordingly, I’ll be removing the spreadsheet from harms way immediately, while hoping it’s possible to make it available in a more legally acceptable form.
Apologies to James and the rest of the Toolkit team for any unintended harm from my oversight.

Related posts:

Comment » | Information Architecture, Intranets, Tools

Lotus Notes User Experience = Disease

September 22nd, 2005 — 12:00am

Lotus Notes has one of the most unpleas­ant and unwel­com­ing User Expe­ri­ences this side of a medium-security prison where the war­den has aspi­ra­tions towards inte­rior design and art instruc­tion. One of the most painful aspects of the Notes expe­ri­ence is the default set­tings for font size and color in the email win­dow. The default font size (for Macs) is on the order of 7 point type, and the default color for unread mes­sages is — iron­i­cally — red. The com­bi­na­tion yields a user expe­ri­ence that resem­bles a bad skin rash.

I call it “angry red microNotes” dis­ease, and it looks like this:

angry_red_micro_notes.png

Over­all, it has an unhealthy affect on one’s state of mind. The under­tones of hos­til­ity and resent­ment run­ning through­out are man­i­fold. And nat­u­rally, it is impos­si­ble to change the default font size and color for the email reader. This is fur­ther con­fir­ma­tion for my the­ory that Notes has yet to escape it’s roots as a thick client for series of uncon­nected data­bases.

After three weeks of suf­fer­ing from angry red microNotes, I real­ized I was lit­er­ally going blind from squint­ing at the tiny type, and went to Google for relief. I found niniX 1.7, a util­ity that allows Mac based Lotus Notes users the abil­ity to edit the binary for­mat Notes pref­er­ences file, and change the font size of the email client. I share it in the hopes that oth­ers may break the chains that blind them. This will only solve half the prob­lem — if some­one can fig­ure out how to change the default color for unread mes­sages to some­thing besides skin rash red, I will hap­pily share with the rest of the suf­fer­ing masses (and appar­ently there are on the order of 118 mil­lion of us out there).

But will it always be this (hor­ri­ble) way?

In Beyond Notes 7.0: IBM Lotus sketches ‘Han­nover’ user expe­ri­ence Peter Bochner of SearchDomino.com says this of the next Notes release, “Notes has often been crit­i­cized for its some­what staid user inter­face. Accord­ing to IBM’s Bis­conti, in cre­at­ing Han­nover, IBM paid atten­tion “to not just the user inter­face, but the user expe­ri­ence.“

Okay… So does that mean I’ll have my choice of dis­eases as themes for the user expe­ri­ence of my col­lab­o­ra­tion envi­ron­ment?
Accord­ing to Ken Bis­conti, IBM Lotus vice pres­i­dent of Work­place, por­tal and col­lab­o­ra­tion prod­ucts, “Through improve­ments such as con­tex­tual col­lab­o­ra­tion and sup­port for com­pos­ite apps, we’ve gone above and beyond sim­ple UI enhance­ment”.

I think sim­ple UI enhance­ment is exactly what Ken and his team should focus on for the next sev­eral years, since they have so much oppor­tu­nity for improvement.

Comment » | Enterprise, Tools, User Experience (UX)

New Web Service for Sparklines

June 27th, 2005 — 12:00am

From someone else named Joe, a free service that generates sparklines:

http://bitworking.org/projects/sparklines/

Now I can plot the truly disatisfying long-term performance of my 401ks using a convenient networked infrastructure service…

Related posts:

Comment » | Tools

Tim Bray and the RDF Challenge: Poor Tools Are A Barrier For The Semantic Web

February 7th, 2005 — 12:00am

In the latest issue of ACMQueue, Tim Bray is interviewed about his career path and early involvement with the SGML and XML standards. While recounting, Bray makes four points about the slow pace of adoption for RDF, and reiterates his conviction that the current quality of RDF-based tools is an obstacle to their adoption and the success of the Semantic Web.
Here are Bray’s points, with some commentary based on recent experiences with RDF and OWL based ontology management tools.
1. Motivating people to provide metadata is difficult. Bray says, “If there’s one thing we’ve learned, it’s that there’s no such thing as cheap meta-data.”
This is plainly a problem in spaces much beyond RDF. I hold the concept and the label meta-data itself partly responsible, since the term meta-data explicitly separates the descriptive/referential information from the idea of the data itself. I wager that user adoption of meta-data tools and processes will increase as soon as we stop dissociating a complete package into two distinct things, with different implied levels of effort and value. I’m not sure what a unified label for the base level unit construct made of meta-data and source data would be (an asset maybe?), but the implied devaluation of meta-data as an optional or supplemental element means that the time and effort demands of accurate and comprehensive tagging seem onerous to many users and businesses. Thus the proliferation of automated taxonomy and categorization generation tools…
2. Inference based processing is ineffective. Bray says, “Inferring meta-data doesn’t work… Inferring meta-data by natural language processing has always been expensive and flaky with a poor return on investment.”
I think this isn’t specific enough to agree with without qualification. However, I have seen analysis of a number of inferrencing systems, and they tend to be slow, especially when processing and updating large RDF graphs. I’m not a systems architect or an engineer, but it does seem that none of the various solutions now available directly solves the problem of allowing rapid, real-time inferrencing. This is an issue with structures that change frequently, or during high-intensity periods of the ontology life-cycle, such as initial build and editorial review.
3. Bray says, “To this day, I remain fairly unconvinced of the core Semantic Web proposition. I own the domain name RDF.net. I’ve offered the world the RDF.net challenge, which is that for anybody who can build an actual RDF-based application that I want to use more than once or twice a week, I’ll give them RDF.net. I announced that in May 2003, and nothing has come close.”
Again, I think this needs some clarification, but it brings out a serious potential barrier to the success of RDF and the Semantic Web by showcasing the poor quality of existing tools as a direct negative influencer on user satisfaction. I’ve heard this from users working with both commercial and home-built semantic structure management tools, and at all levels of usage from core to occasional.
To this I would add the idea that RDF was meant for interpretation by machines not people, and as a consequence the basic user experience paradigms for displaying and manipulating large RDF graphs and other semantic constructs remain unresolved. Mozilla and Netscape did wonders to make the WWW apparent in a visceral and tangible fashion; I suspect RDF may need the same to really take off and enter the realm of the less-than-abstruse.
4. RDF was not intended to be a Knowledge Representation language. Bray says, “My original version of RDF was as a general-purpose meta-data interchange facility. I hadn’t seen that it was going to be the basis for a general-purpose KR version of the world.”
This sounds a bit like a warning, or at least a strong admonition against reaching too far. OWL and variants are new (relatively), so it’s too early to tell if Bray is right about the scope and ambition of the Semantic Web effort being too great. But it does point out that the context of the standard bears heavily on its eventual functional achievement when put into effect. If RDF was never meant to bear its current load, then it’s not a surprise that an effective suite of RDF tools remains unavailable.

Related posts:

Comment » | Semantic Web, Tools

Back to top