After much planning and beer consumption, InfoCamp PDX is a reality. On Febuary 4, 2012, InfoGeeks from around the Pacific Northwest will converge to talk about information: what it is and how we find it, use it, structure it, design it — really whatever we want, because it’s an unconference.

Come join us! Registration’s open now. Bring your burning ideas for a session, or just come and join the conversation. Want to sponsor us? Send us an email at infocamppdx@gmail.com.

Hope to see you there!

I'm almost afraid to click…

Here's a short definition of an ontology that I wrote up the website at work. There's a lot more that can be said, but I think the discussion of why ontologies are useful is of interest.
 

An ontology is a description of the entities in an area of interest, or domain, the attributes of those entities, and the relationships between them. This description is both formal, meaning it can be acted on by a computer, and human-readable.

One of the major strengths of an ontology is that it lets us organize information in terms of the problem we’re trying to solve, not the data we’re collecting. While data remains important in an ontology-based information system, it is structured according to the concepts of the domain, not the table structure of the database it’s stored in. This is important for two reasons: we can formalize relationships between pieces of data that would only be hinted at by foreign keys and naming conventions in a database. More importantly, it frees us to think about our problem space in terms of concepts and abstraction, not data. To take a model-driven approach instead of a data-driven one. Humans think in terms of models, not data. It is models that give meaning to data. As we deal with ever increasing volumes of data, it is models that help us identify what’s important, organize it, hypothesize about it, and discover connections between disparate data.

 

Posted via email from Modelicious

Last week Google began including results from Twitter on their results page. The tweets are accessed through a timeline with a handle you can grab to scroll through results over time.

This is incredibly cool. At the same time, I can’t help noticing that while it presents a lot of information, it’s not immediately clear how to construct meaning from it.

Google talks about using the results to “’replay’ what people were saying publicly about a topic on Twitter.” That seems to describe the usage model pretty accurately: search, scroll through all results, and make of them what you will. It seems to lend itself to historical or anthropological purposes, rather than traditional search.

Here’s some sample tweets returned by searching for “Obama“: This isn’t so great if you’re interested in policy, but highly interesting if you’re investigating the teaparty movement. Ditto with this result:

Up until now, if you were researching a group of people, you would search on the group’s name. With tweets, you really want to search on the topics the group publishes about. So this could change the average information consumer’s search strategies.

The Google Blog suggests this search to “relive” Shaun White’s Olympic glory. The idea of reliving it is interesting, because what’s being relived is not the actual moment, but the response of thousands of people to that moment.

(And, like everything else, it could really use semantic search to filter out stuff like this: )

To sum up: Twitter on Google is very cool. It will change the way we search, but right now not even Google knows a good way to use it. It dumps a huge amount of raw info on the searcher, and leaves it the individual to navigate, sift, and construct meaning out of it.

But, it was only announced this week, and clever people are certainly already at work on innovative ways to build meaning out of the firehose that is the global tweetstream. A semantic search layer? Sentiment analysis? There’s a lot of possibility here.

By the time this posts, Google will probably have rolled this out worldwide. Have you tried it? What do you think?

Posted via email from Modelicious

Is the semantic web a memex?

December 31, 2009

I agree with mc schrafel that the semantic web needs a better metaphor, or really any metaphor, to help people understand and embrace it. I’m just not sure the memex is the right one. It’s not a concept that’s easily recognizable by most people. And I’m not convinced that it’s an accurate metaphor.

Central to Vannevar Bush’s original description of the memex are paths of association between items, the connection made between point a and point b. While ontologies and semantic web apps let us label the relationship between two things, I’ve yet to see an application that lets you capture the path that led you to make that connection.

So for instance Zotero lets me say Paper 1 is related to Paper 2, but not that I followed a link to a citation in paper 1, which led me to a Wikipedia page, which led me to Paper 2. Paper 2 and Paper 1 may have a generally meaningful relationship that any reader would recognize: a shared author, similar subject matter. Or their relationship may be meaningful only to me: there was some association I made along the path from Paper 1 to Paper 2 that may not matter to anyone else. However, that association — the dynamic path leading to the association, not the static association itself — may be a source of information or inspiration to me. Where is the system that lets me preserve it?

To the best of my knowledge, that system doesn’t exist yet. Really, that’s not too surprising: we’re still working on representing the relationship between two things, much less the evolution and lineage of that relationship. There are thorny semantic and user experience questions related to the larger project, especially working across the boundaries of information systems and the semantic web does (or will). But it’s a worthwhile goal, and we should make sure that we make it there and aren’t satisfied with representing static associations. Why? Because doing so creates rich context, that starts to approximate the kind of implicit context humans generate all the time. It grounds are machine representations in human notions of time. And it facilitates that mysterious capacity humans have of sparking new ideas by juxtaposing two apparently unconnected things.

So my answer to my own question at the top of this post — and to dr. schraefel — is: not yet. But maybe someday.

What I did at (info)camp

October 16, 2009

Last weekend I went to Seattle for InfoCamp, an unconference put together by a group of IA/UX/IxD/Library folks. I could only stay for the Saturday session, and due to my (ahem) directional challenges, I got there later than I would have liked. No matter: the day was absolutely inspiring. I spent it having great conversations with some very smart and creative people, and I’ll be chewing over the ideas they sparked for weeks to come.

So, it turns out Seattle is north of Portland these days. Weird. I missed the first half of Axel Roseler‘s keynote, which was too bad. He has interesting things to say about design, creativity, and process. A designer is someone who predicts the future, he said, and illustrated with examples of some pretty radical rethinkings of airplane cockpit and wayfinding interactions.

I was pretty sure I had missed my calling until I followed up the keynote with a session on service design. Service design involves applying design skills to real-world user experience and information systems. For example, making the DMV experience work for the customer, or streamlining processes across agencies. Coming from more of an information architecture than a user experience background, the first thing about it that most appeals to me is how it can be used to pull together a mishmash of accidental systems into a cohesive whole. But listening to a roomful of talented user experience practitioners made me realize how little I know about that world, and how much I should be learning about it and incorporating it into my work.

The service design conversation really took off. You could have spent your entire Saturday at various follow up sessions, and the people who organized it have plans to find a service design project to do in Seattle. I was too tempted by the other offerings to specialize. I wound up at permaculture design for social media sites, and a brainstorming session on integrating taxonomies and social media into corporate intranets.

Both of these left me with a lot of ideas zinging around my brain, and I’ll post about some of them in the next few days. In the meantime, who wants to put together InfoCamp PDX? I don’t think I can wait a whole year to do this again.

A lot of applications claim to be “semantic”.  In some cases it’s easy to understand why. For instance, Zigtag ties its tags back to a taxonomy, so it knows that the tags “New York”, “NYC” and “The Big Apple” refer to the same thing. And that’s kind of semantic-ish.  True Knowledge is built around a sophisticated ontology that understands relationships as they change over time. That’s very semantic.

In other cases, it’s hard to understand where an app’s semantics are. As semantic search becomes more of a buzzword, the term “semantics” gets thrown around freely and, ironically for a word that means “meaning”, loses its meaning.

NetBase has generated a lot of excitement for what seems to be a truly semantic approach to search. They do parts of speech analysis on the text of documents, then put the concepts they find into relationship with each other.

All good, right? But this week, NetBase launched HealthBase, a “health research showcase”. HealthBase was intended to show off their technology. Instead, it pointed up some really big holes in it that make me wonder if there’s anything semantic going on here at all.

TechCrunch has a good story about searches on HealthBase producing questionable results. The most glaring error: a query for “AIDS” returns “Jews” as one of the disease’s causes. The software then goes on to helpfully suggest salt and alcohol as ways to get rid of Jews.

Speaking as a Jew, this suggests all kinds of wildly inappropriate jokes. Ply me with alcohol and salt, and I’ll tell you a few. Leave me sober and not hypertensive, and I’ll point out that this is not actually a case of conspiracy theory run amok, but just some really bad algorithms.

NetBase’s take on the situation was interesting. This is from their response to TechCrunch:

This is an unfortunate example of homonymy, i.e. words that have different meanings.
The showcase was not configured to distinguish between the disease “AIDS” and the verb “aids” (as in aiding someone). If you click on the result “Jew” you see a sentence from a Wikipedia page about 7th Century history: “Hispano-Visigothic king Egica accuses the Jews of aiding the Muslims, and sentences all Jews to slavery. ” Although Wikipedia contains a lot of great health information it also contains non-health related information (like this one) that is hard to filter out.

This is a funny answer: this is the exact problem NetBase’s technology is supposed to solve. Pointing out that it’s hard to solve doesn’t win you any points — you’ve got to actually solve the problem for that.

I’ve got to question what’s going on under the hood here. Granted, natural language processing is far from perfect. But if you’re truly analyzing how words are used in a document, you should be able to tell the difference between the noun that refers to a disease and the verb that refers to helping someone. It’s just coincidence that these concepts are represented by two words with the same spelling. If that trips you up, you must be doing keyword matching. Good old web 1.0, why would anyone fund this or pay for this, we already have search based on it, keyword matching.

Reading between the lines, there are other disturbing implications about Netbase’s approach. They don’t seem to analyze the context of their sources — Yes, Wikipedia contains a lot of non-health related content. Don’t use it for your health knowledge base! They don’t seem to take into account how many times a statement was made — if Jews and AIDS appear together only one time, consider it an outlier. And, they don’t seem to take time into account — AIDS has only been around since the 1980’s, so how could something that happened in the 7th century possibly be relevant?

HealthBase has been “fixed” since the initial uproar, or at least fixed enough to not categorize Jews as an agent of disease (thanks, NetBase!). But given the general cluelessness about semantics in their response, you’ve got to wonder if the fix consisted of tuning their text analytics, or hacking a bunch of workarounds into their code.