Showing posts with label Semantic Web. Show all posts
Showing posts with label Semantic Web. Show all posts

September 24, 2010

The User-centric Web

David Siegel has written a book called “Pull: The Power of the Semantic Web To Transform Your Business” (Dec. 2009).

The main idea is that businesses (suppliers) need to adapt to a new world, where rather than them “push” whatever data they want to us when they want, we (consumers) will be able to get to the information we want and “pull” it whenever we need it (i.e. on demand).

Siegel identifies three types of data online of which less than 1% is currently visible web pages:

  • Public Web—what “we normally see when searching and browsing for information online: at least 21 billion pages indexed by search engines.
  • Deep Web—includes the “large data repositories that requires their internal searches,” such as Facebook, Craigslist, etc.—“about 6 trillion documents generally not seen by search engines.”
  • Private Web—data that “we can only get access to if we qualify: corporate intranets, private networks, subscription based services, and so on—about 3 trillion pages also not seen by search engines.”

In the future, Siegel sees an end of push (i.e. viewing just the Public Web) and instead a new world of pull (i.e. access to the Deep Web).

Moreover, Siegel builds on the “Semantic Web” definition of Sir Tim Berners-Lee who coined the term in the 1990s, as a virtual world where:

  • Data is unambiguous (i.e. means exactly the same things to anyone or any system).
  • Data is interconnected (i.e. it lives online in a web of databases, rather than in incompatible silos buried and inaccessible).
  • Data has an authoritative source (i.e. each piece of information has a unique name, single source, and specified terms of distribution).

While, I enjoyed browsing this book, I wasn’t completely satisfied:

  1. It’s not a tug of war between push and pullthey are not mutually exclusive. Providers push information out (i.e. make information available), and at the same time, consumers pull information in (access it on-demand).
  2. It’s not just about data anymore—it’s also about the applications (“apps”). Like data, apps are pushed out by suppliers and are pulled down by consumers. The apps make the data friendly and usable to the consumer. Rather than providing raw data or information overload, apps can help ready the data for end-user consumption.

All semantics aside, getting to information on the web is important—through a combination of push and pull—but ultimately, making the information more helpful to people through countless of innovative applications is the next phase of the how the web is evolving.

I would call this next phase, the “user-centric web.” It relies on a sound semantic web—where data is unambiguous, interconnected, and authoritative—but also takes it to the next level, serving up sound semantic information to the end-user through a myriad of applications that make the information available in ever changing and intelligent ways. This is more user-centric, and ultimately closer to where we want to be.


Share/Save/Bookmark

May 29, 2010

Internet, Anything But Shallow

Over time, people have transitioned the way they predominantly get their information and learn, as follows:
1) Experiential—people used to learn mostly by doing—through their experiences, although these were usually limited in both time and space.
2) Reading—With the printing press, doing was supplanted by reading and information came from around the world and passed over from generation to generation.
3) Television—Active reading was upended by passive watching television, where the printed word “came alive” in images and sounds streaming right into our living rooms.
4) Virtuality—And now TV is being surpassed by the interactivity of the Internet, where people have immediate access to exabytes of on-demand information covering the spectrum of human thought and existence.

The question is how does the way we learn ultimately affect what we learn and how we think—in other words does sitting and reading for example teach us to think and understand the world differently than watching TV or surfing the Internet? Is one better than the other?

I remember hearing as a kid the adults quip about kids sitting in front of the TV like zombies! And parents these days, tell their kids to “get off of Facebook and get outside and play a little in the yard or go to the mall”—get out actually do something with somebody “real.”

An article in Wired Magazine, June 2010, called “Chaos Theory” by Nicholas Carr states “even as the Internet grants us easy access to vast amounts of information, it is turning us into shallower thinkers, literally changing the structure of our brain.”

Carr contents that the Internet is changing how we think and not necessarily for the better:

1) Information overload: The Internet is a wealth of information, but “when the load exceeds our mind’s ability to process and store it, we’re unable to retain the information or to draw connections with other memories…our ability to learn suffers and our understanding remains weak.”
2) Constant interruptions: “The Internet is an interruption system. It seizes out attention only to scramble it,” though images, videos, hypertext, email, IM, tweets, RSS feeds, and advertisements.
3) “Suckers for Irrelevancy”: “The stream of new information plays to our natural tendency to overemphasize the immediate. We crave the new even when we know it’s trivial.”
4) “Intensive multitasking”: We routinely try to do (too) many things online at the same time, so that we are predominantly in skimming mode and infrequently go into any depth in any one area. In short, we sacrifice depth for breadth, and thereby lose various degrees of our ability in “knowledge acquisition, inductive analysis, critical thinking, imagination, and reflection.”
While I think that Carr makes some clever points about the dangers of Internet learning, I believe that the advantages of the Internet far outweigh the costs.

The Internet provides an unparalleled access to information and communication. It gives people the ability to get more information, from more sources, in more ways, than they would’ve in any of the other ways of learning. We are able to browse and search—skim or dig deep—as needed, anytime, anywhere.

With the Internet, we have access to information that exceeds the experiences of countless lifetimes, our world’s largest libraries—and TV isn’t even a real competitor.

At the end of the day, the Internet is a productivity multiplier like no other in history. Despite what may be considered information overload, too many online interruptions, and our inclinations to multitasking galore and even what some consider irrelevant; the Internet is an unbelievable source of information, social networking, entertainment, and online commerce.

While I believe that there is no substitute for experience, a balance of learning media—from actually doing and reading to watching and interacting online—make for an integrated and holistic learning experience. The result is learning that is diversified, interesting, and provides the greatest opportunity for everyone to learn in the way that suits him or her best.

Moreover, contrary to the Internet making us shallower thinkers as Carr contends, I think that we are actually smarter and better thinkers because of it. As a result of the Internet, we are able to get past the b.s. faster and find what we are looking for and what is actually useful to us. While pure linear reading and thinking is important and has a place, the ability online of the semantic web to locate any information and identify trends, patterns, relationships, and visualize these provides an added dimension that is anything but shallow.

Share/Save/Bookmark

February 4, 2008

Web 3.0 and Enterprise Architecture

While the Web 1.0 is viewed as an information source, and Web 2.0 as participatory, Web 3.0 is envisioned as Semantic (or the Semantic Web).

MIT Technology Review, March 2007 reports in an article entitled “A Smarter Web” by John Borland that Web 3.0 will “give computers the ability—the seeming intelligence—to understand content on the World Wide Web.” The goals is to “take the web and make it …a system that can answer questions, not just get a pile of documents that might hold an answer.”

In The New York Times, November 2007, John Markoff defined Web 3.0 “as a set of technologies that offer efficient new ways to help computers organize and draw conclusions from online data.”

Not only individuals would benefit from the Semantic Web, but companies too that “are awash in inaccessible data on intranets, in unconnected databases, even on employees’ hard drives.” The idea is to bring the data together and make it useful.

Many of you have heard of the Dewey Decimal System for organizing information. Melvin “Dewey was no technologist, but the libraries of his time were as poorly organized as today’s Web. Books were often placed in simple alphabetical order, or even lined up by size…Dewey found this system appalling: order, he believed, made for smoother access to information.” (MIT Technology Review) Melvin Dewey developed in 1876 what became The Dewey Decimal System, a library classification attempts to organize all knowledge.” (Wikipedia) In the Dewey system, books on a similar subject matter are co-located aiding discovery and access to information.

MIT Technology Review contends that like Melvin Dewey, web browser and search engine companies, like Microsoft and Google, want to help consumers locate information more efficiently.

“By the mid-1990’s, the computing community as a whole was falling in love with the idea of metadata, a way of providing Web pages with computer-readable instruction or labels…metadata promised to add the missing signage. XML—the code underlying today’s complicated websites, which describes how to find and display content, emerged as one powerful variety.” The problem with this was that it was not a systematic way of labeling data, since each developer used “their own custom ‘tags’—as if different cities posted signs in related but mutually incomprehensible dialects.”

In 1999, the World Wide Web Consortium (W3C) came up with the Resource Description Framework (RDF) for locating and describing information. Since then the vision has been for “a web that computers could browse and understand much as humans do…analogous to creating detailed road signs that cars themselves could understand and upon which they could act,” independent of human action. However, the obstacles remain for how to create ontologies that everyday busy people would use to relate data across the web—data that is currently described in myriad number of ways today—so that computers could then read and understand the data.

A second area of doubt on the realism of a Semantic Web is whether computers can truly understand the intricacies (or connotations) of human language. For example, can a computer realistically make sense of a word like marriage that can have subtle distinctions of “monogamy, polygamy, same-sex relationships, and civil unions?”

Despite the perceived obstacles, many remain not only fixated, but enamored with the notion of a Semantic Web that can not only provide amazing amounts of information, but also, like a human being, is able to analyze the data holistically, and provide actionable artificial intelligence (AI).

To enterprise architects, the Semantic Web (or Web 3.0) would be an incredible leap forward enabling organizations and individuals to get more intelligence from the web, be more productive, and ultimately provide for more efficient and effective business processes, supported by a higher order of computing enablement. Additionally, for enterprise architects themselves that deal with inordinate amounts of business and technical data—structured and unstructured—Web 3.0 technologies and methods for better mining and analyzing the data would be a welcome capability for advancing the discipline.


Share/Save/Bookmark

November 5, 2007

Semantic Web and Enterprise Architecture

MIT Technology Review, 29 October 2007 in an article entitled, “The Semantic Web Goes Mainstream,” reports that a new free web-based tool called Twine (by Radar Networks) will change the way people organize information.

Semantic Web—“a concept, long discussed in research circles, that can be described as a sort of smart network of information in which data is tagged, sorted, and searchable.”

Clay Shirky, professor in the Interactive Telecommunications Program at New York University says. “At its most basic, the Semantic Web is a campaign to tag information with extra metadata that makes it easier to search. At the upper limit, he says, it is about waiting for machines to become devastatingly intelligent.”

Twine—“Twine is a website where people can dump information that's important to them, from strings of e-mails to YouTube videos. Or, if a user prefers, Twine can automatically collect all the web pages she visited, e-mails she sent and received, and so on. Once Twine has some information, it starts to analyze it and automatically sort it into categories that include the people involved, concepts discussed, and places, organizations, and companies. This way, when a user is searching for something, she can have quick access to related information about it. Twine also uses elements of social networking so that a user has access to information collected by others in her network. All this creates a sort of ‘collective intelligence,’ says Nova Spivack, CEO and founder of Radar Networks.”

“Twine is also using extremely advanced machine learning and natural-language processing algorithms that give it capabilities beyond anything that relies on manual tagging. The tool uses a combination of natural-language algorithms to automatically extract key concepts from collections of text, essentially automatically tagging them.”

A recent article in the Economist described the Semantic Web as follows:

“The semantic web is so called because it aspires to make the web readable by machines as well as humans, by adding special tags, technically known as metadata, to its pages. Whereas the web today provides links between documents which humans read and extract meaning from, the semantic web aims to provide computers with the means to extract useful information from data accessible on the internet, be it on web pages, in calendars or inside spreadsheets.”

So whereas a tool like Google sifts through web pages based on search criteria and serves it up to humans to recognize what they are looking for, the Semantic Web actually connects related information and adds metadata that a computer can understand.
It’s like relational databases on steroids! And, with the intelligence built in to make meaning from the related information.

Like a human brain, the Semantic Web connects people, places, and events seamlessly into a unified and actionable ganglion of intelligence.

For User-centric EA, the Semantic Web could be a critical evolution in how enterprise architects analyze architecture information and come up with findings and recommendations for senior management. Using the Semantic Web, business and technology information (such as performance results, business function and activities, information requirements, applications systems, technologies, security, and human capital) would all be related, made machine readable, and automatically provide intelligence to decision-makers in terms of gaps, redundancies, inefficiencies, and opportunities—pinpointed without human intervention. Now that’s business intelligence for the CIO and other leaders, when and where they need it.

Share/Save/Bookmark