The World Wide Web, after just 12 years in mainstream use, has become an infrastructure accessed by hundreds of millions of people every day, and the medium through which trillions of dollars a year are transacted. In this short period, the Web has already been through a boom, a crippling bust, and a renewal to full grandeur in the modern era of 'Web 2.0'.
But imagine, if you could, a Web in which web sites are not just readable in human languages, but in which information is understandable by software to the extent that computers themselves would be able to perform the task of sharing and combining information. In other words, a Web in which machines can interpret the Web more readily, in order to make it more useful for humans. This vision for a future Internet is known as the Semantic Web.
Why is this useful? Suppose that scientific research papers were published in a Semantic Web language that enabled their content to be integrated with other research publications across the world, making research collaboration vastly more efficient. For example, a scientist running an experiment can publish his data in Semantic format, and another scientist not acquainted with the first one could search for the data and build off of it in real time. Tim Berners-Lee, as far back as 2001, said that this "will likely profoundly change the very nature of how scientific knowledge is produced and shared, in ways that we can now barely imagine."
Some are already referring to the Semantic Web as 'Web 3.0'. This type of labeling is a reliable litmus test of a technology falling into the clutches of emotional hype, and thus caution is warranted in assessing the true impact of it. I believe that the true impact of the Semantic Web will not manifest itself until 2012 or later. Nonetheless, the Semantic Web could do for scientific research what email did for postal correspondence and what MapQuest did for finding directions - eliminate almost all of the time wasted in the exchange of information.
From what little I know about the semantic web concept it sounds like it should be very useful as long as it does not make the process of authoring content too much harder. What might be a useful complimentary technology would be tools that use language analysis techniques to automatically create content. I have worked at many places where it seems that there are hundreds of unwritten answers to unasked questions that each new employee needs to ask a supervisor about constantly as they gain experience. Just when an employee gets up to speed they seem to get hired away or transfered so that all their acquired knowledge walks away with them. I can remember asking supervisors questions only to be told that I would need to call up someone in a different department who did the job ten years ago. If computers could listen to conversations, categorize problems in a semantic database and remember solutions, providing them in response to verbal inquiry, then even new employees would have access to years worth of knowledge without having to worry that the manager is out to lunch when a client or customer needs an answer immediately (or at least thinks they need it immediately).
Posted by: Saul Wall | June 12, 2007 at 01:21 PM
So does this mean research will just be done faster? How will this affect me the average guy eventually?
Posted by: frankwolftown | June 22, 2007 at 03:29 PM
Excellent point Saul Wall.
Posted by: Harsh V | June 26, 2007 at 05:42 AM