Artículos de opinión


800px-Dictionary_through_lens-300x225Dictionaries, encyclopedias, and many other repositories aimed at defining the real world are structured as a set of concepts (with identifiable keywords in each language). Real world things may be classified into these concepts. Internet is a huge repository of content which is made publicly available as a set of web pages that may be visually rendered by browsers.

However, web pages do not usually associate common semantics to its content in an explicit form. This is the reason why Internet search engines have been traditionally limited to syntactic searches based on the included words.

But what if we could progressively conceptualize Internet content as data classified into commonly-defined concepts? For example, in a web page that lists the events that take place in Barcelona, we could specify that the tag with title “Barcelona” is an instance of the common concept “City” and “The Miserables” is an instance of the common concept “Music Event”, which is currently associated to the city “Barcelona.” And what if other pages that have similar information enrich their published contents in the same way?

In that context, Internet searches could be able to provide global semantic answers about, for example, where the musical event “The Miserables” is taking place while additionally checking possible inconsistencies on the sources for improving Internet information quality assurance. There exist initiatives that are already working in this direction (see for example). However, it implies enriching web pages with semantic references based on a commonly accepted definition of concepts (a conceptual schema). In this way, web pages data become instances of such concepts and transforms simple data into semantic information that may be used for better knowledge acquisition. In this situation, all query and processing potential of conceptual schemas would be applicable for the Internet.

Nevertheless, several technical, organizational, and ethical challenges need to be considered: Which “authority” governs the global conceptual schema of Internet information; how can such concept definitions be maintained in a day-to-day changing world; how will website designers mark up their web pages in a standardized way; how can we reach common acceptation of concepts definition; how can we  support different meanings and cultural diversity, etc.

Again, we come back to modeling… but now, the challenge is modeling the world in a world of many perceptions, diversity, and huge dimension… If we contribute to the aim of moving forward in this direction we will be “rethinging” Internet.

Acerca de Albert Tort

Albert Tort es CTO de Sogeti España (grupo Capgemini). Anteriormente fue investigador y profesor de la Universidad Politécnica de Cataluña, donde actualmente es también coordinador del posgrado en Software Quality Assurance. Es especialista en ingeniería de requisitos, modelización conceptual, calidad del software, testing e inteligencia artificial. Su tesis doctoral fue titulada "Testing and Test-Driven Development of Conceptual Schemas” y es autor de diversas publicaciones relacionadas con la ingeniería del software.

0 comments on ““RETHINGING” INTERNET

Deja tu comentario

Introduce tus datos o haz clic en un icono para iniciar sesión:

Logo de

Estás comentando usando tu cuenta de Salir /  Cambiar )

Foto de Facebook

Estás comentando usando tu cuenta de Facebook. Salir /  Cambiar )

Conectando a %s

A %d blogueros les gusta esto: