Actions

Ontolog Forum

The printable version is no longer supported and may have rendering errors. Please update your browser bookmarks and please use the default browser print function instead.

Ontology Summit 2012: Media Kit (draft)

This is the workspace for preparing the OntologySummit2012_MediaKit.

... (this page is undergoing significant revision (Apr 5 2012)!)

What is the Ontology Summit?

Started by Ontolog and NIST in 2006, the Ontology Summit is an annual series of events that involves the ontology community and other communities related to each year's theme. Founded by experienced, senior leaders in the community, each driven by a desire to collaborate across companies, projects, subfields to address issues relevant across the field and to advance the technology and practice generally. The summit is largely a self-organizing, bottom-up, volunteer driven effort, that solicits contributions from participants around the world in both industry and academia. Moreover, the community successfully exploits a variety of communication platforms to share and develop ideas, and work towards a consensus.

Each year's Summit consists of a series events and continued discourse spanning three months, culminating in a free, two-day face-to-face workshop and symposium at the National Institute of Standards (NIST) in Gaithersburg, Maryland, USA. Discussions employ email, chat and teleconferences, including presentations given by experts about issues relevant to each year's theme. A Summit Communique, capturing the most important findings is developed throughout the summit and finalized at the symposium, endorsed by participants.

Bringing together a diverse and dispersed group of volunteers to self-organize and contribute toward a common goal is a particularly difficult collective action problem. A unique attribute of the ontology summit is that it accomplishes this outside the traditional conference framework. Its open IP policy, coupled with continued commitment and support by Ontolog, ensures that all generated material, including ideas, stories, arguments, presentations, audio recordings and transcripts are archived and available in the public domain.

A group of initial volunteers organize and facilitate the summit activities, while seeking panelists and soliciting participation from leading companies and labs. Additionally, the community maintains low barriers to participations, welcoming both well-established professionals and those new to the field. Consequently, personal or corporate agenda are largely overwhelmed in this framework, allowing discussions to focus the challenges associated with each year's theme.

The Ontology Summit program is now co-organized by Ontolog, NIST, NCOR, NCBO, IAOA, NCO_NITRD along with the co-sponsorship of other organizations that are supportive of the Summit goals and objectives.

Ontology for Big Systems

The Cambrian explosion occurred 530 million years ago, as life on earth experienced a sudden increase in diversity and the rate of evolution. Over the past 100 years, we've entered the Cambrian age for information, knowledge and systems, coupled with a constantly evolving technology landscape. The amount of knowledge that is produced, published and shared by humanity has been growing exponentially each year. In the past decade more data has been collected, more video has been produced and more information has been published than in all of previous human history.

At the same time, with the advent of the computer, digital representations, and the Internet, it has been possible to model more of the complexity of systems, connect more people. Moreover, an increasing number of people and organizations are driven to connect their systems to one another. With all this new information (aka Big Data) and all these new systems (aka Big Systems), there has also be an attendant growth in the complexity of systems that model physical phenomena and handle information, their size, their scale, their scope and their interdependence.

To address the problems that have arisen during the current period of information and knowledge, we need novel tools and approaches. Some of the major challenges facing Big Systems stem not only from their scale, but also their scope and complexity. At the same time, there are novel challenges for Big Systems when different, dispersed groups work together toward a common goal, for instance in understanding Climate Change. This leads to a need for better solutions for interoperability among federated systems and for fostering interdisciplinary collaboration. For example, we would expect that systems developed to help evacuate a city in anticipation of a tsunami to operate seamlessly with our navigation systems.

This year's Ontology Summit is titled "Ontology for Big Systems" and seeks to explore, identify and articulate how ontological methods can benefit the various disciplines required to engineer a "big system." The term "big system" is intended to cover a large scope that includes many of the terms encountered in the media such as big data, complex techno-socio-economic systems, intelligent or smart systems, cloud computing, net-centricity and collective intelligence. Established disciplines that fall within the summit scope include (but not limited to) systems engineering, software engineering, information systems modelling, and data mining.

The principal goal of the summit was to bring together and foster collaboration between the ontology community, systems community, and stakeholders of Big Systems. Together, summit participants exchanged ideas on how ontological analysis and ontology engineering might make a difference, when applied in these "big systems." We produced recommendations describing how ontologies fit into Big Systems, as well as providing examples where ontological techniques have been, or could be applied, in domains such as: bioinformatics, electronic health records, intelligence, the smart electrical grid, manufacturing and supply chains, earth and environmental, e-science, cyber-physical systems and e-government. As is traditional with the Ontology Summit series, the results will be captured in the form of a communique, with expanded supporting material provided on the web.

What Are Ontologies?

Here's a surprising fact: we use ontologies all the time. In fact, we're all unwitting ontologists. The mental models we use to interact with our world, are a type of highly internalized, implicit ontology. Our mental models serve to organize and exploit the assumptions we hold about the world - the things that exist in it and how they're related to one another.

For example, when we go to see a doctor, the visit is governed by a set of conventions and shared assumptions: patients wait in a waiting room; the doctor examines patients, patients don't examine the doctor; nurses assist doctors, etc. Yet, while we might use such ontologies all the time, rarely do we notice, let alone care. Of course, mental models are not the only type of ontology, most explicit conceptualizations of the world can be considered a type of ontology.

While we might not consciously think about it, whenever we do anything, we're acting on a vast body of implicit knowledge and belief about what we think exists and is real. If this is true, and ontologies are so pervasive, then why haven't you heard of them? Until recently, ontology was primarily of interest only to philosophers. Ontology, a branch of metaphysics, is concerned with answering big questions such as "What exists?"

Pondering the nature of being, while interesting, is not a priority for most people. Sure, we all answer "What exists" everyday, but we do so pragmatically, often intuitively and almost always implicitly. Our personal ontologies are so internalized that we are rarely aware that we use them.

Ontologies are also embedded in many of our everyday objects and systems. The forks we use are designed based on assumptions about human mouths, hands and the types of foods we eat. A transit system is designed according to assumptions about population density, growth, usage and rates. The objects and systems that pervade our lives carry with them an imprint of the beliefs of their designers.

Ontology engineering arose as an answer to a problem in computing. When humans interact with one another, we rely on a large body of assumed, shared belief about the context, including what kinds of things are in it and how they interact. The fact that we are humans, already means that we share a vast body of broadly similar experiences and knowledge. When we interact with computers that lack this knowledge, they do or conclude things we find bizarre. But we can't, in every interaction, spend time identifying and formulating the contextual background knowledge the computer needs in order understand what we say, do, or represent as data.

So, the idea arose of trying to make this background knowledge explicit for computers. This means that for a given context, we make explicit those background assumptions that humans use to reason with. With such an ontology, a computer would now able to "understand", or at least make assumptions and inferences about the part of the world we made available to it.

Over the past 30 years, as we've come to rely on an ever increasing web of socio-technical systems, we've encountered a slew of new problems. Organizations found that as employees retired or left, their knowledge would leave with them, and in many cases it would cost large amounts of money to maintain or evolve these systems. Similarly, people found that combining two systems was no trivial task. Often, implicit assumptions made by the different designers would contradict one another, making integration impossible. When machines need to talk to one another, or when we want to understand or use a system designed by another person (who might no longer be around), then those implicit assumptions suddenly matter a lot.

A fundamental task for ontology today is to make explicit the implicit assumptions that people or systems make about their relevant portion of the world. This can range from users independently, yet collaboratively creating tag clouds, to search engines providing directories or taxonomies, to organizations developing controlled vocabularies, deploying thesauri and to creating logical models of the world. This makes what we believe accessible to others in a clear, precise way. Forcing us to consider our basic assumptions and bringing to light any subtle disagreements or indeed errors.

In this sense, we engineer ontologies that represent aspects of reality for a particular purpose. The word "ontology" has been used to refer to a wide range of computational artifacts of varying complexity, ranging from folksonomies (tag clouds), controlled vocabularies, taxonomies (Yahoo! directory), thesauri (Wordnet) to logical theories of reality (Basic Formal Ontology, DOLCE).

As Leo Obrst explained in the 2007 Ontology Summit:

  • An ontology defines the terms used to describe and represent an area of knowledge (subject matter).
  • An ontology also is the model (set of concepts) for the meaning of those terms.
  • An ontology thus defines the vocabulary and the meaning of that vocabulary.

One of the most successful applications of ontologies has been in Apple's Siri. When you ask Siri "find me a restaurant", it activates a "Restaurant Ontology" which defines what a "restaurant", "reservation" and "rating" are and how they're related to one another. Siri uses this information to interact "intelligently" and book you that reservation. IBM's Watson also a number of used lightweight ontologies to distinguish between "people," "places," "times" and other categories when playing Jeopardy!

As our world becomes more complex, ontologies are a vital piece of a solution addressing the problems of Big Systems and Big Data. Depending on the intended use, ontologies can:

  • make explicit and accessible, implicit yet vital assumptions about our systems
  • enable integration among systems and data through semantic interoperability
  • improve model design, adaptability and reuse,
  • reduce development and operational costs
  • enhance decision support systems
  • aid in knowledge management and discovery
  • provide a basis for more adaptable systems

Finally, as we move into the knowledge age there is a growing expectation that our systems will be more self-describing and intelligent. In order to engineer such systems, allow intuitive use and meet expectations of all stakeholders, a more consistent and complete use of ontologies and ontological analysis must be made. The 2007 Ontology Summit provides a more thorough (and somewhat more technical) perspective on the exact nature of ontologies.

Challenges for Big Systems

As the term Big System covers a broad scope and includes many different communities, the summit focused on a set of challenges where ontology can provide value. Ontology can help not just model Big Systems and their components, but aslo guide how they can be connected to one another and combined to create new systems and services.

Big Data

One topic that has seen a lot of attention over the past couple of years, especially as the cost of computing has continued its exponential decline, is Big Data. While much of the buzz surrounding the Big Data wave has focused on figuring out how to handle its large volume, there are several other problems that must be addressed.

What happens when a molecular biologist wants to combine their big data with that of a geneticist? What if we want to combine several maps, say add Wikipedia entries to landmarks or pictures uploaded by users?

Designing Big Systems

To do any of these, we need to understand what the data is, that is, we need semantics. To successfully combine Big Data, we need meaning and this is where ontology comes in. Ontology provides a way of capturing the relevant meaning of the data, transforming it into knowledge and adding value.

Another topic that looms large for Big Systems is in their design and construction. As we engineer Big Systems, we increasingly rely on developing models of the system and its components. A large engineering project, say the development of an airplane, nuclear power plant, or car involves many people, often dispersed over large geographic areas working together on the same system. Similarly, groups of people striving to understand Big Systems such as our Earth, its climate, our bodies and so on, also rely on collaboratively building and combining models of both the entire system and its parts.

To do so effectively means that we need to build an ontology of Systems. What exactly is a system and how are its components related to one another and to the whole? The 2012 Ontology Summit dedicated a track to investigating this question, with close participation of the Systems Engineering and Modeling Languages communities.

Integrating Systems

A recurring theme in the two topics above is that of integration and federation. In a decentralized system such as the Internet, it is rarely possible to design from the top down. Rather, many organizations seek to build systems, applications and services that follow a federated model. They strive to integrate multiple, largely independently created data or systems into a coherent whole.

To do so, it is essential that the intended meaning of a model or a system component is communicated, otherwise costly errors may arise. One such extreme is evidenced by the billion dollar mistake made by the European teams making the Airbus380, where two contradictory interpretations of "holes" caused thousands of kilometers of pipes to be recalled and halted production midstream.

Interdisciplinary Collaboration

Lastly, a significant side-effect of all this ontologizing, is that through explicit semantics, one of the great challenges facing humanity today can be addressed. As the volume of our knowledge has grown so quickly and fields have become so specialized, we're losing a lot of time and money to siloed fields. Many of the problems faced by humanity today cut across multiple fields, and very rarely does anyone one person have the adequate skills, background or knowledge to tackle a problem on their own.

More than ever, the ability to connect people and teams in disparate fields, with disparate ways of looking at the world is of utmost importance. One promising approach has been to develop ontologies to help negotiate the different specialized languages used by different communities. Many exciting examples abound today, from the Open Biological and Biomedical Ontologies, to Sage Bionetworks, to the European Union's FutureICT project. Each of the projects aims to provide the necessary tools and infrastructure to connect people with different skill sets, different fields of knowledge to meet the challenges we face today.

The 2012 Ontology Summit provides a unique forum where many of these challenges for Big Systems were discussed and progressed. Engineers talked about their problems and needs, while ontologists provided suggestions for better modeling or understanding of system components. Big Data projects shared how they were successfully using ontologies, and where more effort was needed. Various communities presented different approaches to tackling the challenge of integration on the web, while others discussed how a meaningful notion of quality could be develop.

Stories

The following are a collection of some of the interesting stories that were developed through the course of the 2012 Ontology Summit

Story for Big Data

A key component of the current explosion of knowledge is the proliferation of vast amounts of data. With greater computing power, we're able to encode anyone person's DNA, track our internet usage, credit usage, the experiments at the Large Hadron Collider and so on, each of these activities creates a staggering amount of data.

While the sheer size and scale of these data sets presents a challenge, knowing how to intelligently combine the data means that we must accurately understand the world that this data represents. If we want to combine data from multiple sources, then it becomes all the more important that we understand what each source intended by the publication of the data.

To do this, we need theory. There are limits to blind statistical analysis. We need theory and statistical analysis together. Data publishers need to make explicit what their data represents, the systems that consume and transform To intelligently use this data and combine it for useful ends, involves developing theories about those relevant parts of the world. Especially if we want successful data reuse and adaptability.

There are a variety of groups working towards this vision. For example, the linked open data (URL) seeks to connect distributed data across the net. While there are many data sources available online today, that data is not readily accessible. The LOD cloud aims to create the requisite infrastructure to enable people to seamlessly build "mash-ups" by coming data from multiple sources.

Similarly, there has been a surge of work in bioinformatics, including the Open Biological and Biomedical Ontology, Gene Ontology and other sources which annotate big data with explicit semantics. These initiatives allow research groups to publish findings on genes, gene expression, proteins and so in a standardized consistent manner.

Another example is the FuturICT project funded by the European Union. Its ultimate goal is to understand and manage complex, global, socially interactive systems, with a focus on sustainability and resilience. FuturICT will build a Living Earth Platform, a simulation, visualization and participation platform to support decision-making of policy-makers, business people and citizens.

Story for Model Driven Engineering

One way to express a theory of (a part of) the world is to build a model. Engineers and designers have always built a variety of models to represent parts of their disciplines. Designing a car, a power plant, a transportation system or even the climate relies heavily on creating a computer model of the system. In the computing age, it has become far easier to shared these models and the promise of model reuse has become a desired goal.

Different fields have models of varying sophistication, though in many the semantics - the meaning - of the parts of the model are governed by implicit or inconsistent convention.

First in engineering and slowly in other fields, we're witnessing a gradual shift to explicit semantics. The various sub-disciplines within engineering have evolved from using informal modeling, to using formal languages to model their systems, to underpinning said languages with explicit semantics, to recognizing the importance of understanding the underlying ontology of the elements of the languages.

Current cutting edge research in ontology engineering involves teasing the ontological status of a system component. What does it mean for one to say that a car has a headlamp as a component? What happens to the component if the headlamp is broken or replaced? Is it the same headlamp, is it the same component?

Various standardization efforts are underway as well, from the development of ISO 15926, to providing formal semantics for the Unified Modeling Language. Similarly, groups are working to build repositories of ontologies, or libraries of ontology patterns - snippets that formalize important aspects of reality such as "part-of" or "is-a".

Story for Interoperability

The Internet means that it is far easier for different people in the different parts of the world to share and combine data, information and knowledge. If we want to realize the true potential of this interconnected world it means that we need to be able to combine not just our data, but also our models.

An initiative like Sage Bionetworks might allow a doctor in China to integrate diverse molecular mega-datasets, and reuse a predictive bionetworks built by a team in United States that deploys new insights into human disease biology by a team in France. Each different community views and prioritizes parts of the world according to their own viewpoints and interests.

Similarly, within a single enterprise, the same product may be viewed differently by each of the marketing, engineering, manufacturing, sales and accounting departments. Making sure that these views are, if not harmonized, then aligned so that information can be successfully shared entails solving interoperability.

Semantic analysis is a fundamental, essential aspect of federation and integration. Building value by combining the views of different communities means solving interoperability, and that means negotiating the implicit meaning used by each of these groups.

The Object Modeling Group has recently put out a request for proposal to create a standard to address such issues. Similarly, within the systems engineering community, one examples is ISO 15926 which aims to federate CAD/CAM/PLM systems in industry, business and eco-system-wide (beyound boundary of enterprise) scales.

Stories for Interdisciplinary Collaboration

Similarly, as knowledge has become more specialized, different communities have developed their own bodies of knowledge. Bridging these gaps can unleash a lot of potential, foster innovation, reduce the reinvention of the wheel and accelerate the development of better tools.

While each specialization may use its own jargon and technical language, the underlying reality is the same. Ontologies, in the form of explicit statement of the assumptions in each sub-field can help identify points of overlap and interest between different communities. They can serve as tools to facilitate search and discovery.

The Linked Science effort is a project that aims to create an "executable paper." It hopes to combine publication of scientific data, metadata, results, and provenance information using Linked Data principles, alongside open source and web-based environments for executing, validating and exploring research, using Cloud Computing for efficient and distributed computing and deploying Creative Commons for its legal infrastructure.

Another project, the iPlant Collaborative, is building the requisite cyberinfrastructure to help cross-disciplinary, community-driven groups publish and share information, build models and aid in search. The vision is to develop a cyberinfrastructure that is accessible to all levels of expertise, ranging from students to traditional biology researchers and computational biology experts.

People

--

maintained by the Ontology Summit 2012 Public Relations Champions: Ali Hashemi ... please do not edit