Return to ENVRI Community Home![]()
ENVRI
Services for the Environmental Community
D3.4 ENVRI Reference Model
|
Document identifier: |
D3.4 ENVRI Reference Model |
|
Date: |
28/03/2013 |
|
Activity: |
WP3 |
|
Lead Partner: |
CU |
|
Document Status: |
DRAFT |
|
Dissemination Level: |
RESTRICTED |
|
Document Link: |
<link to the website> |
|
ABSTRACT
It has been recognised that all ENVRI research infrastructures, although are very diverse, have some common characteristics, enabling them potentially to achieve a level of interoperability through the use of common standards for various functions. The objective of ENVRI Reference Model is to develop common ontological framework and standards for the description and characterisation of computational and storage infrastructures in order to achieve greater levels of seamless interoperability between the heterogeneous resources of different infrastructures.
The ENVRI Reference Model is a work-in-progress, hosted by the ENVRI project, intended for interested parties to directly comment on and contribute to.
|
Copyright © Members of the ENVRI Collaboration, 2011. See www.ENVRI.eu for details of the ENVRI project and the collaboration. ENVRI (“ Common Operations of Environmental Research Infrastructures ”) is a project co-funded by the European Commission as a Coordination and Support Action within the 7th Framework Programme. ENVRI began in October 2011 and will run for 3 years. This work is licensed under the Creative Commons Attribution-Noncommercial 3.0 License. To view a copy of this license, visit http://creativecommons.org/licenses/by/3.0/ or send a letter to Creative Commons, 171 Second Street, Suite 300, San Francisco, California, 94105, and USA. The work must be attributed by attaching the following reference to the copied elements: “Copyright © Members of the ENVRI Collaboration, 2011. See www.ENVRI.eu for details of the ENVRI project and the collaboration”. Using this document in a way and/or for purposes not foreseen in the license, requires the prior written permission of the copyright holders. The information contained in this document represents the views of the copyright holders as of the date such views are published.
|
|
Name |
Partner/Activity |
Date |
|
From |
|
|
|
|
Reviewed by |
Moderator: Reviewers: |
|
|
|
Approved by |
|
|
|
|
Issue |
Date |
Comment |
Author/Partner |
|
1.0 |
28/03/13 |
|
Yin Chen Paul Martine Herbert Schentz Barbara Magagna |
|
2.0 |
|
|
|
|
3.0 |
|
|
|
|
4.0 |
|
|
|
This document is a formal deliverable for the European Commission, applicable to all members of the ENVRI project, beneficiaries and Joint Research Unit members, as well as its collaborating projects.
Amendments, comments and suggestions should be sent to the authors.
A complete project glossary is provided at the following page: http://www.ENVRI.eu/glossary.
Frontier environmental research increasingly depends on a wide range of data and advanced capabilities to process and analyse them. The ENVRI project, “Common Operations of Environmental Research infrastructures” is a collaboration in the ESFRI Environment Cluster, with support from ICT experts, to develop common e-science components and services for their facilities. The results will speed up the construction of these infrastructures and will allow scientists to use the data and software from each facility to enable multi-disciplinary science.
The target is on developing common capabilities including software and services of the environmental e-infrastructure communities. While the ENVRI infrastructures are very diverse, they face common challenges including data capture from distributed sensors, metadata standardisation, management of high volume data, workflow execution and data visualisation. The common standards, deployable services and tools developed will be adopted by each infrastructure as it progresses through its construction phase.
Two use cases, led by the most mature infrastructures, will focus the development work on separate requirements and solutions for data pre-processing of primary data and post-processing toward publishing.
The project will be based on a common reference model created by capturing the semantic resources of each ESFRI-ENV infrastructure. This model and the development driven by the test-bed deployments result in ready-to-use systems which can be integrated into the environmental research infrastructures.
The project puts emphasis on synergy between advanced developments, not only among the infrastructure facilities, but also with ICT providers and related e-science initiatives. These links will facilitate system deployment and the training of future researchers, and ensure that the inter-disciplinary capabilities established here remain sustainable beyond the lifetime of the project.
This document describes the ENVRI Reference Model which defines a set of concepts and terms to be used by ENVRI or other environmental research communities, in order to achieve unifying understanding in communications. Built on top of the Open Distributed Processing (ODP) framework, this Reference Model captures functional elements, data flow and dependencies that are common in ENVRI research infrastructures. The Reference Model can be used as the foundation for building reference architectures, and concrete implementations can be derived.
TABLE OF CONTENTS
1 Introduction .........................................................................................................
1.1 Purpose and Scope ...........................................................................................
1.2 Rationales ............................................................................................................
1.3 Basis ......................................................................................................................
1.4 Approaches .........................................................................................................
1.5 Conformance ......................................................................................................
1.6 Related Work ......................................................................................................
1.6.1 Related concepts ...........................................................................................
1.6.2 Related reference models ............................................................................
1.6.3 Related standards .........................................................................................
1.7 Intended Audience ...........................................................................................
1.8 Document Structures ......................................................................................
2 Terminology and Glossary .......................................................................
2.1 Acronyms and Abbreviations .........................................................................
2.2 Terminology ........................................................................................................
3 Model Overview ................................................................................................
3.1 Subsystems of Environmental Infrastructures .......................................
3.1.1 Data Acquisition ............................................................................................
3.1.2 Data Curation ................................................................................................
3.1.3 Data Access ....................................................................................................
3.1.4 Data Processing .............................................................................................
3.1.5 Community Support ......................................................................................
3.2 Subsystem Relationships ................................................................................
3.3 Common Functions within Common Subsystems ...................................
4 ENVRI Reference Model .............................................................................
4.1 Enterprise Viewpoint .......................................................................................
4.1.1 Common Communities .................................................................................
4.1.2 Common Community Roles ..........................................................................
4.1.3 Common Community Behaviours ...............................................................
4.2 Information Viewpoint ....................................................................................
4.2.1 Overview .........................................................................................................
4.2.2 Components
4.2.3 Data Lifecycle ................................................................................................
4.2.4 Static Schemata ............................................................................................
4.2.5 Subsystems ....................................................................................................
4.3 Computational Viewpoint ...............................................................................
4.3.1 Computational modelling .............................................................................
4.3.2 Data Acquisition ............................................................................................
4.3.3 Data Curation ................................................................................................
4.3.4 Data Access ....................................................................................................
4.3.5 Data Processing .............................................................................................
4.3.6 Community Support ......................................................................................
4.3.7 Brokered Data Export ..................................................................................
4.3.8 Brokered Data Import ..................................................................................
4.3.9 Brokered Data Query ....................................................................................
4.3.10 Citation ...........................................................................................................
4.3.11 Internal Data Staging ...................................................................................
4.3.12 Processed Data Import .................................................................................
4.3.13 Raw Data Collection .....................................................................................
4.3.14 Request Verification .....................................................................................
4.3.15 Resource Registration ..................................................................................
4.3.16 Task Verification ...........................................................................................
5 Conclusion .............................................................................................................
6 References ............................................................................................................
Appendixes .................................................................................................................
Table of Figures
Figure 3.2: Radial depiction of ENVRI-RM requirements with the minimal model highlighted.
Figure 4.1 : Common Communities
Figure 4.2: Roles in the Data Acquisition Community
Figure 4.3: Roles in the Data Curation Community
Figure 4.4: Roles in the Data Publication Community
Figure 4.5: Roles in the Data Service Provision Community
Figure 4.6: Roles in the Data Usage Community
Figure 4.7: Behaviours of the Data Acquisition Community
Figure 4.8: Behaviours of the Data Curation Community
Figure 4.9: Behaviours of the Data Publication Community
Figure 4.10: Behaviours of The Data Service Provision Community
Figure 4.11: Behaviours of the Data Usage Community
Figure 4.12: Overview of Information Viewpoint
Figure 4.13: Information Viewpoint Components
Figure 4.14 Information Objects
Figure 4.15: Information Objects Action Types
Figure 4.16: Instances of Information Objects
Figure 4.19: State after measurements or observation
Figure 4.20: Global Interpretability of Data
Figure 4.21: Data Ready to Use
Figure 4.22: Permanent Data Acquisition
Figure 4.23: Permanent Data Curation
Figure 4.24: Permanent Data Access
Figure 4.25: Tracing of Provenance
Figure 4.30: Community Support
Figure 4.31: Brokered Data Export
Figure 4.32: Brokered Data Import
Figure 4.33: Brokered Data Query
Figure 4.34: Internal Data Staging
Figure 4.35: Processed Data Import
Figure 4.36: Raw Data Collection
Figure 4.37: Request Verification
Figure 4.38: Resource Registration
Figure 4.39: Task Verification
It has been recognised that all ENVRI research infrastructures, although very diverse, have some common characteristics, enabling them potentially to achieve a greater level of interoperability through the use of common standards for various functions. The objective of the ENVRI Reference Model is to develop a common ontological framework and standards for the description and characterisation of computational and storage infrastructures in order to achieve greater levels of seamless interoperability between the heterogeneous resources of different infrastructures.
The ENVRI Reference Model serves the following purposes [ 1 ]:
This document describes the ENVRI Reference Model which:
The Reference Model provides an abstract conceptual model; it does not impose a specific architecture nor does it impose any specific design decisions on the design of an infrastructure.
The initial model focuses on the urgent and important issues prioritised for ENVRI research infrastructures including data preservation, discovery and access, and publication. It defines a minimal set of computations to support these requirements. The core set will be extended incrementally over the course of the ENVRI project. The initial model does not cover engineering mechanisms or the applicability of existing standards or technologies.
Environmental issues will dominate the 21 st century [2]. Research infrastructures which provide advanced capabilities for data sharing, processing and analysis enable excellent research and play an ever-increasing role in the environmental sciences. The ENVRI project gathers 6 EU ESFRI [1] environmental infrastructures (ICOS [2] , EURO-Argo [3] , EISCAT-3D [4] , LifeWatch [5] , EPOS [6] , and EMSO [7] ) in order to develop common data and software components and services. The results will accelerate the construction of these infrastructures and improve interoperability among them. The experiences gained from this endeavour will also benefit the building of other top-level research infrastructures.
The primary objective of ENVRI is to agree on a reference model for joint operations. This will enable greater understanding and cooperation between users as fundamentally the model will serve to provides a universal reference framework for discussing many common technical challenges facing all of the ESFRI-ENV infrastructures. By drawing analogies between the reference components of the model and the actual elements of the infrastructures (or their proposed designs) as they exist now, various gaps and points of overlap can be identified [ 3 ].
The ENVRI Reference Model is based on the design experiences of the state-of-the are environmental research infrastructures, with a view of informing future implementation. It tackles multiple challenging issues encountering many existing initiatives, such as data streaming and storage management; data discovery and access to distributed data archives; linked computational, network and storage infrastructure; data curation, data integration, harmonisation and publication; data mining and visualisation, and scientific workflow management and execution. It uses Open Distributed Processing (ODP), a standard framework for distributed system specification, to describe the model.
To our best knowledge there is no existing reference model for environmental science research infrastructures of this scope. This work intends to make a first attempt, which can serve as a basis to inspire future research explorations.
There is an urgent need to create such a model, as we are at the beginning of a new era. The advances in automation, communication, sensing and computation enable experimental scientific processes to generate data and digital objects at unprecedentedly great speeds and volumes. Many infrastructures are starting to be built to exploit the growing wealth of scientific data and enable multi-disciplinary knowledge sharing. In the case of ENVRI, most investigated RIs are in their planning / construction phase. The high cost attached to the construction of environmental infrastructures require cooperation on the sharing of experiences and technologies, solving crucial common e-science issues and challenges together. Only by adopting a good reference model can the community secure interoperability between infrastructures, enable reuse, share resources and experiences, and avoid unnecessary duplication of effort.
The contribution of this work is threefold:
The ENVRI Reference Model is built on top of the Open Distributed Processing (ODP) framework. ODP is an international standard for architecting open, distributed processing systems. It provides an overall conceptual framework for building distributed systems in an incremental manner [ 4 - 7 ].
The reasons for adopting the ODP framework in the ENVRI project come from three aspects:
ODP adopts the object modelling approach to system specification. ISO/IEC 10746-2 [ 5 ] includes the formal definitions of the concepts and terminology adopted from object models, which provide the foundation for expressing the architecture of ODP systems. The modelling concepts fall into three categories [ 4 , 5 ]:
ODP is best known for its use of viewpoints. A viewpoint (on a system) is an abstraction that yields a specification of the whole system related to a particular set of concerns. The ODP reference model defines five specific viewpoints as follows [ 4 , 6 ]:
The initial ENVRI Reference Model covers 3 ODP viewpoints: the enterprise, information, computational viewpoints.
The approach to creating the ENVRI Reference Model is based on the analysis of the requirements of six representative environmental research infrastructures.
The ODP standard is used as the modelling and specification framework, which enables the designers from different organisations to work independently and collaboratively.
The development starts from a core model and will be incrementally extended based on the community common requirements and interests.
The reference model will be evaluated by examining the feasibilities in implementations, and the refinement of the reference model will be based on community feedback.
A conforming environmental research infrastructure should support the common subsystems described in Chapter 3 and the functional and information model described in Chapter 4.
The ENVRI Reference Model does not define or require any particular method of implementation of these concepts. It is assumed that implementers will use this reference model as a guide while developing a specific implementation to provide identified services and content. A conforming environmental research infrastructure may provide additional services to users beyond those minimally required computations defined in this document.
Any descriptive (or prescriptive) documents that claim to be conformant to the ENVRI Reference Model should use the terms and concepts defined herein in a similar way.
A reference model is an abstract framework for understanding significant relationships among the entities of some environment. It consists of a minimal set of unifying concepts, axioms and relationships within a particular problem domain. [ 8 ]
A reference model is not a reference architecture. A reference architecture is an architectural design pattern indicating an abstract solution that implements the concepts and relationships identified in the reference model [ 8 ]. Different from a reference architecture, a reference model is independent from specific standards, technologies, implementations or other concrete details. A reference model can drive the development of a reference architecture or more than one of them [ 9 ].
It could be argued that a reference model is, at its core, an ontology . Conventional reference models e.g., OSI[ 10 ] , RM-ODP [ 4 ] , OAIS[ 11 ] , are built upon modelling disciplines. Many recent works, such as the DL.org Digital Library Reference Model [ 9 ], are more ontology-like.
Both models and ontologies are technologies for information representation, but have been often been applied separately in different domains. Modelling approaches have risen to prominence in the software engineering domain over the last ten to fifteen years [ 1 2 ]. Traditionally, software engineers have taken very pragmatic approaches to data representation, encoding only the information needed to solve the problem in hand, usually in the form of language data structures or database tables. Modelling approaches are meant to increase the productivity by maximising compatibility between systems (by reuse of standardised models), simplifying the process of design (by models of recurring design patterns in the application domain), and promoting communication between individuals and teams working on the system (by a standardisation of the terminology and the best practices used in the application domain) [ 1 3 ]. On the other hand, ontologies have been developed by the Artificial Intelligence community since the 1980s. An ontology is a structuring framework for organising information. It renders shared vocabulary and taxonomies which models a domain with the definition of objects and concepts and their properties and relations. These ideas have been heavily drawn upon in the notion of the Semantic Web. [ 1 3 ]
Traditional views tend to distinguish the two technologies. The main points of argument include but are not limited to:
However, these separations between the two technologies are rapidly disappearing in recent developments. Study [ 1 3 ] shows that ‘all ontologies are models’, and ‘almost all models used in modern software engineering qualify as ontologies.’ As evidenced by the growing number of research workshops dealing with the overlap of the two disciplines (e.g., SEKE [ 16 ] , VORTE [17] , MDSW [18] , SWESE [19] , ONTOSE [20] , WoMM [21] ), there has been considerable interests in the integration of software engineering and artificial intelligence technologies in both research and practical software engineering projects.[ 1 3 ]
We tend to take this point of view and regard the ENVRI Reference Model as both a model and an ontology. The important consequence is that we can explore further in both directions, e.g., the reference model can be expressed using a modelling language, such as UML (UML4ODP). It can then be built into a tool chain, e.g., to plugin to an integrated development environment such as Eclipse, which makes it possible to reuse many existing UML code and software. On the other hand, the reference model can also be expressed using an ontology language such as RDF or OWL which can then be used in a knowledge base. In this document we explore principally from modelling aspects. In another ENVRI task, T3.4, the ontological aspect of the reference model will be exploited.
Finally, a reference model is a standard . Created by ISO in 1970, OSI is probably among the earliest reference models, which defines the well-known 7-layered network communication. As one of the ISO standard types, the reference model normally describes the overall requirements for standardisation and the fundamental principles that apply in implementation. It often serves as a framework for more specific standards [ 22 ]. This type of standard has been rapidly adopted, and many reference models exist today, which can be grouped into 3 categories, based on the type of agreement and the number of people, organisations or countries who were involved in making the agreement:
Some examples from each of the categories are discussed below, with emphasis on approaches and technologies.
In this category, we look at those defined by international organisations , such as the Advancing Open standards for the Information Society (OASIS), the Consultative Committee for Space Data Systems (CCSDS), and the Open Geospatial Consortium (OGC).
The Open Archival Information System (OAIS) Reference Model [ 11 ] is an international standard created by CCSDS and ISO which provides a framework, including terminology and concepts for archival concept needed for Long-Term digital information preservation and access.
The OASIS Reference Model for Service Oriented Architecture (SOA-RM) [ 8 ] defines the essence of service oriented architecture emerging with a vocabulary and a common understanding of SOA. It provides a normative reference that remains relevant to SOA as an abstract model, irrespective of the various and inevitable technology evolutions that will influence SOA deployment.
The OGC Reference Model (ORM) [ 23 ], describes the OGC Standards Baseline, and the current state of the work of the OGC. It provides an overview of the results of extensive development by OGC Member Organisations and individuals. Based on RM-ODP's 5 viewpoints, ORM captures business requirements and processes, geospatial information and services, reusable patterns for deployment, and provides a guide for implementations.
The Reference Model for the ORCHESTRA Architecture (RM-OA) [ 24 ] is another OGC standard. The goal of the integrated project ORCHESTRA (Open Architecture and Spatial Data Infrastructure for Risk Management) is the design and implementation of an open, service-oriented software architecture to overcome the interoperability problems in the domain of multi-risk management. The development approach of RM-OA is standard-based which is built on the integration of various international standards. Also using RM-ODP standard as the specification framework, RM-OA describes a platform neutral (abstract) model consisting of the informational and functional aspects of service networks combining architectural and service specification defined by ISO, OGC, W3C, and OASIS. [ 24 ]
In this category, we discuss those created by non-formal standard organisations.
The LifeWatch Reference Model [ 25 ], developed by the EU LifeWatch consortium, is an specialisation of the RM-OA standard which provides the guidelines for the specification and implementation of the biodiversity research infrastructure. Inherited from RM-OA, the reference model uses the ODP standard as the specification framework.
The Digital Library Reference Model [ 9 ] developed by DL.org consortium introduces the main notations characterising the whole digital library domain, in particular, it defines 3 different types of systems: (1) Digital Library, (2) Digital Library System, and (3) Digital Library Management System; 7 core concepts characterising the digital library universe: (1) Organisation, (2) Content, (3) Functionality, (4) User, (5) Policy, (6) Quality, and (7) Architecture; and 3 categories of actors: (1) DL End-Users (including, Content Creators, Content Consumers, and Digital Librarians), (2) DL Managers (Including, DL Designer, and DL System Administrators) , and (3) DL Software Developers.
The Workflow Reference Model [ 26 ] provides a common framework for workflow management systems, identifying their characteristics, terminology and components. The development of the model is based on the analysis of various workflow products in the market. The workflow Reference Model firstly introduces a top level architecture and various interfaces it has which may be used to support interoperability between different system components and integration with other major IT infrastructure components. This maps to the ODP Computational Viewpoint. In the second part, it provides an overview of the workflow application program interface, comments on the necessary protocol support for open interworking and discusses the principles of conformance to the specifications. This maps to the ODP Technology Viewpoint.
The Agent System Reference Model [ 27 ] provides a technical recommendation for developing agent systems, which captures the features, functions and data elements in the set of existing agent frameworks. Different from conventional methods, a reverse engineering method has been used to develop the reference model, which starts by identifying or creating an implementation-specific design of the abstracted system; secondly, identifying software modules and grouping them into the concepts and components; and finally, capturing the essence of the abstracted system via concepts and components.
The Data State Reference Model [ 28 ] provides an operator interaction framework for visualisation systems. It breaks the visualisation pipeline (from data to view) into 4 data stages (Value, Analytical Abstraction, Visualisation Abstraction, and View), and 3 types of transforming operations (Data Transformation, Visualisation Transformation and Visual Mapping Transformation). Using the data state model, [ 29 ] analyses 10 existing visualisation techniques including, 1) scientific visualisations, 2) GIS, 3) 2D, 4) multi-dimensional plots, 5) trees, 6) network, 7) web visualisation, 8) text, 9) information landscapes and spaces, and 10) visualisation spread sheets. The analysis results in a taxonomy of existing information visualisation techniques which help to improve the understanding of the design space of visualisation techniques.
The Munich Reference Model [ 30 ] is created for adaptive hypermedia applications which is a set of nodes and links that allows one to navigate through the hypermedia structure and that dynamically “adapts” (personalise) various visible aspects of the system to individual user’s needs. The Munich Reference Model uses an object-oriented formalisation and a graphical representation. It is built on top of the Dexter Model layered structure, and extends the functionality of each layer to include the user modelling and adaptation aspects. The model is visually represented using in UML notation and is formally specified in Object Constraint Language (which is part of the UML).
While these works use a similar approach to the development of the reference model as the ENVRI-RM, which is based on the analysis of existing systems and abstracts to obtain the ‘essence’ of those systems, a major difference is that these works have not normally met with significant feedback or been formally approved by an existing community, with the consequence that they express less authority as a standard.
Data Distribution Service for Real-Time Systems (DDS) [ 31 ], an Object Management Group (OMG) standard, is created to enable scalable, real-time, dependable, high performance, interoperable data exchanges between publishers and subscribers. DDS defines a high-level conceptual model as well as a platform-specific model. UML notations are used for specification. While DDS and the ENVRI share many similar views in design and modelling, DDS focuses on only one specific issue -- to model the communication patterns for real-time applications; on the other hand, ENVRI aims to capture a overall picture of requirements for environmental research infrastructures [ 31 ]. While DDS and the ENVRI share many similar views in design and modelling, DDS focuses on only one specific issue -- to model the communication patterns for real-time applications; on the other hand, ENVRI tends to capture the general requirements for environmental research infrastructures.
Published by the web standards consortium OASIS in 2010, the Content Management Interoperability Services (CMIS) [ 32 ] is an open standard that allows different content management systems to inter-operate over the Internet. Specially, CMIS defines an abstraction layer for controlling diverse document management systems and repositories using web protocols. It defines a domain model plus web services and Restful AtomPub bindings that can be used by applications to work with one or more Content Management repositories/systems. However as many other OASIS standards, CMIS is not a conceptual model and is highly technology dependent [ 32 ].
The intended audience of this document is the ENVRI community as well as other organisations or individuals that are interested in understanding the top level technical architecture which underpins the construction of such an architecture. In particular, the intended primary audience of this documents includes [ 33 ]:
The documents is also intended for Research Infrastructure leaders and Service Centre staffs.
The document can be read by others who want to better understand the ENVRI ongoing work, to gain understanding necessary to make contributions to the standardisation processes of environmental research infrastructures.
The rest of the document is arranged as follows:
Chapter 2 provides the terminology and concepts defined throughout the document.
Chapter 3 presents an overview of the ENVRI Reference Model, which consists of 5 common subsystems identified in a pre-study work. The concepts of these entities and their relationship is discussed.
Chapter 4 detailed describes the ENVRI Reference Model from ODP three Viewpoints, the Enterprise, Information and Computational.
Chapter 5 concludes this work.
Appendix A presents the full list of the required functionalities which is the result of the investigation of the common requirements of ENVRI Research Infrastructures. It is not part of the recommendation, and is provided for the convenience of the reader.
Appendix B provides full data lifecycle modelled in Information Viewpoint (Section 4.2.3).
CCSDS Consultative Committee for Space Data Systems
CMIS Content Management Interoperability Services
DDS Data Distribution Service for Real-Time Systems
ENVRI Environmental Research Infrastructure
ENVRI_RM ENVRI Reference Model
ESFRI European Strategy Forum on Research Infrastructures
ESFRI-ENV RI ESFRI Environmental Research Infrastructure
GIS Geographic Information System
IEC International Electrotechnical Commission
ISO International Organisation for Standardization
OAIS Open Archival Information System
OASIS Advancing Open standards for the Information Society
ODP Open Distributed Processing
OGC Open Geospatial Consortium
OMG Object Management Group
ORCHESTRA Open Architecture and Spatial Data Infrastructure for Risk Management
ORM OGC Reference Model
OSI Open Systems Interconnection
OWL Web Ontology language
SOA Service Oriented Architecture
SOA-RM Reference Model for Service Oriented Architecture
RDF Resource Description Framework
RM-OA Reference Model for the ORCHESTRA Architecture
RM-ODP Reference Model of Open Distributed Processing
UML Unified Modelling Language
W3C World Wide Web Consortium
UML4ODP Unified Modelling Language For Open Distributed Processing
Access Broker: Broker for facilitating data access requests.
Access Control: A functionality that approves or disapproves of access requests based on specified access policies.
Acquisition Service: An o versight service for an integrated data acquisition instrument network.
Active role : A active role is typically associated with a human actor.
Admin Gateway Host : A gateway service or equivalent construct used to interact with the infrastructure in an administrative capacity.
Authentication : A functionality that verifies a credential of a user.
Authentication Service : Security service responsible for the authentication of external agents making requests of infrastructure services.
Authorisation : A functionality that specifies access rights to resources.
Authorisation Service: Security service responsible for the authorisation of all requests made of infrastructure services by external agents.
Behaviour : A behaviour of a community is a composition of actions performed by roles normally addressing separate business requirements.
Capacity Manager : An active role, which is a person who manage and ensure that the IT capacity meets current and future business requirements in a cost-effective manner.
Catalogue Service : An o versight service for metadata management with the data curation subsystem.
Citation Service : Access service for resolving citations used in external publications to refer to artefacts within a research infrastructure.
Community: A collaboration which consists of a set of roles agreeing their objective to achieve a stated business purpose.
Community Gateway Host : A gateway service or equivalent construct used by the scientific community to interact with the infrastructure.
Community Support Subsystem : A subsystem that provides functionalities to manage, control, and track users’ activities and supports users to conduct their roles in the community.
Concept: Name and definition of the meaning of a thing. (abstract or real thing). Human readable definition by sentences, machine readable definition by relations to other concepts (machine readable sentences)
Conceptual Model: A collection of concepts, their attributes and their relations.
Data Access Subsystem: A subsystem that enables discovery and retrieval of data housed in data resources.
Data Acquisition Community . A community, which collects raw data and bring (streams of) measures into a system.
Data Acquisition Subsystem : A subsystem that collects raw data and brings the measures or data streams into a computational system.
Data Acquisition System: A passive role, which is a computer system which provides functionalities to automate the process of data acquisition.
Data Analysis : A functionality that inspects, cleans, transforms data, and provides data models with the goal of highlighting useful information, suggesting conclusions, and supporting decision making.
Data Assimilation : A functionality that combines observational data with output from a numerical model to produce an optimal estimate of the evolving state of the system.
Data Cataloguing : A functionality that associates a data object with one or more metadata objects which contain data descriptions.
Data Citation : A functionality that assigns an accurate, consistent and standardised reference to a data object, which can be cited in scientific publications.
Data Collection : A functionality that obtains digital values from a sensor instrument, associating consistent timestamps and necessary metadata.
Data Collector : An active role, which is a person who prepares and collects data. The purpose of data collection is to obtain information to keep on record, to make decisions about important issues, or to pass information on to others.
Data Consumer : Either an active or a passive role, which is an entity who receives and use the data.
Data Curation Community: A community, which curates the scientific data, maintains and archives them, and produces various data products with metadata.
Data Curation Subsystem: A subsystem that facilitates quality control and preservation of scientific data.
Data Curation System : A passive role, which is a system a data curator interacted with to conduct his/her role.
Data Curator: An active role, which is a person who verifies the quality of the data, preserve and maintain the data as a resource, and prepares various required data products.
Data Discovery and Access : A functionality that retrieves requested data from a data resource by using suitable search technology.
Data Discovery and Access System : A passive role, which is a system facilitating discovery and access of published data.
Data Export Service : An O versight service for export of data from the data curation subsystem.
Data Extraction : A functionality that retrieves data out of (unstructured) data sources, including web pages ,emails, documents, PDFs, scanned text, mainframe reports, and spool files.
Data Identification : A functionality that assigns (global) unique identifiers to data contents.
Data Import Service : An Oversight service for the import of new data into the data curation subsystem.
Data Mining : A functionality that supports the discovery of patterns in large data sets.
Data Originator : Either an active or a passive role, which provide the digital material to be made available for public access.
Data Processing Control : A functionality that initiates the calculation and manages the outputs to be returned to the client.
Data Processing Subsys tem : A subsystem that aggregates the data from various resources and provides computational capabilities and capacities for conducting data analysis and scientific experiments.
Data Processor : Used by the data curation subsystem to perform simple processing of newly imported / exported data and invoke the data processing subsystem for more complex tasks.
Data Product Generation : A functionality that processes data against requirement specifications and standardised formats and descriptions.
Data Provider : Either an active or a passive role, which is an entity providing the data to be used.
Data Publication : A functionality that provides clean, well-annotated, anonymity-preserving datasets in a suitable format, and by following specified data-publication and sharing policies to make the datasets publically accessible or to those who agree to certain conditions of use, and to individuals who meet certain professional criteria.
Data Publication Community: A community that assists the data publication, discovery and access.
(Data Publication) Repository : A passive role, which is a facility for the deposition of published data.
Data Quality Checking : A functionality that detects and corrects (or removes) corrupt, inconsistent or inaccurate records from data sets.
Data Query Service: An o versight service for the querying of data resources in the data curation subsystem.
Data Resource Host : A data store within the data curation subsystem.
Data Resource Registrar : Registration service for integrating new data resources into the data curation subsystem.
Data Storage & Preservation : A functionality that deposits (over long-term) the data and metadata or other supplementary data and methods according to specified policies, and makes them accessible on request.
Data Service Provision Community: A community that provides various services, applications and software/tools to link, and recombine data and information in order to derive knowledge.
Data Transmission : A functionality that transfers data over communication channel using specified network protocols.
Data Usage Community: A community who makes use of the data and service products, and transfers the knowledge into understanding.
Design of Measurement Model : A behaviour that designs the measurement or monitoring model based on scientific requirements.
Education or Trainee : An active role, a person, who makes use of the data and application services for education and training purposes.
ENVRI Reference Model : A common ontological framework and standards for the description and characterisation of computational and storage systems of ESFRI environmental research infrastructures.
Environmental Scientist : An active role, which is a person who conduct research or perform investigation for the purpose of identifying, abating, or eliminating sources of pollutants or hazards that affect either the environment or the health of the population. Using knowledge of various scientific disciplines, may collect, synthesize, study, report, and recommend action based on data derived from measurements or observations of air, food, soil, water, and other sources.
Execution Resource Host: Part of the execution platform provided by the data processing subsystem.
External Resource : A resource not within the infrastructure acting temporarily as a data source or destination for data export.
General Public, Media or Citizen (Scientist) : An active role, a person, who is interested in understanding the knowledge delivered by an environmental science research infrastructure, or discovering and exploring the knowledgebase enabled by the research infrastructure.
Identification Service: Oversight service for identifier assignment and resolution.
Instrument Host : An instrument (integrated raw data source) in the data acquisition subsystem.
(Measurement Model) Designer : An active role, which is a person who design the measurements and monitoring models based on the requirements of environmental scientists.
Measurer : An active role, which is a person who determines the ratio of a physical quantity, such as a length, time, temperature etc., to a unit of measurement, such as the meter, second or degree Celsius.
Metadata: Data describing data, usually using a predefined structure (schema).
Metadata Harvesting : A functionality that (regularly) collects metadata (in agreed formats) from different sources.
Passive Role : A passive role is typically associated with a non-human actor.
PID generator : A passive role, a system which assigns persist global unique identifiers to a (set of) digital object.
PID registry : A passive role, which is an information system for registering PIDs.
Policy or Decision Maker : An active role, a person, who makes decisions based on the data evidences.
Process Control : A functionality that receives input status, applies a set of logic statements or control algorithms, and generates a set of analogue / digital outputs to change the logic states of devices.
Process Coordination Service : An o versight service for data processing tasks deployed on infrastructure execution resources.
Process Resource Registrar : Registration service for integrating new execution resources into the data processing subsystem.
Private Sector (Industry investor or consultant) : An active role, a person, who makes use of the data and application service for predicting market so as to make business decision on producing related commercial products.
Provenance: The pathway of data generation from raw data to the actual state of data.
Quality Assessment (QA): Assessment of details of the data generation, including the check of the plausibility of the data. Usually the quality assessment is done by predefined checks on data and their generation process.
QA notation: Notation of the result of a Quality Assessment. This notation can be a nominal value out of a classification system up to a comprehensive (machine readable) description of the whole QA process.
Observer : An active role, which is a person who receives knowledge of the outside world through the senses, or records data using scientific instruments.
Reference Mode: A reference mode is an abstract framework for understanding significant relationships among the entities of some environment.
Resource Registration : A functionality that creates an entry in a resource registry and inserts resource object or a reference to a resource object in specified representations and semantics.
role : A role in a community is a prescribing behaviour that can be performed any number of times concurrently or successively.
Service Consumer : Either an active or a passive role, which is an entity using the services provided.
Service Registry : A passive role, which is an information system for registering services.
Scientific Modelling and Simulation : A functionality that supports the generation of abstract, conceptual, graphical or mathematical models, and to run an instance of the model.
Scientist or Researcher : An active role, a person, who makes use of the data and application services to conduct scientific research.
(Scientific) Workflow Enactment : A specialisation of Workflow Enactment, which support of composition and execution a series of computational or data manipulation steps, or a workflow, in a scientific application. Important processes should be recorded for provenance purposes.
Semantic Annotation: link from a thing (single datum, data set, data container) to a concept within a conceptual model, enabling the discovery of the meaning of the thing by human and machines.
Semantic Mediator : A passive role, which is a system or middleware facilitating semantic mapping discovery and integration of heterogeneous data.
Service Provider : Either an active or a passive role, which is an entity providing the services to be used.
Sensor : A passive role, which is a converter that measures a physical quantity and converts it into a signal which can be read by an observer or by an (electronic) instrument.
Sensor network: A passive role, which is a network consists of distributed autonomous sensors to monitor physical or environmental conditions.
Storage : A passive role, which is memory, components, devices and media that retain digital computer data used for computing for some interval of time.
Storage Administrator : An active role, which is a person who has the responsibilities to the design of data storage, tune queries, perform backup and recovery operations, raid mirrored arrays, making sure drive space is available for the network.
Subsystem : A subsystem is a set of capabilities that collectively are defined by a set of interfaces with corresponding operations that can be invoked by other subsystems. Subsystems are disjoint from each other.
Technician : An active role, which is a person who develop and deploy the sensor instruments, establishing and testing the sensor network, operating, maintaining, monitoring and repairing the observatory hardware.
Technologist or Engineer : An active role, a person, who develop and maintains the research infrastructure.
User Behaviour Tracking : A behaviour enabled by a Community Support System that to track the Users . If the research infrastructure has identity management, authorisation mechanisms, accounting mechanisms, for example, a Data Access Subsystem is provided, then the Community Support System either include these or work well with them.
User Profile Management : A behaviour enabled by a Community Support System that to support persistent and mobile profiles, where profiles will include preferred interaction settings, preferred computational resource settings, and so on.
User Working Space Management : A behaviour enabled by a Community Support System that to support work spaces that allow data, document and code continuity between connection sessions and accessible from multiple sites or mobile smart devices.
User Working Relationships Management : A behaviour enabled by a Community Support System that to support a record of working relationships, (virtual) group memberships and friends.
User Group Work Supporting : A behaviour enabled by a Community Support System that to support controlled sharing, collaborative work and publication of results, with persistent and externally citable PIDs.
Upload Broker : Broker for facilitating data upload requests from external contributors .
The ENVRI-RM is partitioned into five subsystems, each with its own domain of responsibility: Data Acquisition , Data Curation , Data Access , Data Processing and Community Support . The partitioning of the reference model into subsystems is based broadly on a notion of data life-cycle evident in all existing research infrastructures investigated.
This lifecycle begins with the acquisition of raw data from a network of integrated data collecting instruments (seismographs, weather stations, robotic buoys, human observations, etc.) which is then pre-processed and curated within a number of data stores belonging to an infrastructure or one of its delegate infrastructures. This data is then made accessible to authorised requests by parties outwith the infrastructure, as well as to services within the infrastructure. This results in a nature partitioning of data acquisition, curation and access. In addition, data can be extracted from parts of the infrastructure and made subject to data processing, the results of which can then be resituated within the infrastructure. Finally, the community support subsystem provides tools and services required to handle data outside of the core infrastructure and reintegrate it when necessary.
Each subsystem should provide a set of capabilities via interfaces invoked by the other subsystems. In ODP, an interface is simply an abstraction of the behaviour of an object that consists of a subset of the interactions expected of that object together with the constraints imposed on their occurrence.
The data acquisition subsystem of a research infrastructure collects raw data from registered sources to be stored and made accessible within the infrastructure.
The data acquisition subsystem collects raw data from sensor arrays and other instruments, as well as from human observers, and brings those data into the system. Within the ENVRI-RM, the acquisition subsystem is considered to begin upon point of data entry into the modelled system, the general maintenance and deployment of sensor stations and human observers being outside the scope of ENVRI. Acquisition is typically distributed across a network of observatories and stations. Data acquired is generally assumed to be non-reproducible, being associated with a specific (possibly continuous) event in time and place; as such, the assignment of provenance (particularly data source and timestamp) is essential. Real-time data streams may be temporarily stored, sampled, filtered and processed (e.g., based on applied quality control criteria) before being passed on for curation. Control software is often deployed to manage and schedule the execution and monitoring of data flows. Data collected by the acquisition subsystem must ultimately be transferred to the data curation subsystem for preservation, usually within a specific time period.
The data curation subsystem of a research infrastructure stores, manages and ensures access to all persistent data-sets produced within the infrastructure.
The data curation subsystem facilitates quality control and preservation of scientific data. The subsystem is typically implemented across one or more dedicated data centres. Data handled by the subsystem include raw data products, metadata and processed data; where possible, processed data should be reproducible by executing the same process on the same source data-sets. Operations such as data quality verification, identification, annotation, cataloguing, replication and archival are often provided. Access to curated data from outside the infrastructure is brokered through the data access subsystem. There is usually an emphasis on non-functional requirements for data curation satisfying availability, reliability, utility, throughput, responsiveness, security and scalability criteria.
The data access subsystem of a research infrastructure enables discovery and retrieval of scientific data subject to authorisation.
The data access subsystem enables discovery and retrieval of data housed in data resources managed by the data curation subsystem. Data access subsystems often provide gateways for presenting or delivering data products. Query and search tools may be provided which allow users or upstream services to discover data based on metadata or semantic linkages. Data handled by the access subsystem need not be homogeneous. When supporting heterogeneous data, different types of data (often pulled from a variety of distributed data resources) may be converted into uniform representations with uniform semantics resolved by a data discovery service. The subsystem may provide services for harvesting, compressing and packaging (meta)data as well as encoding services for secure data transfer. Data access is controlled using authentication and authorisation policies. Despite the name, the access subsystem may also provide services for importing data into the infrastructure.
The data processing subsystem of a research infrastructure provides a toolbox of services for performing a variety of data processing tasks.
The data processing subsystem is able to aggregate data from various sources and conduct a range of experiments and analyses upon that data. Data handled by the subsystem are typically derived and recombined via the data access subsystem. The data processing subsystem is expected to offer operations for statistical analysis and data mining as well as facilities for conducting scientific experiments, modelling and simulation, and scientific visualisation. Performance requirements for processing scientific data tend to be concerned with scalability which may be addressable at the level of engineering (e.g., by making use of Grid or Cloud services).
The community support subsystem of a research infrastructure exists to support users of an infrastructure in their interactions with that infrastructure.
The community support subsystem manages, controls and tracks users' activities and supports users to conduct their roles in their communities. Data 'handled' within the subsystem are typically user-generated data and communications. The community support subsystem may support interactive visualisation, standardised authentication, authorisation and accounting protocols, and the use of virtual organisations. The subsystem is considered to encircle the other four subsystems, describing the interface between the research infrastructure and the wider world in which it exists.
As shown in Figure 3.1, amongst the five subsystems can be identified seven major points-of-reference wherein interfaces between subsystems can be implemented.
Figure 3.1: Illustration of the major points-of-reference between different subsystems of the ENVRI-RM.
These points-of-reference are as follows:
Depending on the distribution of resources in an implemented infrastructure, some of these reference points may not be present in the infrastructure. They take particular importance however when considering scenarios where a research infrastructure delegates subsystems to other client infrastructures. For example, EPOS and LifeWatch both delegate data acquisition and some data curation activities to client national or domain-specific infrastructures, but provide data processing services over the data held by those client infrastructures. Thus reference points 4 and 5 become of great importance to the construction of those projects.
Analysis of the common requirements of the six ESFRI project infrastructures affiliated with the ENVRI project has resulted in the identification of a number of common functionalities. These functionalities can be partitioned amongst the five subsystems of the ENVRI-RM and presented as interfaces of each subsystem. They encompass a range of concerns, from the fundamental ( e.g. data collection and storage, data discovery and access and data security) to more specific challenges ( e.g. data versioning, instrument monitoring and interactive visualisation).
Figure 3.2: Radial depiction of ENVRI-RM requirements with the minimal model highlighted.
In order to better manage the range of requirements, and in order to ensure rapid publication of incremental refinements to the ENVRI-RM, as highlighted in Figure 3.2, a minimal model has been identified which describes the fundamental functionality necessary to describe a functional environmental research infrastructure. By initially focusing on this minimal model, it then becomes practical to produce an incomplete specification of the ENVRI-RM which nonetheless reflects the final shape of the ENVRI-RM without the need for significant refactoring. Further development of the ENVRI-RM will focus on designated priority areas based on feedback from the contributing ESFRI project representatives.
The definitions of the minimal set of functions are given as follows. The definition of the full list of common functions are provided in Appendix A.
(A) Data Acquisition Subsystem
Process Control : A functionality that receives input status, applies a set of logic statements or control algorithms, and generates a set of analogue / digital outputs to change the logic states of devices.
Data Collection : A functionality that obtains digital values from a sensor instrument, associating consistent timestamps and necessary metadata.
Data Transmission : A functionality that transfers data over a communication channel using specified network protocols.
(B) Data Curation Subsystem
Data Quality Checking : A functionality that detects and corrects (or remove) corrupt, inconsistent or inaccurate records from datasets.
Data Identification : A functionality that assigns (global) unique identifiers to data contents.
Data Cataloguing : A functionality that associates a data object with one or more metadata objects which contain data descriptions.
Data Product Generation : A functionality that processes data against requirement specifications and standardised formats and descriptions.
Data Storage & Preservation : A functionality that deposits (over the long-term) data and metadata or other supplementary data and methods according to specified policies, and then to make them accessible on request.
(C) Data Access Subsystem
Access Control: A functionality that approves or disapproves of access requests based on specified access policies.
Metadata Harvesting : A functionality that (regularly) collects metadata in agreed formats from different sources.
Resource Registration : A functionality that creates an entry in a resource registry and inserts a resource object or a reference to a resource object with specified representation and semantics.
Data Publication : A functionality that provides clean, well-annotated, anonymity-preserving datasets in a suitable format, and by following specified data-publication and sharing policies to make the datasets publically accessible or to those who agree to certain conditions of use, and to individuals who meet certain professional criteria.
Data Citation : A functionality that assigns an accurate, consistent and standardised reference to a data object, which can be cited in scientific publications.
Data Discovery and Access : A functionality that retrieves requested data from a data resource by using suitable search technology.
(D). Data Processing Subsystem
Data Assimilation : A functionality that combines observational data with output from a numerical model to produce an optimal estimate of the evolving state of the system.
Data Analysis : A functionality that inspects, cleans, transforms data, and to provide data models with the goal of highlighting useful information, suggesting conclusions, and supporting decision making.
Data Mining : A functionality that supports the discovery of patterns in large datasets.
Data Extraction : A functionality that retrieves data out of (unstructured) data sources, including web pages, emails, documents, PDFs, scanned text, mainframe reports, and spool files.
Scientific Modelling and Simulation : A functionality that supports of the generation of abstract, conceptual, graphical or mathematical models, and to run an instance of the model.
(Scientific) Workflow Enactment : A specialisation of Workflow Enactment, which support of composition and execution a series of computational or data manipulation steps, or a workflow, in a scientific application. Important processes should be recorded for provenance purposes.
Data Processing Control : A functionality that initiates the calculation and manage the outputs to be returned to the client.
(E) Community Support Subsystem
Authentication : A functionality that verifies the credentials of a user.
Authorisation : A functionality that specifies access rights to resources.
The ENVRI Reference Model is structured according to the Open Distributed Processing (ODP) standard. As such, the Reference Model is defined from five different perspectives. In the context of ENVRI, which uses ODP to define an 'archetypical' environmental research infrastructure rather than a specific (implemented) infrastructure, three viewpoints take particular priority – the Enterprise , Information and Computational viewpoints.
The remaining two viewpoints ( Engineering and Technology ) are more relevant to specific instances of research infrastructure. Nevertheless, the ENVRI Reference Model will address these viewpoints to some extent in future revisions.
The Enterprise Viewpoint of the ENVRI-RM intends to capture the requirements for an environmental research infrastructure from the perspective of the people who perform their tasks and achieve their goals as mediated by the infrastructure. Modelling in this viewpoint uses a reverse engineering method, which derives the principles and properties of model objects through the analysis of the structure and functionality of the real-world systems.
In a pre-study, we have investigated 6 representative environmental research infrastructures. By examining their computational characteristics, we have identified 5 common subsystems. The fundamental reason of the division of the 5 subsystems is based on the observation that all applications, services and software tools are designed and implemented around 5 major physical resources: the sensor network, the storage, the (internet) communication network, application servers and client devices.
Correspondingly, human activities which interact with the 5 subsystems in order to collaboratively conduct scientific research, from data collection to the deliverance of scientific results, can also be grouped in the same way. Such groups are the so-called communities in ODP. In this viewpoint, we examine what those communities are, what kind of roles they have, and what main behaviours they act out.
A community is a collaboration which consists of a set of roles agreeing their objective to achieve a stated business purpose.
In the ENVRI-RM, we distinguish 5 activities, seen as communities in accordance to the 5 common sub-systems. As shown in Figure 4.1, the 5 communities are, data acquisition , data curation , data publication , data service provision , and data usage community . The definition of the communities are based on their objectives.
Figure 4.1 : Common Communities
A role in a community is a prescribing behaviour that can be performed any number of times concurrently or successively. A role can be either active (typically associated with a human actor) or passive (typically associated with a non-human actor).
In the following, we identify active roles in relation to people associated with a research infrastructure:
Note, an individual may be a member of more than one community.
A system (or part of it) and the hardware facilities which active roles interact with are modelled as passive roles.
The main objectives of the data acquisition community is to bring measurements into the system. The measurement and monitoring models are designed based on the requirements of environmental scientists. Such a design decides what data is to be collected and what metadata is to be associated with it, such as experimental information and instrument conditions. Technicians configure and calibrate a sensor or a sensor network to satisfy the experiment specifications. In the case where human sensors are to be used, observers or measurers input the measures to the system, e.g., by using mobile devices. Data collectors interact with a data acquisition system to prepare the data or control the flow of data and automatically collect and transmit the data.
As shown in Figure 4.2, the following roles are identified in the data acquisition community:
Figure 4.2: Roles in the Data Acquisition Community
The data curation community responds to provide quality data products and maintain the data resources. Consider a typical data curation scenario: when data is being imported into a curation system, a curator will perform the quality checking of the scientific data. Unique identifiers will be assigned to the qualified data, which will then be properly catalogued by associating necessary metadata, and stored or archived. The main human roles interacting with or maintaining a data curation system are data curators who manage the data and storage administrators who manage the storage facilities.
As shown in Figure 4.3, we identified the following roles in this community:
Figure 4.3: Roles in the Data Curation Community
The objectives of the data publication community are to publish data and assist discovery and access. We consider the scenarios described by Kahn's data publication model [ 34 ]: an originator , i.e., a user with digital material to be made available for public access, makes the material into a digital object. A digital object is a data structure whose principal components are digital material, or data, plus a unique identifier for this material, called a handle (and, perhaps, other material). To get a handle , the user requests one from an authorized handle generator . A user may then deposit the digital object in one or more repositories , from which it may be made available to others (subject, to the particular item’s terms and conditions, etc.). Upon depositing a digital object in a repository, its handle and the repository name or IP address is registered with a globally available system of handle servers . Users may subsequently present a handle to a handle server to learn the network names or addresses of repositories in which the corresponding digital object is stored. We use a more general term "PID" instead of " handle" (thus, " PID registry " instead of " handle servers "), and identify the key roles involved in the data publication process including, a data originator, a PID generator, a repository, and a PID registry.
The published data are to be discovered and accessed by data consumers. A semantic mediator is used to facilitate the heterogeneous data discovery.
In summary, as shown in Figure 4.4, the following roles are involved in the data publication community:
Figure 4.4: Roles in the Data Publication Community
The data service provision community provides various application services such as data analysis, mining, simulation and modelling, visualisation, and experimental software tools, in order to facilitate the use of the data. We consider scenarios of service oriented computing paradigm which is adopted by the ENVRI implementation model, and identify the key roles as below. These concepts are along the lines of the existing standards such as OASIS Reference Model for Service Oriented Architecture.
As shown in Figure 4.5, roles in the data service provision community include:
Figure 4.5: Roles in the Data Service Provision Community
The main role in the data usage community is a user who is the ultimate consumer of data, applications and services. Depending on the purposes of usage, a user can be one of the following active roles:
Figure 4.6: Roles in the Data Usage Community
A behaviour of a community is a composition of actions performed by roles normally addressing separate business requirements. In the ENVRI-RM, the modelling of community behaviours is based on analysis of the common requirements of the ENVRI research infrastructure which has resulted in a list of common functions . The initial model focuses on the minimal set of requirements . A community behaviour can be either a single function or a composition of several functions from the function list.
Figure 4.7 depicts the main behaviours of the data acquisition community including:
Figure 4.7: Behaviours of the Data Acquisition Community
The main behaviours of the data curation community are depicted in Figure 4.8 which include:
Figure 4.8: Behaviours of the Data Curation Community
As shown in Figure 4.9, a data publication community may perform the following behaviours:
Figure 4.9: Behaviours of the Data Publication Community
Figure 4.10 depicts the behaviours modelled for the data service provision community, which inclued:
These are general behaviours of a service-oriented computing model. In the context of environmental science research infrastructures, a data service provision community will focus on the implementation of domain special services, in particular those supporting Data Assimilation, Data Analysis , Data Mining, Data Extraction, Scientific Modelling and Simulation, (Scientific) Workflow Enactment. ( See Chapter 2, Terminology and Glossary , for the definitions of these functionalities.)
Figure 4.10: Behaviours of the Data Service Provision Community
Finally, a data usage community may have the following behaviours, which is depicted in Figure 4.11:
Figure 4.11: Behaviours of the Data Usage Community
The goal of the information viewpoint is to model the shared information which is handled by the system to provide a common understanding. The focus lies on the information itself, without considering any platform-specific or implementation details and independent from the computational interfaces and functions that manipulate the data or the nature of technology used to store it. Thus it is similar to ontologies aiming to provide a common and unique interpretation of the shared information entities of a particular domain. The information viewpoint specifies the types of the information handled by the systems and the relationships by these types. It describes how the state of the data in the systems evolve as the system operates and defines the nature of accepted actions by the system and the constraints on the data and its processing.
The Information Viewpoint Model consists of 4 parts:
As shown in Figure 4.13, Information Viewpoint consists of 4 collections of components: Information Objects is the collection of all modelled objects. Only the objects and their statuses that related to the interoperability are provided; Information Action Types is the collection of all action types related to the information objects; Information Object Instances is the collection of instances of information objects; and Data States is the collection of the statuses of persisted data within a data lifecycle.
Information objects model the data manipulated by the system about entities in the real world. They interact with other information objects.
As shown in Figure 4.14, the Information objects consist of:
Figure 4.14: Information Objects
This is the description of the measurement itself
The measurement description can be part of the metadata or the semantic annotation.
As the measurement design is important for the correct interpretation of the data, it is important that the measurement description can be understood. in case of data sharing: can be understood by the outside. a machine readable measurement description is desirable.
Data as it comes out of the measurement instrument, sensor or is produced by the observer
Description, explaining the result of the QA assessment.
This can be:
QA_annotation can be seen as a special annotation. It must be unambiguous and the meaning must be understandable. In case of data sharing it must be understandable to the "outside world".
A unique identifier for any kind of entity
The unique identifier can be
The critical issues of unique identifiers are:
Data about data, in scientific applications is used to describe, explain, locate, or make it easier to retrieve, use, or manage an information resource.
There have been numerous attempts to classify the various types of metadata. As one example, NISO (National Information Standards Organisation) distinguishes between three types of metadata based on their functionality: Descriptive metadata, which describes a resource for purposes, such as discovery and identification; Structural metadata, which indicates how compound objects are put together; and Administrative metadata, which provides information to help manage a resource. But this is not restrictive. Different applications may have different ways to classify their own metadata.
Metadata is generally encoded in a metadata schema which defines a set of metadata elements and the rules governing the use of metadata elements to describe a resource. The characteristics of metadata schema normally include: the number of elements, the name of each element, and the meaning of each element. The definition or meaning of the elements is the semantics of the schema, typically the descriptions of the location, physical attributes, type (i.e., text or image, map or model), and form (i.e., print copy, electronic file). The value of each metadata element is the content. Sometimes there are content rules and syntax rules. The content rules specify how content should be formulated, representation constraints for content, allowable content values and so on. And the syntax rules specify how the elements and their content should be encoded. Some popular syntax used in scientific applications include Some popular syntax includes:
Such syntax encoding allows the metadata to be processed by a computer program.
Many standards for representing scientific metadata have been developed within disciplines, sub-disciplines or individual project or experiments. Some widely used scientific metadata standards include:
Metadata can be fused with the data or kept separately.
Two aspects of metadata give rise to the complexity in management:
A fundamental category of existence and a smallest entity of a conceptual model.
A concept can be part of
Collection of concepts, can be unstructured or structured (e.g. glossary, thesaurus, ontology)
Usually the description of a concept and/or a relation defines the concept in a human readable form.
Concepts within ontologies and their relations can be seen as machine readable sentences. Those sentences can be used to establish a self-description. It is, however, practice today, to have both, the human readable description and the machine readable description. In this sense a conceptual model can also be seen as a collection of human and machine readable sentences.
Conceptual models can reside within the persistence layer of a data provider or a community or outside.
Conceptual models can be fused with the data (e.g. within a network of triple stores) or kept separately.
Enumeration for the different statuses of data
These statuses are just referential statuses
The instantiated chain of data genesis can is expressed in data provenance.
data provenance is a annotation system for any step in the data lifecycle, usually using identifiers for the data in their different statuses and for the description of the different statuses.
data which are persistently stored. In their lifecycle they may have certain statuses:
A backup of the data.
A lot of data are originally stored at unsecure and low level persistence. In order to make them available long term backups are necessary .
Information that traces all the pathway of data
Any action creates an entrance into the data provenance at least containing:
This can be:
Information actions model the information processing in the system. Every action is associated with at least one object. Actions cause state changes in the objects that participate in them.
Figure 4.14 shows a collection of all action types used to model the information viewpoint.
Only actions that are relevant for the interoperability with the outside world are listed.
Request a unique identifier from a server and add it to the data.
Add a pointer to the data to the unique identifier server.
Measure parameter(s) or observe an event.
Perform a measurement or observation to produce measurement results.
Add additional information according to a predefined schema (metadata schema) – partially overlaps annotate data.
Add information about the action and the managed data to the data provenance.
Figure 4.15: Information Objects Action Types
Link data with meaning (concepts of predefined local or global conceptual models).
This can be done by adding a pointer to concepts within a conceptual model to the data.
If e.g. concepts are terms in and SKOS/RDF thesaurus, published as linked data then this would mean entering the URL of the term describing the meaning of the data.
There is no exact borderline between metadata and semantic annotation.
Establish a local or global model of interrelated concepts.
This establishment has the following issues:
Duplicating data to an additional data storage.
A special backup is a long term persistence.
Actions to enhance the quality of data:
Quality checks can be carried out at any point in the chain of data genesis.
Quality checks can be supported by software developed especially for that process (e.g. statistic tolerance checks).
Execute a sequence of metadata / data request --> interpret result --> do a new request.
Usually this sequence helps to deepen the knowledge about the data.
Classically this sequence can:
It can be supported by special software that helps to carry out that sequence of data request and interpretation of results.
Review, which produces data, which will never be changed again.
Produce the status "finally reviewed". It is good practice to simultaneously add an annotation for this status.
Usually this is coupled with some sort of archiving and versioning.
Execute transformation rules for values (mapping from one unit to another unit) or translation rules for concepts (translating the meaning from one conceptual model to another conceptual model, e.g. translating code lists).
In general use data:
Make data accessible for the public .
This can be done by:
Send a data request to a data container.
In general there exist two different types of data access:
(e.g. OGC services first call a web content service then call a web feature service)
Any request can be directly sent to a service or distributed by a broker.
Send a data request to a metadata container.
In general there exist two different types of data access:
(e.g. OGC services first call a web content service then call a web feature service)
Any request can be directly sent to a service or distributed by a broker.
Mapping of data and/or concepts has to be expressed by rules.
These rules have to be written in a language that can be understood by a mapping machine.
The mapping rule has at least to include the following information:
Specify the details of the method of observations/measurements.
This e.g. includes the specification of a measurement device type and its settings, measurement/ observation intervals.
Specify design of sampling or observation:
Make data persistent.
Figure 4.16 shows the collection of instances of information objects which are information objects existing more than once having several instances.
Instances of information objects are needed for two purposes:
Concepts with a commitment of a whole data sharing community. Usually those concepts are part of global conceptual models (global Thesauri like GEMET / EuroVoc / AGROVOC or global ontologies like Gene Ontology , ... ).
A concept of
A concept can be local or global depending on the size of the community which commits to it and only if considered in relation to each other.
Figure 4.16: Instances of Information Objects
In summary, the statuses of persisted data within the data lifecycle is shown in Figure 4.17.
The overview of the pathway of the data lifecycle is shown in Figure 4.18, which is part of a comprehensive overview of all actions applied on objects (See Appendix B).
Static Schemata specifies a minimum set of constraints for data sharing, in order to:
There static schemata are described, state after measurements or observation , global interpretability of data , and data ready to use .
The state after measurements or observation is illustrated in Figure 4.19, which is a c ollection of constraints which help to avoid loss of information or wrong information drawn out of data.
Figure 4.19: State after measurements or observation
The global interpretability of data is illustrated in Figure 4.20, which is a collection of constraints which are needed for transporting the meaning of data in order to help for correct interpretation by data users of the "global world".
Figure 4.20: Global Interpretability of Data
Figure 4.21 describes the status data ready used which is a status necessary for the data access from outside.
Figure 4.21: Data Ready to Use
In order to provide an easy way of observing, this subsection regroup the information objects into subsystems defined in Chapter 3.
The ongoing Data Acquisition
Precondition: working hypothesis:
A design for observation / measurements, is established and described in the setup description. Then the measurement methode is specified including specifications for the measurement devices.
The measurement or observation leads to results which are persisted (stored).
Figure 4.22: Permanent Data Acquisition
The ongoing curation of data
Precondition:
Raw data are stored results of observations and measurements
All the data curation steps are optional
Figure 4.23: Permanent Data Curation
The ongoing access to data:
Figure 4.24: Permanent Data Access
Tracing of provenance bridges all subsystems recording the whole life cycle of data. It records actions of objects, action involved objects, and statuses of objects.
Figure 4.25: Tracing of Provenance
An environmental research infrastructure provides scientific data and the means to interact with that data. A number of different services and processes must be implemented if the infrastructure is to be able to acquire, store and provide access to a significant body of data. Indeed simply providing the ability to catalogue and annotate data, or to even visualise data requests, requires that a number of core functions be realised.
Each of the five subsystems of the ENVRI reference model must provide a number of computational objects which can be distributed across implemented services and resources in a concrete research infrastructure. Each object provides a number of interfaces by which its functionality can be accessed; each object also has facilities to invoke the functionality possessed by other objects by those interfaces. The computational viewpoint of each of the five subsystems is provided along with descriptions of their computational objects and interfaces:
A number of reference interactions describing the interaction between objects in different subsystems is also provided:
The aggregation of these interactions forms a minimal computational model for the ENVRI reference model.
The computational viewpoint of ODP is concerned with the modelling of computational objects and their interfaces. Each interface provides a set of invocations by which an object can be interacted with or by which an object can interact with another. Complex interactions of objects can be realised by the use of binding objects which encapsulate the coordination of multiple objects via multiple interfaces. Specific interactions can be specified in more detail, breaking them down into their constituent invocations.
Major computational interactions are modelled as collaborations between computational objects. Objects are linked via interfaces: operation interfaces are used to specify the service invocations which can be called between objects whilst stream interfaces describe data channels between objects by which bulk datasets can be moved.
Each reference point diagram involves a central binding object used to connect all computational objects which must interact at this reference point. The binding object serves to illustrate which computational objects (and thus computational functions) should be available at some point during the delivery of data; whether a particular object is invoked at the data source, destination or some other interim point is not considered to be important within the reference model.
The continuous acquisition of data requires the existence of an integrated network of instruments from which data can be extracted. Whether the instruments represent individual sensors, sensor arrays or something even further removed from the physical recording of measurements, an 'instrument' from the computational perspective represents a data source which produces new data over time and which can be directly queried for that data at any time. Integration, in this instance, refers to the fact that these instruments are considered part of the infrastructure (as opposed to being provided by an outside agency). Instruments can therefore be read from within the infrastructure without the need for the data access subsystem to verify the instrument's identity and validate any requests made.
The data acquisition subsystem is computationally described as a set of instrument hosts (representing the computational functionality of instruments) monitored and managed by an acquisition service . The acquisition service fields all directives sent to the instrument network (via its acquire data interface), schedules data collection and performs any re-calibrations of instruments requested within the instruments' operation boundaries (via the configure instrument interface). It is also responsible for negotiating the import of data into the data curation subsystem via the import data interface. Under the oversight of the acquisition service, instrument hosts can deliver raw data to data resource hosts within the data curation subsystem in order to (continuously) move data directly from instruments to data stores.
Oversight service for an integrated data acquisition instrument network.
An acquisition service object collects the computational functions required to manage a network of data-collecting instruments integrated into an infrastructure. Any service or set of services implementing the acquisition service is responsible for integrating new instruments into the network, re-calibrating existing instruments (within the permitted operational bounds of those instruments) and managing the extraction of data from all instruments in the data acquisition network.
The acquisition service invokes two interfaces:
The acquisition service provides one interface:
An instrument (integrated raw data source) in the data acquisition subsystem.
An instrument is a source of raw data for a research infrastructure. Instruments are distinguished from other data sources in that they are fully integrated into a data acquisition network within an infrastructure's data acquisition subsystem, which allows their status to be continually monitored and their output directly channelled into a data curation system without the intercession of an intermediate broker. Instrument hosts collect the computational functions required to interact with instruments directly.
An instrument host can deliver raw data to the data curation subsystem.
An instrument host provides one interface:
Scientific data must be collected, catalogued and made accessible to all authorised users for indefinite periods of time. The accessibility requirement in particular dictates that infrastructures provide facilities to ensure the easy availability of data products, generally by replication (avoiding data loss by system failure, as well as more optimised data retrieval for larger, distributed infrastructures), the use of persistent identifiers (providing a long-term means to precisely locate specific data-sets) and the use of metadata catalogues (providing a means to search for and discover data).
The data curation subsystem is computationally described as a set of data resource hosts (representing the computational functionality of data resources as presented to the larger infrastructure) monitored and managed by a number of curation services via their configure data resource interfaces. These curation services include:
Data resource hosts represent logically distinct data stores with their own internal data management regimes. These stores can be queried directly (without the need to export whole datasets) via the query resource interface. All data resources must be able to import datasets and export datasets on demand.
Oversight service for metadata management with the data curation subsystem.
A metadata service object collects the computational functions required to manage the attribution and distribution of metadata with regard to data-sets stored within the data curation subsystem of a research infrastructure. No assumption is made as to whether metadata is stored separately or integrated into data-sets. Exported data-sets should be packaged with their metadata, a responsibility fulfilled by the metadata service.
The catalogue service may invoke one interface:
The catalogue service provides two interfaces:
Oversight service for export of data from the data curation subsystem.
A data export service object collects the computational functions required to coordinate the export of datasets from the data curation subsystem of a research infrastructure. In particular, it is responsible for locating requested data and selecting the data resources from which the data will be retrieved, then ensuring that the retrieved data is packaged with all pertinent metadata for use outside the infrastructure.
The data export service invokes two interfaces:
The data export service provides one interface:
Oversight service for the import of new data into the data curation subsystem.
A data import service object collects the computational functions required to integrate new data into the data curation subsystem of a research infrastructure. The import service coordinates the collection of raw data delivered by the data acquisition subsystem; this entails the refactoring of data streams into discrete, identifiable datasets and the further delivery of those data-sets to suitable data resources for curation. The import service also handles the ingestion of user-provided datasets where permitted, and the curation of derived results generated by the data processing subsystem.
The data import service invokes two interfaces:
The data import service provides one interface:
Used by the data curation subsystem to perform simple processing of newly imported / exported data and invoke the data processing subsystem for more complex tasks.
The data processor object encapsulates functions within the data curation subsystem to refactor and prepare data products going into and out of the subsystem. Importantly, the data processor also allows the data processing subsystem to interact with data integration and retrieval processes, permitting standard process workflows to be put in place between the data curation and other subsystems.
A data processor can invoke one interface:
A data processor provides one interface:
Oversight service for the querying of data resources in the data curation subsystem.
The data query service collects the computational functions required to query data resources in the data curation subsystem of a research infrastructure. The query service is responsible for the distribution and aggregation of sub-queries where data from multiple datasets has been requested. Operations which actually require the movement of entire datasets onto external resources are handled by the data export service.
The data query service invokes one interface:
The data query service provides one interface:
A data store within the data curation subsystem.
A data resource object represents a data store integrated into the data curation subsystem of a research infrastructure. It contains the computational functionality required to store and maintain data products produced within the infrastructure, as well as to provide access on demand to authorised agents. A data resource must be able to interpret internal data requests, obey directives from curation services and adhere to data management policies such as (where applicable) versioning, replication and persistence.
A data resource host can import and export datasets .
A data resource host provides three interfaces:
Registration service for integrating new data resources into the data curation subsystem.
The data resource registrar collects the computational functions required to register and configure new data resources within the data curation subsystem of a research infrastructure. The registrar ensures that new resources can be identified within the system and provide all expected functions of a data resource; this entails the deployment of a new data resource host object to handle all service invocations.
A data resource registrar provides one interface:
Oversight service for identifier assignment and resolution.
Persistent identifiers are generated by a service either provided within or outside of the infrastructure. The identification service object collects the functions required to interact with such a service, acquiring identifiers for all artefacts which require them and providing a resolution service for cited identities.
Different versions of artefacts, where maintained separately, are assumed to have different identifiers, but those identifiers should share a common root such that the family of versions of a given artefact can be retrieved in one transaction, or only the most recent (or otherwise dominant) version is returned.
An identification service provides two interfaces:
A core responsibility of a research infrastructure is to provide access to the scientific data held within to the broader scientific community without. This access can be provided in a number of ways: by allowing the export of data-sets, by allowing the data to be read via a client, by providing interactive visualisations via a portal, etc . Beyond the actual mechanism of data retrieval however is the issue of data discovery. Specific data-sets may be found via citation (the de-referencing of persistent identifiers), by browsing catalogues or by search over metadata. Data requests can be formulated which extract specific data elements from one or more datasets, returning the aggregation of all discovered results.
The data access subsystem provides a number of data access brokers, which act as intermediaries for access of data held within the data curation subsystem. These brokers are responsible for verifying the agents making access requests and validating those requests before sending them on to the relevant data curation service.
The data access subsystem plays host to the authorisation service (which can authorise requests via its authorise request interface) and the authentication service (which can authenticate users via its authenticate user interface); these services are invoked by all brokers before forwarding requests onto the data curation subsystem.
Broker for facilitating data access requests.
An access broker object intercedes between the data access subsystem and the data curation subsystem, collecting the computational functions required to make direct requests of data curations services on behalf of a user. It is the responsibility of an access broker to validate all requests and to verify the identity and access privileges of agents making requests. It is not permitted for an outside agency or service to access the data curation subsystem of a research infrastructure by any means other than via an access broker.
An access broker can invoke two interfaces:
An access broker provides one interface:
Security service responsible for the authentication of external agents making requests of infrastructure services.
An authentication service provides one interface:
Security service responsible for the authorisation of all requests made of infrastructure services by external agents.
An authorisation service provides one interface:
Access service for resolving citations used in external publications to refer to artefacts within a research infrastructure.
A citation service provides one interface:
Broker for facilitating data upload requests from external contributors.
An upload broker object intercedes between the data access subsystem and the data curation subsystem, collecting the computational functions required to upload user data into a research infrastructure for curation. It is the responsibility of an upload broker to validate all upload requests and to verify the identity and privileges of agents wanting to contribute data. It is not permitted for user data to be inserted into the data curation subsystem of a research infrastructure by any means other than via an upload broker.
An upload broker can invoke one interface:
An upload broker provides one interface:
The processing of data can be tightly integrated into data handling systems, or can be delegated to a separate collection of services invoked on demand; generally the more involved the processing, the more likely that separate resources will be required, particularly for tasks which require some form of high performance computing. The provision of dedicated processing services becomes significantly more important when large quantities (terabyte scale and higher) of data are being curated within a research infrastructure, especially for scientific data, which is often subject to extensive post-processing and analysis in order to extract new results. The data processing subsystem of an infrastructure encapsulates the dedicated processing services made available to that infrastructure, either within the infrastructure itself or delegated to a client infrastructure.
The data processing subsystem is computationally described as a set of execution resource hosts (representing the computational functionality of registered execution resources) monitored and managed by a process coordination service . The process coordination service delegates all processing tasks sent to the data processing subsystem (received via its stage task interface) to particular execution resources, coordinates multi-stage workflows and initiates execution (via the execute tasks interfaces of selected execution resource hosts). Data may need to be staged onto individual execution resources and results retrieved for curation – this is handled by their hosts via the stage task data and deliver results stream interfaces respectively.
It is also possible to register new resources as execution resources, should the data processing subsystem need scaling up. This is handled by the process resource registrar via its register process resource interface.
Part of the execution platform provided by the data processing subsystem.
An execution resource host can be used to stage task data and deliver results back into the data curation subsystem.
An execution resource host provides one interface:
Oversight service for data processing tasks deployed on infrastructure execution resources.
A process coordination service can invoke three interfaces:
A process coordination service provides one interface:
Registration service for integrating new execution resources into the data processing subsystem.
A process resource registrar provides one interface:
As important as the internal composition and management of an infrastructure is, equally important is how the broader scientific community interacts with that infrastructure. The principle way that prospective users interact with an infrastructure is via infrastructure-managed portals (generically referred to as 'scientific gateways'). Given a multi-faceted infrastructure, there may exist numerous gateways into the infrastructure engineered towards particular user roles and usecases. These gateways may provide additional user services beyond simple access to curated data and to data processing services; they may also provide 'social computing' services such as user profiling, reputation mechanisms and workflow sharing.
The community support subsystem also encompasses 'external' resources – resources not part of the infrastructure (or its federated client systems), but which are instead transitory resources temporarily brought in by a user as a data source or destination for bulk data transfers.
The community support subsystem plays host to a number of gateway host objects, representing the computational functionality of various assorted scientific gateways and portals.
The community support subsystem is considered to contain external resources , from which data can be imported into the infrastructure (via the import data product steam interface) or into which data can be exported from the infrastructure (via the export data product stream interface).
Figure 4.30: Community Support
A gateway service or equivalent construct used to interact with the infrastructure in an administrative capacity.
An admin gateway host can invoke one interface:
A gateway service or equivalent construct used by the scientific community to interact with the infrastructure.
A community gateway host can invoke three interfaces:
A resource not within the infrastructure acting temporarily as a data source or destination for data export.
An external resource is not assumed to provide or require any particular operational interfaces, but is assumed to be able to send and receive data via its import data product and export data product stream interfaces.
Exporting data from curated data resources requires that the export be brokered by the data access subsystem before data can be retrieved from the data curation subsystem and delivered to a designated external resource. The interaction between the data access and data curation subsystems in this context is illustrated below.
Figure 4.31: Brokered Data Export
Brokered data export provides three exposed interfaces: access data , by which the services of the data access broker are enlisted; request metadata , by which other parts of the infrastructure can contribute additional metadata information to package with the exported data product; and stage task , by which data processing services can be enlisted in the post-processing of exported data.
Any request for the export of datasets (base or derived) from a research infrastructure must be verified and validated by an access broker in accordance with the data access policies of the infrastructure. The retrieval of curated data is managed by collaboration between the access broker and the data export service via its export data interface. It is assumed that the access broker has been provided information about the data source to which data is to be delivered; it is likewise assumed that this external resource is accessible via any and all required protocols. The data export service can prepare and configure data resources for data ingestion via the configure data resource interfaces on the respective data resource hosts .
The data export service oversees the interaction of services involved in the extraction of curated data out of the data curation subsystem and the establishment of delivery channels between external resources and internal curation resources – this is represented by the coordinate data export interface to the brokered data export binding object. The role of the brokered data export binding object is to link all computational objects involved in the retrieval of a data product from the data curation subsystem and the extraction of all pertinent metadata.
Given a delivery channel established between an external resource and a data resource host, overseen by the data export service, all curated data must be re-packaged into a self-describing data product which can be examined independently of the research infrastructure. Any externally-meaningful metadata associated with exported data must be packaged with the extracted data product. In certain circumstances, there may also be additional post-processing required in order to ensure that the data fulfils the original access request and does not violate any additional security constraints. The extraction of catalogued metadata is handled by the catalogue service invoked via the export metadata interface. Any post-processing is handled by the data processor invoked via the process data interface; the data processor may use its stage task interface to further invoke services provided by the data processing subsystem if necessary to carry out its function.
Importing data from sources other than the acquisition network requires that the import be brokered by the data access subsystem before data can be delivered into the data curation subsystem. The interactions between the two subsystems in this context is illustrated below.
Figure 4.32: Brokered Data Import
Brokered data import provides two exposed interfaces: upload data , by which the services of the data upload broker are enlisted; and stage task , by which data processing services can be enlisted in the pre-processing of newly curated data.
Any request to upload data into a research infrastructure must be verified and validated by an upload broker in accordance with the data access policies of the infrastructure. The curation of imported data is managed by collaboration between the upload broker and the data import service via its import data interface. It is assumed that the upload broker has been provided information about the data source from which data is to be extracted; it is likewise assumed that this external resource is accessible via any and all required protocols. The data import service can prepare and configure data resources for data ingestion via the configure data resource interfaces on the respective data resource hosts .
The data import service oversees the interaction of services involved in the ingestion of imported data into the data curation subsystem and the establishment of delivery channels between external resources and internal curation resources – this is represented by the coordinate data import interface to the brokered data import binding object. The role of the brokered data import binding object is to link all computational objects involved in the delivery of external data products into the data curation subsystem and its conversion (where necessary) into curated datasets stored within data resources.
Given a delivery channel established between an external resource and a data resource host, overseen by the data import service, any data product must be re-packaged into an infrastructure-compliant dataset with its own persistent identifier. Moreover any metadata stored within the data product or evident in its provenance should be extracted and added to all pertinent catalogues. In certain circumstances, there may also be additional pre-processing required before the data can be stored. The assignment of persistent identifiers is handled by the identification service invoked via its acquire identifier interface. The cataloguing of extracted metadata is handled by the catalogue service invoked via its catalogue data interface. Any pre-processing is handled by the data processor invoked via its process data interface; the data processor may use its process data interface to further invoke services provided by the data processing subsystem if necessary to carry out its function.
Querying curated data resources requires that the request be brokered by the data access subsystem before any results will be retrieved from the data curation subsystem and delivered to request source. The interaction between the data access and data curation subsystems in this context is illustrated below.
Figure 4.33: Brokered Data Query
Brokered data query provides two exposed interfaces: access data , by which the services of the data access broker are enlisted; and stage task , by which data processing services can be enlisted in the processing of the query should any such extraordinary processing be required.
Any data request made to a research infrastructure must be verified and validated by an access broker in accordance with the data access policies of the infrastructure. The querying of curated data is managed by collaboration between the access broker and the data query service via its query data interface. Any results not requiring a full export of a (derived) dataset are returned via this interface. The data query service is then responsible for distributing the query across all necessary data resources (which may be greater than one for complex data requests) via the distribute query interface to the brokered data query binding object. The role of the brokered data query binding object is to link all computational objects involved in the aggregation of results to a data query request including possibly objects belonging to the data processing subsystem in complex cases.
Individual data resources identified as having required data by the data query service are queried via the query resource interfaces of their respective data resource host objects. If processing of data (either before or after aggregation of results) is required, the data processor can be invoked via its process data interface; for particularly complex requests, the data processor can invoke the services of the data process subsystem at large via the stage task interface.
The citation of datasets involves reference to persistent identifiers assigned to objects by a research infrastructure. Such citations are resolved by referring back to the infrastructure, which can then return a report describing the data cited. This simple interaction is illustrated below.
Citation provides two exposed interfaces: resolve citation , by which the link between persistent identifiers and data objects is established; and access data , by which datasets referred to by a citation can be examined or retrieved.
The citation of datasets is managed by the citation service in the data access subsystem, which extracts the persistent identifier embedded in a given citation and then refers to the identification service in the data curation subsystem via its resolve identifier interface.
The internal staging of data within an infrastructure for processing requires coordination between the data processing subsystem (which handles the actual processing workflow) and the data curation subsystem (which holds all scientific datasets within the infrastructure). The interactions between the two subsystems in this context is illustrated below.
Figure 4.34: Internal Data Staging
Internal data staging provides two exposed interfaces: stage task , by which a processing workflow is specified and scheduled; and import data , by which derived datasets and results produced in processing can be formally ingested into the infrastructure.
Internal data staging is directed by the process coordination service in accordance with the task specifications given to it and any available infrastructure resources. The export of curated datasets onto execution resources is managed by collaboration between the process coordination service and the data export service via its export data interface. The process coordination service can configure execution resources and initiate processing of selected execution resources within the infrastructure by invoking the execute task interfaces of the execution resource hosts representing those resources. The data export service can prepare data resources for data export via the configure data resource interfaces on their respective data resource hosts .
The data export service oversees the interaction of services involved in the delivery of curated data into the data processing subsystem and the establishment of delivery channels between data and execution resources – this is represented by the coordinate data export interface to the internal data staging binding object. The role of the internal data staging binding object is to link all computational objects involved in the staging of curated data onto execution resources. Note that it is perfectly valid for a data resource to also be an execution resource, likely for more efficient conduct of standard analyses on datasets. In this case, the 'staging' of data becomes trivial, with the data export service merely confirming that the data is suitably prepared for the processing task to be instigated by the process coordination service.
Once data has been confirmed as having been staged onto a suitable execution resource, data processing can proceed on that resource and results imported back into the data curation subsystem either as revisions of existing datasets or as new derived data.
The formal ingestion of data derived during internal processing of existing datasets within an infrastructure requires coordination between the data processing subsystem (which produces the new results) and the data curation subsystem (which is responsible for all scientific datasets within the infrastructure). The interactions between the two subsystems in this context is illustrated below.
Figure 4.35: Processed Data Import
Processed data import provides two exposed interfaces: stage task , by which a processing workflow is specified and scheduled; and export data , by which more data can be extracted from the data curation subsystem for further processing.
Processed data import is directed by the process coordination service in accordance with the task specifications given to it and the existence of uncurated results on any of the infrastructure resources it manages. The retrieval of derived datasets from execution resources is managed by collaboration between the process coordination service and the data import service via its import data interface. The process coordination service can configure execution resources and initiate extraction of derived data from selected execution resources within the infrastructure by invoking the execute task interfaces of the execution resource hosts representing those resources. The data import service can prepare selected data resources for data ingestion via the configure data resource interfaces on their respective data resource hosts .
The data export service oversees the interaction of services involved in the delivery of derived data into the data curation subsystem and the establishment of delivery channels between data and execution resources – this is represented by the coordinate data import interface to the processed data import binding object. The role of the processed data import binding object is to link all computational objects involved in the retrieval of processed data from execution resources. Note that it is perfectly valid for an execution resource to also be the data resource which is to curate a given set of results. In this case, the 'retrieval' of data becomes trivial, with the data import service merely confirming that the data is suitably prepared for curation.
Given a delivery channel established between an execution resource host and a data resource host, overseen by the data import service, any derived data product must be re-packaged into an infrastructure-compliant dataset with its own persistent identifier. Moreover any metadata stored within the data product or evident in its provenance should be extracted and added to all pertinent catalogues. The assignment of persistent identifiers is handled by the identification service invoked via its acquire identifier interface. The cataloguing of extracted metadata is handled by the catalogue service invoked via its catalogue data interface. It is assumed that any further processing of data required for curation is handled as part of the data processing activity already extant when producing the data.
The collection of raw scientific data requires coordination between the data acquisition subsystem (which extracts the raw data from instruments) and the data curation subsystem (which packages and stores the data). The interactions between the two subsystems in this context is illustrated in Figure 4.36.
Raw data collection provides two exposed interfaces: acquire data , by which the behaviour of the acquisition service can be configured; and stage task , by which data processing services can be enlisted in the pre-processing of newly curated data.
Figure 4.36: Raw Data Collection
Raw data acquisition is motivated by the acquisition service in accordance with its configuration. The curation of acquired data is managed by collaboration between the acquisition service and the data import service via its import data interface. The acquisition service can configure and calibrate the behaviour of instruments in the acquisition network by invoking the configure instrument interface for the instrument hosts representing those instruments. The data import service can prepare and configure data resources for data ingestion via the configure data resource interface on the respective data resource hosts .
The data import service oversees the interaction of services involved in the ingestion of raw data into the data curation subsystem and the establishment of delivery channels between instruments and data resources – this is represented by the coordinate data import interface to the raw data collection binding object. The role of the raw data collection binding object is to link all computational objects involved in the delivery of raw data into the data curation subsystem and its conversion into curated datasets stored within data resources.
Given a delivery channel established between an instrument host and a data resource host, overseen by the data import service, raw data must be packaged into datasets in accordance with data curation policy. Each dataset should be assigned a persistent identifier and appropriate metadata describing its nature and provenance and subject to any pre-processing dictated by the research infrastructure. The assignment of persistent identifiers is handled by the identification service invoked via its acquire identifier interface. Any metadata generated should be catalogued – this is handled by the catalogue service via its catalogue data interface. Any pre-processing is handled by the data processor invoked via its process data interface; the data processor can use its stage task interface to further invoke services provided by the data processing subsystem if necessary to carry out its function.
Making a request of a research infrastructure requires that the request be brokered by the data access subsystem, which entails that the source of the request be authenticated and the request itself authorised. The interaction between a community gateway or portal and the data access subsystem in the context of data management is illustrated below.
Figure 4.37: Request Verification
Request verification provides three interfaces into a research infrastructure depending on the data request being made of it: query data , by which the scientific data held by the infrastructure can be queried; export data , by which curated datasets can be retrieved from the infrastructure and delivered to a particular location; and import data , by which data held by external agents or services can be brought into the infrastructure for curation.
Requests are assumed to be made by some sort of community tool such as a scientific gateway, represented by a community gateway host . Data requests are made using the request access interface of a request verification binding object. The role of a request verification binding object is to link together all computational objects involved in verifying and then passing on a data request to the data access subsystem. Thus the intercession of the authentication service (via its authenticate user interface) and the authorisation service (via its authorise request interface) is required. Only once these services have verified the request as being permissible can the request be passed on to the appropriate broker.
Requests involving the reading or export of datasets held within the data curation subsystem are brokered by an access broker via its access data interface. Requests involving the uploading of data into the data curation subsystem are brokered by an upload broker via its upload data interface. Complex requests may involve multiple brokers acting in parallel.
The registration of new resources (whether to support data curation or data processing) requires that a request be made via suitable administrative channels, authorised and handled by some kind of registration service situated in the affected subsystem. The interaction between an administrative gateway or portal and the relevant subsystems in this context is illustrated below.
Figure 4.38: Resource Registration
Resource registration is considered to be a self-contained interaction. An administrative gateway represented by an admin gateway host invokes the request registration interface of a resource registration binding object. The role of the resource registration binding object is to link all computational objects involved in the registration of resources. Registration requires that the source of the registration request be authenticated, the request itself be authorised and the resource being registered be assigned a persistent identifier within the infrastructure and where appropriate an identifier which can identify the resource outside of it. Authentication is handled by the authentication service via its authenticate user interface. Authorisation is handled by the authorisation service via its authorise request interface. Identification is handled by the identification service via its acquire identifier interface.
If a data resource is being registered, then the resource profile is passed on to the data curation subsystem; specifically a data resource registrar via its register data resource interface. If a processing or execution resource is being registered, then the resource profile is passed on to the data processing subsystem; specifically a process resource registrar via its register process resource interface. It is possible for a resource to be both a data resource and an execution resource, in which case both subsystems must be informed and must separately ensure adequate configuration.
Making a request of a research infrastructure requires that the request be brokered by the data access subsystem, which entails that the source of the request be authenticated and the request itself authorised. The interaction between a community gateway or portal and the data access subsystem in the context of data processing is illustrated below.
Figure 4.39: Task Verification
Task verification provides three interfaces into a research infrastructure depending on the requirements of the request being made of it: execute task , by which requested tasks are scheduled to be executed on resources belonging to the infrastructure; export data , by which curated datasets required for task execution are moved onto execution resources; and import data , by which results generated by data processing are moved into the data curation subsystem.
Requests are assumed to be made by some sort of community tool such as a scientific gateway, represented by a community gateway host . Data processing requests are made using the request task interface of a task verification binding object. The role of a task verification binding object is to link together all computational objects involved in verifying and then passing on a task request to the data processing subsystem. Thus the intercession of the authentication service (via its authenticate user interface) and the authorisation service (via its authorise request interface) is required. Only once these services have verified the request as being permissible can the request be passed on to the process coordination service via its stage task interface.
All data processing tasks are overseen by the process coordination service, which coordinates the execution of individual tasks via its execute task interface, stages data via its export data interface and arranges the curation of any results produced via its import data interface. A complex task may involve multiple invocations of these interfaces.
Much work remains. Stronger correspondence between the three primary viewpoints is necessary to ensure that the three sub-models are synchronised in concept and execution. Further refactoring of individual components and further development of individual elements is to be expected as well. Further development of the presentation of the model is also essential, in order to both improve clarity to readers not expert in ODP and in order to promote a coherent position.
[1] W. Los, "Introduction to ENVRI and the workshop objectives," in ENVRI Frascati Meeting 5-7 Feb 2013 , Presentation. Frascati, Italy, 2013.
[2] "Global Change: Towards global research infrastructures," European Commission, Directorate-General For Research and Innovation , 2012.
[3] S. Sorvari. "Envrionmental reseach in harmony," International Innovation - Disseminating, science researhc and technology . Dec. 2012 Page 28, 2012. Available: http://www.research-europe.com/magazine/ENVIRONMENT/2012-15/index.html
[4] SO/IEC, "ISO/IEC 10746-1: Information technology--Open Distributed Processing--Reference model: Overview," ISO/IEC Standard , 1998.
[5] ISO/IEC, "ISO/IEC 10746-2: Information technology--Open Distributed Processing--Reference model: Foundations," ISO/IEC Standard , 2009.
[6] ISO/IEC, "ISO/IEC 10746-3: Information technology--Open Distributed Processing--Reference model: Architecture," ISO/IEC Standard , 2009.
[7] ISO/IEC, "ISO/IEC 10746-4: Information technology--Open Distributed Processing--Reference model: Architecture Semantics, " ISO/IEC Standard , 1998.
[8] OASIS, "Reference Model for Service Oriented Architecture 1.0," OASIS Standard , 2006.
[9] L. Candela, and A. Nardi, "The Digital Library Reference Model," DL.org , 2010.
[10] ISO/IEC, "Open System Interconnection (OSI), ISO/IEC 7498-1," ISO/IEC Standard , 1994.
[11] CCSDS, "Reference Model for an Open Archival Information System (OAIS)," CCSDS Standard , 2012.
[12] C. Atkinson, M. Gutheil, and K. Kiko, "On the Relationship of Ontologies and Models," Lecture Notes in Informatics, Gesellschaft für Informatik, Bonn, INI Proceedings, 1996.
[13] D. C. Schmidt, "Model-Driven Engineering," IEEE Computer vol. 39, 2006.
[14] N. F. Noy, and D. L. McGuinness, "Ontology Development 101: A Guide to Creating Your First Ontology," Stanford Knowledge Systems Laboratory Technical Report KSL-01-05 and Stanford Medical Informatics Technical Report SMI-2001-0880 , 2001.
[15] P. Tetlow, J. Z. Pan, D. Oberle, E. Wallace, M. Uschold, and E. Kendall, "Ontology Driven Architectures and Potential Uses of the Semantic Web in Systems and Software Engineering," W3C Standard , 2006.
[16] SEKE, "International Conference on Software Engineering (SEKE 2005)".
[17] VORTE, "International Workshop on Vocabularies, Ontologies and Rules for The Enterprise (VORTE 2005-2013)".
[18] MDSW, "The Model-Driven Semantic Web Workshop (MDSW 2004)".
[19] SWESE, "Workshop on Semantic Web Enabled Software Engineering (SWESE 2005-2007)".
[20] ONTOSE, "Workshop on Ontology, Conceptualizations and Epistemology of Software and Systems Engineering (ONTOSE 2005-2009)".
[21] WoMM, "Workshop on Meta-Modeling and Corresponding Tools (WoMM 2005)" .
[22] Kwaaitaal, M. Hoogeveen, and T. V. D. Weide, "A Reference Model for the Impact of Standardisation on Multimedia Database Management Systems," Computer Standards & Interfaces, vol. 16, pp. 45-54, 1994.
[23] OGC, "OGC Reference Model," Open Geospatial Consortium , OGC Standard, 2011.
[24] T. Uslander, "Reference Model for the ORCHESTRA Architecture (RM-OA) V2,", Open Geospatial Consortium , OGC Standard, 2007.
[25] V. Hernandez-Ernst, et al. , "LIFEWATCH. Deliverable 5.1.3: Data & Modelling Tool Structures -- Reference Model," the EU LifeWatch consortium , 2010.
[26] D. Hollingsworth, "The Workflow Reference Model," the Workflow Management Coalition , 1995.
[27] Mayk, and W. C. Regli, "Agent Systems Reference Model Release Version 1.0a," US Army Communications and Electronics Command Research Development and Engineering Center (CERDEC) , 2006.
[28] E. H. Chi, and J. T. Riedl, "An Operator Interaction Framework for Visualization Systems," Symposium on Information Visualization (InfoVis '98), 1998.
[29] E. H. Chi, "A Taxonomy of Visualisation Techniques using the Data State Reference Model," Proceedings of the IEEE Symposium on Information Visualization 2000 (InfoVis'00), 2000.
[30] N. Koch, and M. Wirsing, "The Munich Reference Model for Adaptive Hypermedia Applications," in 2nd International Conference on Adaptive Hypermedia and Adaptive Web-based Systems, Proceedings. P. De Bra, P. Brusilovsky, and R. Conejo (eds.) LNCS 2347, ©Springer Verlag, pp. 213-222, 2002.
[31] OMG, "Data Distribution Service for Real-time Systems Version 1.2," OMG Standard , 2007.
[32] OASIS, "Content Management Interoperability Services (CMIS) Version 1.0," OASIS Stanard , 2011.
[33] Hardisty, "WP3 Progress and Issues for Full Plenary with SAB, Wednesday 6th Feb 2013," in ENVRI Frascati Meeting, 5-7 Feb 2013 , ed. Frascati, Italy, 2013
[34] R. Kahn and R. Wilensky, "A framework for distributed digital object services", International Journal on Digital Libraries (2006) 6(2):115-123 , 2006.
[35] Barros, M. Dumas and P. Oaks, "A Critical Overview of the Web Services Choreography Description Language (WS_CDL)", BPTrends , Mar. 2005.
|
A |
Data Acquisition Subsystem |
|
|
No |
Functions |
Definitions |
|
A.1 |
Instrument Integration |
A functionality that creates, edits and deletes a sensor. |
|
A.2 |
Instrument Configuration |
A functionality that sets-up a sensor or a sensor network. |
|
A.3 |
Instrument Calibration |
A functionality that controls and records the process of aligning or testing a sensor against dependable standards or specified verification processes. |
|
A.4 |
Instrument Access |
A functionality that reads and/or updates the state of a sensor. |
|
A.5 |
Configuration Logging |
A functionality that collects configuration information or (run-time) messages from a sensor (or a sensor network) and output into log files or specified media which can be used by routine troubleshooting and in incident handling. |
|
A.6 |
Instrument Monitoring |
A functionality that checks the state of a sensor or a sensor network which can be done periodically or when triggered by events. |
|
A.7 |
(Parameter) Visualisation |
A functionality that outputs the values of parameters and measured variables a display device. |
|
A.8 |
(Real-Time) (Parameter/Data) Visualisation |
A specialisation of (Parameter) Visualisation which is subject to a real-time constraint. |
|
A.9 |
Process Control |
An interface that provide operations to receive input status, apply a set of logic statements or control algorithms, and generate a set of analogy and digital outputs to change the logic states of devices. |
|
A.10 |
Data Collection |
An interface that provides operations to obtain digital values from a sensor instrument, associating consistent timestamps and necessary metadata. |
|
A.11 |
(Real-Time) Data Collection |
A specialisation of Data Collection which is subject to a real-time constraint. |
|
A.12 |
Data Sampling |
An interface that provides operations to select a subset of individuals from within a statistical population to estimate characteristics of the whole population. |
|
A.13 |
Noise Reduction |
An interface that provides operations to remove noise from scientific data. |
|
A.14 |
Data Transmission |
A interface that provides operations to transfer data over communication channel using specified network protocols. |
|
A.15 |
(Real-Time) Data Transmission |
A specialisation of Data Transmission which handles data streams using specified real-time transport protocols. |
|
A.16 |
Data Transmission Monitoring |
An interface that provides operations to check and report the status of data transferring process against specified performance criteria. |
|
B |
Data Curation Subsystem |
|
|
No |
Functions |
Definitions |
|
B.1 |
Data Quality Checking |
An interface that provides operations to detect and correct (or remove) corrupt, inconsistent or inaccurate records from data sets. |
|
B.2 |
Data Quality Verification |
An interface that provides operations to support manual quality checking. |
|
B.3 |
Data Identification |
An interface that provides operations to assign (global) unique identifiers to data contents. |
|
B.4 |
Data Cataloguing |
An interface that provides operations to associate a data object with one or more metadata objects which contain data descriptions. |
|
B.5 |
Data Product Generation |
An interface that provides operations to process data against requirement specifications and standardised formats and descriptions. |
|
B.6 |
Data Versioning |
A interface that provides operations to assign a new version to each state change of data, allow to add and update some metadata descriptions for each version, and allow to select, access or delete a version of data. |
|
B.7 |
Workflow Enactment |
An interface that provide operations or services to interprets predefined process descriptions and control the instantiation of processes and sequencing of activities, adding work items to the work lists and invoking application tools as necessary. |
|
B.8 |
Data Storage & Preservation |
An interface that provides operations to deposit (over long-term) the data and metadata or other supplementary data and methods according to specified policies, and make them accessible on request. |
|
B.9 |
Data Replication |
An interface that provides operation to create, delete and maintain the consistency of copies of a data set on multiple storage devices. |
|
B.10 |
Replica Synchronisation |
An interface that provides operations to export a packet of data from on replica, transport it to one or more other replicas and to import and apply the changes in the packet to an existing replica. |
|
C |
Data Access Subsystem |
|
|
No |
Functions |
Definitions |
|
C.1 |
Access Control |
An interface that provides operations to approve or disapprove of access requests based on specified access policies. |
|
C.2 |
Resources Annotation |
An interface that provides operations to create, change or delete a note that reading any form of text, and to associate them with a computational object. |
|
C.3 |
(Data) Annotation |
A specialisation of Resource Annotation which allows to associate an annotation to a data object. |
|
C.4 |
Metadata Harvesting |
An interface that provides operations to (regularly) collect metadata (in agreed formats) from different sources. |
|
C.5 |
Resource Registration |
An interface that provides operations to create an entry in a resource registry and insert resource object or a reference to a resource object in specified representations and semantics. |
|
C.6 |
(Metadata) Registration |
A specialisation of Resource Registration, which registers a metadata object in a metadata registry. |
|
C.7 |
(Identifier) Registration |
A specialisation of Resource Registration, which registers an identifier object in an identifier registry. |
|
C.8 |
(Sensor) Registration |
A specialisation of Resource Registration which registers a sensor object to a sensor registry. |
|
C.9 |
Data Conversion |
An interface that provides operations to convert data from one format to another format. |
|
C.10 |
Data Compression |
An interface that provides operations to encode information using reduced bits by identifying and eliminating statistical redundancy. |
|
C.11 |
Data Publication |
An interface that provides operations to provide clean, well-annotated, anonymity-preserving datasets in a suitable format, and by following specified data-publication and sharing policies to make the datasets publically accessible or to those who agree to certain conditions of use, and to individuals who meet certain professional criteria. |
|
C.12 |
Data Citation |
An interface that provides operations to assign an accurate, consistent and standardised reference to a data object, which can be cited in scientific publications. |
|
C.13 |
Semantic Harmonisation |
An interface that provides operations to unify similar data (knowledge) models based on the consensus of collaborative domain experts to achieve better data (knowledge) reuse and semantic interoperability. |
|
C.14 |
Data Discovery and Access |
An interface that provides operations to retrieve requested data from a data resource by using suitable search technology. |
|
C.15 |
Data Visualisation |
An interface that provides operations to display visual representations of data. |
|
D |
Data Processing Subsystem |
|
|
No |
Functions |
Definitions |
|
D.1 |
Data Assimilation |
An interface that provides operations to combine observational data with output from a numerical model to produce an optimal estimate of the evolving state of the system. |
|
D.2 |
Data Analysis |
An interface that provides operations to inspect, clean, transform data, and to provide data models with the goal of highlighting useful information, suggesting conclusions, and supporting decision making. |
|
D.3 |
Data Mining |
An interface that provides operations to support the discovery of patterns in large data sets. |
|
D.4 |
Data Extraction |
A interface that provides operations to retrieve data out of (unstructured) data sources, including web pages ,emails, documents, PDFs, scanned text, mainframe reports, and spool files. |
|
D.5 |
Scientific Modelling and Simulation |
An interface that provides operations to support of the generation of abstract, conceptual, graphical or mathematical models, and to run an instance of the model. |
|
D.6 |
(Scientific) Workflow Enactment |
A specialisation of Workflow Enactment, which support of composition and execution a series of computational or data manipulation steps, or a workflow, in a scientific application. Important processes should be recorded for provenance purposes. |
|
D.7 |
(Scientific) Visualisation |
An interface that provides operations to graphically illustrate scientific data to enable scientists to understand, illustrate and gain insight from their data. |
|
D.8 |
Service Naming |
An interface that provides operations to encapsulate the implemented name policy for service instances in a service network. |
|
D.9 |
Data Processing |
An interface that provides operations to initiate the calculation and manage the outputs to be returned to the client. |
|
D.10 |
Data Processing Monitoring |
An interface that provides operations to check the states of a running service instance. |
|
E |
Community Support Subsystem |
|
|
No |
Functions |
Definitions |
|
E.1 |
Authentication |
An interface that provides operations to verify a credential of a user. |
|
E.2 |
Authorisation |
An interface that provides operations to specify access rights to resources. |
|
E.3 |
Accounting |
An interface that provides operation to measure the resources a user consumes during access for the purpose of capacity and trend analysis, and cost allocation. |
|
E.4 |
(User) Registration |
A specialisation of Resource Registration which registers a user to a user registry. |
|
E.5 |
Instant Messaging |
An interface that provides operation for quick transmission of text-based messages from sender to receiver. |
|
E.6 |
(Interactive) Visualisation |
An interface that provides operations to enable users to control of some aspect of the visual representations of information. |
|
E.7 |
Event Notification |
An interface that provide operations to deliver message triggered by predefined events. |
[1] ESFRI, the European Strategy Forum on Research Infrastructures, is a strategic instrument to develop the scientific integration of Europe and to strengthen its international outreach.
[2] ICOS, http://www.icos-infrastructure . eu/ , is a European distributed infrastructure dedicated to the monitoring of greenhouse gases (GHG) through its atmospheric, ecosystem and ocean networks.
[3] EURO-Argo, http://www. e uro-argo.eu/ , is the European contribution to Argo, which is a global ocean observing system.
[4] EISCAT-3D, http://www.eiscat3d.se/ , is a European new-generation incoherent-scatter research radar for upper atmospheric science.
[5] LifeWatch, http://www.lifewatch.com/ , is an e-science Infrastructure for biodiversity and ecosystem research.
[6] EPOS, http://www.epos-eu.org/ , is a European Research Infrastructure on earthquakes, volcanoes, surface dynamics and tectonics.
[7] EMSO, http://www.emso-eu.org/management/ , is a European network of seafloor observatories for the long-term monitoring of environmental processes related to ecosystems, climate change and geo-hazards.
Stichting EGI에게 부여된 무료 Atlassian Confluence Community License로 실행됩니다. 오늘 Confluence를 평가해 보세요.