Data federation addresses the problem of uniformly accessing multiple, possibly heterogeneous data sources, by mapping them into a unified schema, such as an RDF(S)/OWL ontology or a relational schema, and by supporting the execution of queries, like SPARQL or SQL queries, over that unified schema. Data explosion in volume and variety has made data federation increasingly popular in many application domains. Hence, many data federation systems have been developed in industry and academia, and it has become challenging for users to select suitable systems to achieve their objectives. In order to systematically analyze and compare these systems, we propose an evaluation framework comprising four dimensions: (i) federation capabilities, i.e., query language, data source, and federation techniques; (ii) data security, i.e., authentication, authorization, auditing, encryption, and data masking; (iii) interface, i.e., graphical interface, command line interface, and application programming interface; and (iv) development, i.e., main development language, deployment, commercial support, open source, and release. Using this framework, we thoroughly studied 51 data federation systems from the Semantic Web and Database communities. This paper shares the results of our investigation and aims to provide reference material and insights for users, developers and researchers selecting or further developing data federation systems.
The convenience of digitization, the variety of data descriptions, and the discrepancy in personal preferences have led large enterprises to store massive amounts of data in a variety of formats, ranging from structured relational databases to unstructured flat files. According to the prediction by Reinsel et al. [1], the global data volume will reach 163 zettabytes by 2025, and half of that data will be produced by enterprises.
Since data becomes more valuable if enriched and fused with other data, decision-makers need to consider data distributed in different places and with different formats in order to get valuable insights that support them in their daily activities. However, data explosion in volume, variety, and velocity – i.e., the “3Vs” of Big Data [2,3] – increases complexity and makes the traditional ways of data integration [4–6], such as data warehousing [7,8], not only more costly in terms of time and money but also unable to guarantee the freshness of data. Integration solutions developed in a more agile way are thus demanded especially in the Big Data context. Data federation is a technology that makes this possible today, that is becoming more and more appealing in both industry and academia, and that has been studied for a long time in different communities such as the Database and (more recently) the Semantic Web ones.
Data federation systems (also known as federated database systems) are traditionally defined as a type of meta-database management system that transparently maps multiple autonomous database systems into a single federated database [9,10]. The key task of data federation systems is federated query answering, that is to provide users with the ability of querying multiple data sources under a uniform interface. Such an interface usually consists of a query language over a unified schema, such as SQL [11] over a relational schema or SPARQL [12] over an RDF(S)/OWL [13–15] ontology, this interface being often closely related or restricting the query languages and schemas of supported data sources. Unlike in traditional pipelines for data extraction, transformation, and loading (ETL) often used in data warehouse systems, federated query answering is achieved by data virtualization [16,17], i.e., all the data are kept in situ and accessed via a common semantic layer on the fly, with no data copy, movement, or transformation. As a result, federated query answering via data virtualization reduces the risk of data errors caused by data migration and translation, decreases the costs (e.g., time) of data preparation, and guarantees the freshness of data. Compared to centralized solutions, though, accessing multiple data sources on the fly renders query answering more challenging [18–20] and requires sophisticated optimization strategies to be devised. Besides federated query answering, modern data federation systems also offer a wide range of other important capabilities for data management, such as read-and-write data access for enabling users to both access and modify the data in the sources, data security for protecting the sensitive data of users and implementing secure data access, and data governance for managing the availability, usability, and integrity of the data.
Data federation is an active field and many data federation systems have been and are being developed. For example, FedX [21,22] and Teiid [23] are two systems supporting respectively SPARQL query answering over multiple SPARQL endpoints (i.e., standardized HTTP services [24] that can process SPARQL queries) and SQL query answering over multiple heterogeneous data sources, like relational databases, structured files and web services. More generally, current data federation systems include both industrial systems, mostly developed by software companies and more mature, and academic systems, mostly developed by research organizations and providing newer functionalities. Moreover, federated query answering facilities are often included in modern data management systems aimed at heterogeneous big data. These systems include logical data warehouses [25–27], data lakes [28–31], and polystores [32–36], and can be seen to all intents and purposes as special cases of data federation systems. All the aforementioned systems present substantial overlap in terms of adopted techniques and extra capabilities offered to users, while differences in the exposed unified interface may be often bridged – e.g., by using Ontology-Based Data Access (OBDA) [37] to adapt SQL over a federated relational schema to SPARQL over an OWL ontology – this way enabling the use of a data federation system in additional scenarios with respect to the ones it was primarily developed for – e.g., use a robust industrial SQL-based data federation system to create a “virtual” knowledge graph for Linked Open Data publishing. Therefore nowadays, users have access to a large number of data federation systems to choose among, but selecting the right system for a specific task requires collecting, analyzing, and comparing the capabilities and techniques of many systems, which is very time-consuming: for industrial systems, the information needed is usually fragmented and scattered, and the official documents often consist of hundreds of pages; for academic systems, conversely, end-user documentation is typically poor or unavailable, and system features are described in academic publications, when available.
This survey tries to shed some light on this complex matter by analyzing 51 state-of-the-art data federation systems, jointly covering systems from the Semantic Web and the Database communities thanks to their substantial interchangeability and their commonalities in implemented techniques and features. The considered systems, selected by following a rigorous and well-founded methodology, comprise 33 industrial systems under active development and with public official documentation, and 18 academic systems. This work has a twofold goal: help end users in identifying the systems best suited to their applications and tasks, and allow researchers and developers to gain more insights into the capabilities, techniques, strengths, and weaknesses of current systems, this way informing further work in the field.
In order to compare the considered systems from the perspective of data federation in a uniform way, this survey proposes a qualitative evaluation framework consisting of four dimensions further refined into several sub-dimensions, which we defined by considering and classifying the aspects that play crucial roles in the users’ choice of a system for employment in their applications and tasks:
The federation capabilities dimension concerns the federated query answering features offered by a system over multiple data sources, both homogeneous and heterogeneous in type. It is further refined into three closely related sub-dimensions: data source, query language, and federation techniques.
The data security dimension concerns the capabilities of a system of safeguarding the data in the sources participating in the federation from unwanted actions by unauthorized users, especially when such data is sensitive or private. It is refined into five sub-dimensions: authentication, authorization, auditing, encryption, and data masking.
The interface dimension concerns the usability of the systems. It is further divided into the three sub-dimensions of graphical interface, command line interface, and application programming interface, so as to measure the ability of supporting users in fully appreciating, accessing, and exploiting the features implemented by a system.
The development dimension, finally, concerns the development, release and support practices adopted by system vendors. Its five sub-dimensions of main development language, deployment, commercial support, open source, and release, aim overall at assessing the maturity of the systems and the possibilities for users to get help from vendors, and to maintain and improve the systems by themselves, if needed.
For all the 51 considered data federation systems, we collect information along the proposed four dimensions by consulting the official documentation of each system, as well as its related publications. Note that since not all the features of these systems are properly documented, our analysis is conducted using our best efforts.
This survey adds to an existing body of literature [20,38–43] that reviews the approaches and systems for federated query answering under multiple perspectives. For example, Oguz et al. [20] evaluate seven SPARQL federation query engines by focusing on their query evaluation techniques, while Azevedo et al. [42] study the modern data federation systems (including BigDAWG [33], CloudMdsQL [35], Myria [34], and Apache Drill [44]) by focusing on their features, owners, goals, and main components. Compared with all these works and summing up, we make the following contributions:
We carried out an extensive review of academic literature and documentation about industrial solutions to identify a large number of data federation systems from the Semantic Web and the Database communities.
We provide a framework for investigating data federation systems in a uniform and qualitative way by taking into account aspects of interest for data federation end users, developers and researchers.
We analyze the identified systems through the proposed framework, this work amounting to an extensive analysis covering 51 systems and 4 main evaluation dimensions overall divided into 16 sub-dimensions. To the best of our knowledge, this is the most extensive analysis on data federation so far in terms of investigated systems and considered aspects.
As a by-product of our analysis, we make explicit the common capabilities of current data federation systems, such as the capability of handling heterogeneous data sources, or the query optimization techniques used.
We discuss remaining open problems and challenges and point out the research directions that are interesting and valuable for pursuit.
The remainder of the survey is organized as follows. Section 2 presents an outline of data federation. Section 3 illustrates the overall methodology of the survey work. Section 4 describes the proposed framework for systems assessment and comparison. Section 5 lists and provides a summary of the selected systems. Section 6 thoroughly analyzes the capabilities of these systems according to the proposed framework. Section 7 discusses related work. Section 8 concludes by discussing open problems and challenges as well as giving directions for further work. Appendices A and B respectively provide further details on the specific sources supported by the systems and on our methodology. A Web version of the tables in this paper, including possible corrections and integrations, is available on GitHub.1
This section provides an overview of the main concepts underlying data federation that are addressed in this paper, for readers not already familiar with them.
The core task of data federation is federated query answering [20,38–41]. For a set of autonomous and possibly heterogeneous data sources, the goal of federated query answering is to provide a uniform interface, typically as a unified query language over a unified schema, to access the data of these sources in situ, i.e., without first copying the data to centralized storage. Given a user query over the unified schema, this task is carried out by issuing and orchestrating the evaluation of native sub-queries targeting the data sources of the federation.
Figure 1 depicts the typical architecture of a federated query engine providing federated query answering. Unified schema, mappings, metadata catalog are key components, which respectively provide a unified schema of the data sources participating in the federation, map the data in the sources to the unified schema, and provide statistical information about the data sources as well as the information of how these data sources can be accessed. For example, for a relational database, if the unified schema is an RDF ontology, then there exist mappings that map the tables of this database to the classes and properties of the ontology, and the metadata catalog could list the relevant content statistics, such as the number of rows of the referred tables, used in federated query optimization. Formally, a data federation instance usually consists of three components , where is a set of data sources which can be relational databases, NoSQL databases, structured files, data warehouses, and so on; is the unified schema for these n sources, such as an RDF(S) ontology or relational schema; and is a set of mappings that map the data of the sources participating in the federation into the elements conforming to the unified schema . Then accessing multiple data sources staying in situ simultaneously is carried out by evaluating queries Q expressed in terms of the unified schema (such as SPARQL queries when is an RDF ontology, and SQL queries when is a relational schema) via the following steps:
Query parsing. This step deals with the syntactic issues of Q, i.e., checking whether the input queries are syntactically correct w.r.t. the adopted query language(s) as well as the unified schema. Some engines also transform Q into an algebraic form, such as a tree structure using internal nodes to denote operations (e.g., join, union, or projection) and leaf nodes to denote accessed relations.
Typical architecture of a federated query engine (inspired by Oguz et al. [20]).
Source selection and query partition. This step selects suitable data sources for each algebraic component of Q, and partitions Q into smaller sub-queries (i.e., query chunks) accordingly, based on the mappings from the data sources to the unified schema . Approaches for source selection can be index-based, such as the “triple pattern-wise source selection” for SPARQL queries [45,46], and a way for query partitioning is to try to “push down” the evaluation of the operators to the data sources, rather than perform such evaluation at the level of the federation engine.
Query optimization & query plan generation. This step computes an execution plan of the partitioned sub-queries , establishing in which order to evaluate the sub-queries and which algorithms to use for joining their answers (e.g., bind join, hash join, etc), based on the metadata catalog. Existing approaches may be rule-based (i.e., via predefined and deterministic heuristic rules) or cost-based (i.e., choose the lowest-cost execution plan according to some heuristic cost function).
Query plan execution. This step, finally, evaluates the decomposed sub-queries over the corresponding data sources via the mappings and the metadata catalog, and generates the answers of the original query Q. Note that, if the query language that the data source supports is different from the query language of the federation engine, a translation based on the mappings is needed to translate the sub-query into the one supported by the data source.
An example of federated query answering.
Next, we use an example to further clarify the inner workings of federated query answering.
Suppose we have a data federation instance modeling information about a large enterprise, as per the one in Fig. 2. Here and are two data sources storing information about two different departments. Concretely, is a relational database from the Sales department storing the information about products being sold, whereas is a NoSQL database from the Human Resources department storing information about each employee of the enterprise. The unified schema of the federation instance is an RDF ontology including the classes :Product, and :Inspector, as well as the properties :hasCode, :hasInspector, :hasName, and :hasSalary. The set contains mappings from the data in to the terminology :Product, :hasCode, and :hasInspector of , as well as the mappings from the data in to the terminology :Inspector, :hasName, and :hasSalary.
Suppose we want to retrieve the names of inspected products as well as the names and salary of their relative inspectors. For this purpose, we formulate a SPARQL query such as Q from Fig. 2, consisting of five triple patterns . We send Q to the federation engine for evaluation over the data federation instance. As the first step, the engine checks the syntax of Q w.r.t. the syntax of SPARQL and the classes and properties declared in . After the syntactic check, the engine identifies the sources of each triple pattern in Q, and further partitions Q into sub-queries according to some query partition strategy. In our example, by exploiting the mapping set , the federation engine selects source for triple patterns , , and , and selects source for triple patterns and . Then, by adopting exclusive groups, i.e., a push down strategy for query partition and optimization [21,22], the engine computes a partition of Q, by grouping together the triple patterns corresponding to the same source, so that joins among them are pushed down to the source and a minimal number of federated joins are evaluated. After that, the engine computes a plan for evaluating . A possible plan is the following: reformulate query into a SQL query and query into a NoSQL query , according to the mappings definitions in ; dispatch to and to , and evaluate them in a parallel way; merge the returned answers for and to generate the answers of the initial query Q.
Transparent vs explicit federated query answering
Example of SPARQL query under the explicit federation setting (left-hand side), and its counterpart under the transparent setting (right-hand side).
From the perspective of whether the data source information is transparent for end users, federated query answering can be classified into transparent federation (the one we have discussed so far) and explicit federation [9,45]. Transparent federation gives users the impression to query one single data source despite data being distributed and possibly coming from heterogeneous sources [45]. Hence, it is recognized as a general and ideal2
A simplified setting is one where the unified schema is simply a merge of the source schemas, and the user explicitly states in the query the sources against which it should be evaluated. In such a scenario, we talk about explicit data federation. This approach is built-in into SPARQL 1.1 through its dedicated SERVICE keyword, and therefore is supported by any SPARQL-based system fully compliant with SPARQL 1.1, including systems not primarily focusing on data federation. Figure 3, left-hand side, shows an example of query formulated under the explicit federation setting, asking for the data from a local RDF store and an explicitly specified remote RDF store. The right-hand side of the same figure shows the same query formulated under the transparent federation setting, assuming that foaf:knows and foaf:name are properties belonging to the unified schema.
Compared with transparent federated query answering, the explicit scenario does not require a procedure of source selection for delivering its task of accessing and joining multiple data sources. However, the burden is placed on end users, and this might constitute a major hindrance in case they are not familiar with the data sources participating in the federation and the data therein contained.
However, the transparent setting is not devoid of drawbacks. For instance, users lose the ability of communicating with specific data sources directly. Moreover, the transparent situation needs to maintain a unified schema mapped to multiple data sources, which means that it is more sensitive to schema updates: when the schema of a source is updated, the unified schema and the mappings may also need to be updated.
Other capabilities
As mentioned earlier, beyond the core feature of federated query answering, data federation has evolved to offer a wide range of additional capabilities supporting more powerful and intelligent forms of data consumption and management. Next, we list some noteworthy capabilities supported by federation systems of this survey.
Data security. It provides techniques for protecting users’ privacy and sensitive data from leakage. Take the data federation platform Denodo as an example. The “unified security management” of Denodo offers a single point to control the access to any piece of information. Different users of Denodo are only allowed to access either filtered or masked data by using the Denodo role-based security model. Interested readers can refer to the official documents3
Data update. It provides the capability of enabling users to both read and write the data of the sources participating in the federation. For example, the SPARQL federation engine FedX4
so as to make users able to modify the data of the SPARQL endpoints, and the SQL federation engine Denodo supports SQL data manipulation language (SQL DML) with the motivation of making users able to modify the data stored in the source databases;
Data quality. It provides the techniques for guaranteeing the correctness and consistency of data. Take the SAS Federation Server6
as an example. Data quality on SAS Federation Server is implemented through a “SAS Quality Knowledge Base (QKB)”, allowing for the specification of a set of methods and rules for data quality, such as rules to cleanse the data.
Survey methodology
This survey work stems from our needs for selecting suitable data federation systems for heterogeneous data integration. Collecting, analyzing, and comparing the existing systems on data federation is a very time-consuming process. Sharing the results of our study can benefit readers interested in data federation solutions, such as end-users (consumers), developers, researchers and students. In this section, we present the overall methodology used for our study. Figure 4 provides a snapshot of our methodology, which consists of the identification of the considered systems, the design of the system evaluation framework, and the evaluation of the systems through the framework.
The overall methodology of the survey work.
The methodology for system selection
As shown in Fig. 4, the systems considered in our survey are mainly identified through a four stage process: designing keywords and questions, searching in the search engines, finding the candidate data federation systems, and filtering according to the inclusion criteria. The bulk of candidate systems collection and filtering required three months, between the end of 2020 and the beginning of 2021. Although sharing the same stages, the criteria for selecting academic systems and industrial ones are a bit different. For clarity, in the following, we describe the selection of academic systems and industrial systems separately.
The selection of academic systems The considered academic systems are selected by reviewing the academic publications found via surfing the Google Scholar search engine. As a first step, we designed the following keywords to find the potential systems:
Note that for obtaining “more” results, we did not use any operator, like “AND” and “OR”, in the search phrases. After searching these keywords, we speed-read the titles and abstracts of more than 2000 academic publications from libraries such as SpringerLink, IEEE Xplore, ACM Digital Library, and so on. By evaluating these titles and abstracts, we selected and downloaded 295 academic publications for further in-depth reading, consisting of papers, technical reports, PhD and master theses whose primary focus is on data federation. They include a majority of system-specific publications and 17 system comparison publications, which range from data federation surveys to benchmarks, system evaluations and PhD theses reviewing the topic.
After reviewing these publications, we identified a total of 56 academic data federation systems that we narrowed down to a final selection of 18 representative systems based on the following inclusion criteria:
Scope. The system must focus on the problem of query federation, or introduce a data federation system.
Venue. The system must be described in formal publications such as papers in journals or conference proceedings, and not only in preprints or technical reports.
Availability. The system source code and official website must be available, either linked from the system publications or findable from the authors’ GitHub profiles (e.g., the SPLENDID system).
Relevance. The system must satisfy at least one of the following criteria: it should be mentioned in system comparison publications; it should provide federation of heterogeneous data sources (e.g., RDBs and CSV files); it should ensure data security.
Citations. For period ⩾2020, we do not consider citations. For period 2015–2019, systems should have at least 10 citations. For each prior period ⩽2008, 2009–2011 and 2012–2014, we only consider the system having the largest number of citations among the ones matching the previous criteria.
The citations criterion aims at keeping the scope of this survey manageable and focused on newer systems, also considering that most systems earlier than 2015 are covered by other system comparison publications and their source code is more likely to be unavailable, making them less interesting to our intended audience. Note that to apply this criterion, a system is classified into the year of its most recent conference or journal publication and a system number of citations is obtained by summing the Google Scholar citations of all its collected publications as of 2022/06/07.
Fine details of the selection process are provided in Appendix B, which reports on: (i) the collected 295 academic publications in terms of aggregated statistics (Section B.1) and full bibliography (Section B.4); (ii) the collected 17 system comparison publications, in terms of metadata, compared systems and considered aspects (Section B.2); and (iii) the selection of 18 systems out of the 56 identified ones, based on attained inclusion criteria (Section B.3).
The selection of industrial systems To find candidate industrial systems we adopted the Google search engine. We employed the following generic keywords/questions, aiming at including as many systems as possible:
where X denotes a data federation system already known by us, like Teiid and Denodo. We collected, deduplicated and reviewed the search results, looking for the websites of industrial data federation systems. Some search results already corresponded to a system website. Others were instead discussing more broadly about data federation/virtualization or recommending/listing/comparing systems referring to data federation, virtualization or integration, in which case we browsed page links to identify any referenced system website. As a result, we collected the official websites of 72 candidate systems that may provide (due to noise in search results following the use of generic keywords) the capability of data federation. We then consulted these websites, read the systems descriptions and documentation carefully, and eventually selected 33 industrial systems for our survey work that strictly meet all the following inclusion criteria:
Scope. The system must actually provide the capability of data/query federation.
Community. There should be evidence for a user community around the system, e.g., via usage statistics and user messages in fora, mailing lists, issue trackers and the like.
Documentation. Official system documentation must be publicly available, to support both (perspective) users and ourselves in conducting the analyses reported in this survey.
Active development. There must have been at least a system release since 2015/10, i.e., in the last five year since the time we started this survey (2020/10).
The concrete information of the systems found, as their names, owners, and descriptions, can be found in Section 5.
The methodology for designing the evaluation framework
The generation of the system evaluation framework.
To design a framework for evaluating data federation systems in a uniform and qualitative way, also considering the intended audience of this survey, we focus on answering the following question (see framework design in Fig. 4):
What aspects of data federation systems are relevant for end users, developers and scholars?
While in principle answers can be obtained by interviewing these three groups (e.g., via questionnaires), this approach presents two main difficulties: (i) it is hard to identify a representative sample to interview; and (ii) it is hard for interviewees to answer the question in a general and comprehensive way. Instead, we rely on the fact that data federation is an established domain that has been studied for decades in both the Database and the Semantic Web communities, leading to a large body of information from which to extract the aspects of interest that answer our question. Concretely, we consider three information sources:
Academic publications. We look for aspects deemed important by other surveys on data federation, or that are frequent in academic publications referring to data federation systems.
Official documents. We look for aspects commonly present in official documents of data federation systems, such as user and developer guides.
Web pages. We look for aspects that are often considered when comparing data federation systems.7
The system evaluation framework, consisting of four dimensions with sub-dimensions, is generated by combining, classifying, and further refining the identified aspects. The full process is depicted in Fig. 5. Starting from our original question (Requirement box), we report the “raw” aspects identified in academic publications, official documents and web pages (Identification box). They were classified into four categories (Classification box), which then underwent a series of refining steps (Refinement box), guided both by the information sources we reviewed and by our own expertise as researchers and developers, as well as our own experience on the data federation task and systems. This refinement results in a final evaluation framework that addressess the information needs of the different audience groups targeted by this survey:
End users. They have the concrete need of integrating and federating data sources, and might lack technical skills like programming. Hence, aspects relevant to them are whether the system is able of handling their data sources, whether it provides a query language that they are familiar with, whether it offers a graphical interface to help them to set up a data federation instance easily, whether it provides the services for solving the problems they may encounter, whether it provides the techniques for protecting their data from leakage, and whether it is robust enough so as to withstand the technical difficulties that may be encountered in production (e.g., load spikes, temporary source unavailability, etc).
Developers. Their need is to work with the systems at a lower-level than end users, for instance through programming interfaces, so as to enrich the functionalities delivered by their own applications. Other developers might also be interested in the source code of the systems themselves, for the purpose of extending it with new functionalities, e.g., to support more complex data consumption scenarios.
Researchers and students. They conduct research or studies on data/query federation. Thus, the aspects of interest for them relate to the knowledge of the capabilities of the systems, or of the strategies they adopt.
All of the aforementioned aspects of interest are captured by the dimensions and sub-dimensions of our evaluation framework, as it will be detailed in Section 4.
The methodology of system evaluation
After identifying the considered systems and the evaluation framework, we use such framework to investigate and analyze the capabilities, strengths, and weakness of the considered systems, e.g., the capability of handling data heterogeneity. Finally, we point out some open problems and challenges that might be addressed by further research.
The framework for system evaluation and comparison
In this section, we present our framework for analyzing and comparing the selected systems under a user and application perspective in a uniform and qualitative way. Our framework, shown in the right part of Fig. 5, consists of four dimensions: federation capabilities, data security, interface, and development. Each dimension is further characterized by sub-dimensions (16 in total). In the remainder of this section we discuss each of these dimensions, and relative sub-dimensions, in detail.
Federation capabilities dimension
This dimension evaluates the main task of data federation systems, i.e., federated query answering, in terms of data source, query language, and federation techniques.
Data source sub-dimension The types of supported data sources usually play a key role when choosing a data federation system. For example, if a company has massive CSV files that need to be virtually integrated with data stored in MySQL, then it will preferably take into consideration systems supporting CSV files and MySQL at the same time. This sub-dimension also permits users to distinguish whether a system focuses on heterogeneous or homogeneous data sources. Roughly speaking, the more different types of data sources a system supports, the more capable that system is in accessing heterogeneous data. By reviewing the data sources supported by the considered systems, we design six types of data sources, like relational and graph-based, to inspect this sub-dimension. The concrete information will be introduced in Section 6.1.
Query language sub-dimension We consider the query language(s) provided to users for accessing and managing the data in the federated sources. Generally speaking, a federation system should preferably adopt a standard query language that is familiar to most people, like SPARQL or SQL. In this way, users do not need to learn a new query language when using the system, and existing tools and resources for the adopted language can be reused. We considered the systems developed within the Semantic Web and Database communities, but not limited to these two kinds. Hence, we characterize this sub-dimension into SPARQL, SQL, and Other.
Federation technique sub-dimension We refer to the typical architecture for federated query answering described in Section 2, and assess the main techniques adopted by a system. We focus on the techniques for metadata catalog, unified schema and mappings, source selection and query partition, query optimization and plan generation, and query execution. The motivation is to help readers in forming a general idea about the techniques employed by each system.
Data security dimension
As a data-centric application, data federation offers a single logical point to integrate data from multiple sources that may contain sensitive and private data (e.g., financial transactions, users’ contact information, or medical procedures). The protection of such data represents a crucial problem for obtaining the trust of users and data providers. This problem is further complicated by the risk of leaking sensitive information through analysis and correlation of otherwise non-sensitive data from separate sources. Therefore, the data security dimension considers whether a data federation system has the ability of safeguarding data from unwanted actions of unauthorized users, and it is further organized in sub-dimensions according to the system’s support for the most common data security mechanisms.
Authentication sub-dimension Authentication refers specifically to accurately identifying users before they have access to data. It is the act of validating that users are whom they claim to be, and is the first step in any data security process. The most common authentication mechanism is a username and password combination. Other common authentication mechanisms use shared keys, PIN numbers, or security tokens.
Authorization sub-dimension Authorization is a mechanism for granting or denying access to a resource based on identity. More generally, it consists in defining an access policy, and is usually implemented through a set of declarative security roles which can be associated to users. Authorization is different from authentication, and usually happens after authentication.
Auditing sub-dimension Data auditing logs and reports on events like users’ accesses, modifications, changes of ownership, or permissions regarding sensitive data. Audit procedures increase visibility on data operations and are instrumental to the investigation and prevention of data breaches and other data security incidents.
Encryption sub-dimension Data encryption algorithms transform the original data into an unreadable format so that only authorized users having the corresponding key can decrypt and read the information. Encryption is commonly employed on data transiting between the system and the user, and possibly on data stored, cached, or otherwise materialized within the system as well, to protect them from unauthorized low level accesses.
is the process of masking (i.e., obscuring, deleting, or otherwise scrambling) specific pieces of accessed data, so as to ensure that sensitive information is not exposed to unauthorized parties (e.g., users, developers, system administrators). Data masking may use lossless techniques such as encryption or tokenization9
that allow retrieving the original unmasked value if the required information is available (e.g., the decryption key), but this feature is not a requirement and is not provided by many masking approaches that just aim at hiding sensitive data (e.g., the simple replacement of data with random values, or with ‘∗’ characters). Also, differently from encryption that may operate on the whole communication channel between the system and the user, data masking typically operate on the individual pieces of sensitive data (e.g., a table column or row field).
Interface dimension
The ultimate goal of system development is to support users in fully appreciating, accessing, and exploiting the features implemented by the system. Its achievement largely depends on the interface(s) offered to users for interacting with the system, which ultimately determine the ease of use, i.e., the usability, of a system. Such interfaces are the subject of this dimension, whose sub-dimensions are organized according to the different types of interfaces commonly offered by systems.
Graphical interface sub-dimension Setting up a data federation system is typically a complex task involving an extensive amount of configuration, e.g., for connecting the federated data sources, acquiring their necessary metadata, and setting up the system components. For example, Teiid supports the use of a complex XML configuration file10
to define a federated database, there called a Virtual Database (VDB). Without fully understanding the syntax and components of this file, building a VDB is hard for users, especially for the less-technical ones. A graphical user interface may greatly ease the configuration process, as well as other administration and operation tasks, and thus largely affects the user friendliness of a system.
Command line interface sub-dimension Data federation systems are typically used as components of larger information systems, where they need to be integrated with other components, such as business intelligence (BI), customized dashboards, or machine learning tools, to support or handle much more complex applications and tasks. To that respect, a command line interface provides a first, simple solution for automatically invoking the functionalities of a data federation system in other programs or scripts of a larger information system.
Application programming interface sub-dimension A further, more flexible integration mechanism is represented by application programming interfaces (APIs) offered by the data federation system, such as web APIs or client libraries in various programming languages (e.g., ODBC/JDBC drivers). Such APIs make it easier for developers to connect, configure, and operate an instance of the system at run-time within other applications.
Development dimension
This dimension considers the development, release, and support practices of a system, with its sub-dimensions capturing the aspects that are most relevant when matching the non-functional requirements of a user (in terms of, e.g., performance, robustness, flexibility, sustainability).
Main development language sub-dimension The main programming language(s) used to develop the core functionalities of a system influence system requirements (e.g., a Java Runtime Environment is required for the Java language), performance, customization, and integration options (e.g., embedding the system as a library), and consequently affect the system’s fitness for use in an intended user application.
Deployment sub-dimension The hardware/software infrastructure required to run a system, as well as its economic viability, are influenced by the deployment options offered for the system. At one end of the spectrum, we have on-premises deployment where the user obtains the software, possibly for a one-time license fee, and is in charge of its deployment, maintenance (e.g., updates) and configuration, which may occur on any machine(s) under the user control (i.e., “on the premises” of the user). The other end of the spectrum is represented by Software as a Service (SaaS),11
whereby the system is offered as a pre-deployed service maintained by the provider, and the user only cares about configuring and using the service on a subscription basis, where costs may depend on “how long” (e.g., hours) and “how much” (e.g., number of queries, transfered data) the service is used. In between, Infrastructure as a Service (IaaS) and Platform as a Service (PaaS) are intermediate options where system deployment and maintenance are up to the user (as for on-premises deployment), but the system comes bundled with infrastructural resources, such as virtual machines or middleware, of a cloud provided (e.g., Amazon AWS, Microsoft Azure, Google Cloud Platform), these resources managed to different extents by the user (IaaS) or the provider (PaaS). Examples are container platforms like Kubernetes or OpenShift, or cloud marketplaces where virtual machines pre-configured with the system are obtained and subscription fees are divided among system and infrastructure providers.
Commercial support sub-dimension Learning how to best use an unfamiliar and complex system and dealing with any issue preventing its normal operation are time-consuming activities, which may result in additional costs or even in economic losses due to system downtimes. Therefore, the availability of commercial support, e.g., in the form of training, timely bug fixes, and installation and customization services, plays a keys role when choosing a system.
Open source sub-dimension Systems whose source code is made freely available for modification and redistribution offer users more options for integrating the system while matching specific application requirements, for improving the system itself, and for maintaining the system even if it is no more supported by authors.
Release sub-dimension We consider the release history and practices of a system, focusing on the number of releases and the time between the first and the last release of the system. Generally speaking, the longer this time and the more numerous the releases, the more mature and robust the system typically is, since each new release is obtained by adding new functions or fixing some issues in the previous one. For example, the first release (v1.0) of the Denodo platform was in 2002 and the last here considered (v8.0) was in 2020. Thus, Denodo development has been active for almost 20 years, which makes it potentially more robust than some other younger systems.
Overview of the selected data federation systems
Before reporting on the application of the framework of Section 4, we provide here the list and a brief overview of the selected systems involved in our evaluation and comparison, also to help readers become more familiar with the current offer on data federation, both industrial and academic, as a whole. For the data federation systems developed in the context of the Semantic Web community, more academic ones and less industrial ones were found. On the contrary, for the systems developed within the context of the relational databases community, more industrial ones and less academic ones were identified.
Summary of the selected data federation systems. Academic systems in italics
Table 1 lists the selected systems alphabetically, reporting for each one its name with relevant references where to gather detailed information, academic (name in italics) or industrial nature, provider, and a one sentence description introducing the system (in its latest version) and complementing the detailed information reported in the next sections. Note that here and in the following, the information for industrial systems (33 in total) was mainly extracted from their official websites, while for the academic systems (18 in total), information was mostly extracted and summarized from their academic publications, although we also considered their online documentation if available.
On the whole, the table exhibits a substantial variability in terms of system provider, nature, and their main characteristics. Providers range from university and research institutions for academic systems, to open source organizations, specialized companies, and major corporations for industrial systems. Systems range from database engines (RDBMS, graph databases, triple stores, polystores, and other multi-model systems) whose storage services are augmented with data federation capabilities, to purely mediator systems specifically focusing on data federation, possibly complemented with accessory functionalities (e.g., security). Some industrial systems can be accessed only as cloud services (SaaS).
System evaluation and analysis
In this section, we investigate and analyze in more detail each of the systems overviewed in Section 5, while applying the four dimensions of the proposed framework. The main goal is to better understand the main characteristics of each system and to reveal its strengths and weaknesses with respect to the main task of data federation. Notice that all the systems we investigated have been considered as per their latest version (last update on November 20th, 2021).
Federation capabilities dimension
In this subsection, we evaluate the selected systems with a special attention to their capabilities to support federated query answering. In doing this, we will highlight the query languages that are supported, the data sources each system is able to manage, and the adopted federation techniques. Concerning the first two aspects, a synthetic overview of the query languages and the types of data sources supported by the investigated systems is presented in Table 2. The concrete data source implementations (e.g., MySQL) supported by each system are instead listed in Table 7 of Appendix A.
Evaluation of query language and data source sub-dimensions. Academic systems in italics. “–” denotes feature/information not found in the systems’ official documentation, websites, or academic publications, to the best of our efforts
Query language For columns 2–4 of Table 2, we can make the following observations:
With no significant distinction between industrial or academic systems, the standard and popular query languages SQL and SPARQL are adopted by most of these systems to query the data involved in the federation. This choice definitely eases the integration of the system with other possible interacting applications. Notice also that BigDAWG, CloudMds, Myria, and SAS Federation Server use alternative languages inspired by SQL to support the required capabilities in the distributed federation environment. Instead, Neo4j adopts the declarative graph language Cypher [74] as its underlying query language, with the motivation of making graph data querying easy to learn, understand, and use by the final users.
There exist very few systems that adopt multiple query languages at the same time. Among them, for instance, AllegroGraph supports SPARQL and Prolog simultaneously; GraphDB provides the capability of processing SPARQL, SQL, and Cypher queries; and Virtuoso takes both SPARQL and SQL as its query languages. This situation can be explained by taking into account that (i) the importance or necessity of supporting multiple query languages is unknown or ignored, and (ii) supporting multiple query languages within the very same system requires a lot of work from an engineering and development point of view.
Some of the academic SPARQL-based systems support only BGP-like queries, such as Obi-Wan [75] and Squerall [96]. Other systems support general SPARQL queries but their publications only discuss federation techniques tailored towards BGPs, such as CostFed [61], HiBISCuS [66], Ontario [28], PolyWeb [32], SAFE [89], SemaGrow [93] and SPLENDID [46]. General SPARQL support may be achieved by relying on a fully-fledged SPARQL engine like RDF4J12
https://rdf4j.org/
(formerly Sesame) that supports further operators such as UNION and OPTIONAL.
For systems supporting SPARQL federation, only a few systems, like Amazon Neptune and Apache Jena, provide the capability of explicit query federation via the SERVICE keyword. Among non-SPARQL systems, only CloudMdsQL does not support transparent federation.
Data source Uniformly evaluating and analyzing systems in terms of supported data sources is a challenging task for several reasons. Firstly, system providers usually adopt different standards and granularity to describe the data sources they support. Some systems classify supported data sources differently and possibly in incompatible ways. For example, relational sources all go under the databases class in Teiid,13
solely classifies the supported data sources into IBM data sources, third-party data sources, and files. Secondly, systems may list as supporting both a generic data access interface (e.g., JDBC, ODBC, ADO.NET, OLE DB, SPARQL HTTP protocol, etc) and some data sources available through that interface, with different meanings. Often, the listed sources are just examples or special cases for which additional capabilities are implemented, and additional sources may be configured (e.g., by tuning the employed SQL dialect) and connected through the interface. In some cases, however, the listed sources are simply the only ones supported through the interface, which we thus disregard in our assessment. Finally, sources not supported directly by a system, may be supported indirectly by combining the system with a suitable third-party adapter component, such as a SQL connector exposing a non-relational data source (e.g., MongoDB) through a standard relational interface (e.g., JDBC), as further discussed in Appendix A. Since such combinations are potentially limitless and the feasibility of each should be assessed (e.g., to verify whether combined components are actually compatible), we here consider only directly supported sources and further discuss the issue in Section 6.5. Overall, all the aforementioned factors make it difficult to assess the supported data source sub-dimension uniformly and precisely.
In order to understand the status quo of handling the variety dimension of big data in the data federation setting, after inspecting the data sources supported by each system, we take the following 6 types of sources into consideration: (i) Relational, including SQL-based RDBMS, (federated) relational query engines, and distributed/cloud relational stores; (ii) Graph-based, including SPARQL endpoints, RDF triple stores and property graphs; (iii) Aggregate-oriented, including key-value stores, wide-column stores, document stores and other NoSQL stores and search engines that organize data as “aggregates” [104], ranging from opaque values to arbitrarily complex nested documents;18
We use the broad “aggregate-oriented” category due to the difficulty of classifying many NoSQL stores into a single fine-grained category (e.g., Amazon DynamoDB is independently classified as key-value, wide-column, or document store by different academic and web sources).
(iv) Structured Files such as CSV, JSON and XML; (v) Web Service Paradigms to access arbitrary web sources, such as HTTP/REST and SOAP/WSDL (vs. specific web APIs like Twitter one); and (vi) Other. We manually classified each occurrence of a specific data source (e.g., MySQL, MongoDB) among the ones supported by a system, under one of the considered 6 data source types (e.g., relational and aggregate-oriented, respectively), depending on how the specific source is accessed by the system and also relying on established system classifications (e.g., DB-Engines [105] and Database of Databases [106] catalogs). We use “Other” as a container for all those infrequently supported sources not covered by the former 5 types, such as directory services, streaming and event data processing systems, specialized databases (e.g., for time series data) and protocols (e.g., IMAP), and various specialized web APIs. We remark that source classification is not global across systems but rather local to each data federation system supporting that source, so for instance a multi-model database like Virtuoso (when used as a source) may be classified as relational if accessed via SQL, or as graph-based if access occurs via SPARQL.
By combining Table 2 and Table 7, we can observe the following:
Industrial systems usually support more data sources than academic systems (respectively, 3.2 vs 1.9 distinct source types per system on average). Consider for example Data Virtuality, which covers all the source types we considered. It is an unsurprising conclusion, since industrial systems usually focus more on coverage.
As for the systems covering multiple, possibly heterogeneous, types of data sources, no matter whether industrial or academic, relational sources have been considered extensively, and most of the mainstream RDBMS implementations have been supported (cf. second column of Table 7). This may be caused by the dominant role of relational sources in organizing data. This dominant role, along with the generality and well-understood semantics of the relational model, might also partially explain the proliferation of SQL connectors/adapters for non-relational data sources (see discussion in Appendix A). Such proliferation facilitates, for a data federation system supporting the connector/adapter data access interface (e.g., JDBC), extending the support to additional, unanticipated data sources.
Structured files like JSON, XML, and CSV, because of their importance and wide use, are also directly supported as native data sources by many systems considered in this survey (24 out of 51, i.e., 47%). Other systems not directly supporting structured files may instead support the database systems commonly used for storing and indexing the kind of data of these files (e.g., MongoDB and Elasticsearch for JSON data).
Aggregate-oriented sources mostly consist of NoSQL systems (cf. the fourth column of Table 7), exhibit overall support (24 systems out of 51, i.e., 47%) similar to the one for graph-based sources and structured files, and are present both in industrial systems (18 out of 33, i.e., 55%) and, to a lesser degree, in academic systems (6 out of 18, i.e., 33%).
Web service paradigms, although important (many sources are available only as web services), are considered less often (10 systems out of 51, i.e., 20%). This may be caused by the difficulty of implementing federated query answering over such kind of data, as their data models (where defined) and access patterns (usually restricted) are very dissimilar from the ones exposed by the data federation system to its users.
Other sources in our classification consist mostly of specialized web APIs (cf. last column of Table 7) and are supported by industrial systems (18 out of 33, i.e., 55%) more than academic systems (2 out of 18, i.e., 11%).
Systems supporting SQL queries focus on relational sources (21 systems out of 22, i.e., 95%) while graph-based sources have rarely been taken into account (5 out of 22, i.e., 23%). Conversely, systems supporting SPARQL queries focus on graph-based sources (25 systems out of 27, i.e., 93%) but support relational sources more frequently (10 out of 27, i.e., 37%) than SQL systems do with graph-based sources.
Federation techniques Besides the supported query languages and data sources, we also considered the specific techniques used by each of the selected systems. Table 3 organizes such techniques according to the main components of a typical data federation system as shown in Fig. 1. Note that the categories Unified schema and mappings and Source selection and query partition are only suitable for transparent federation. For each technique, we provide references to the literature and a list of systems for which the adoption of such technique is stated in official documents or publications. Hence, the lack of the indication of a particular system under a particular technique has to be interpreted as unavailable information, and not as negative information. This holds true especially for closed-source industrial systems, where information about these technical aspects is often covered scarcely or not covered at all in systems’ documentation. We next discuss each element of Table 3.
Summary of the main techniques used in federated query answering, grouped by affected main component of a typical data federation system. For each technique, we provide references to the literature describing the technique, as well as example systems known to implement the technique
Metadata catalog. A fundamental classification of federation techniques for this component distinguishes between techniques where the metadata catalog is automatically built out of source metadata accessed in a standard way (e.g., via the SQL “Information Schema”), and techniques that allow for manual provision of such metadata by users. These technique families are complementary and a system may adopt one or both of them (e.g., Denodo, see Table 3 for other examples). Manually supplied metadata may be described through self-defined dialects, such as the XML syntax of Teiid and the RDF molecule template of Ontario. Alternatively, some systems adopt standard languages, such as the VoID [94] vocabulary for Linked Data [126] (e.g., Squerall) or the SQL extension for the “Management of External Data”, SQL/MED [127] (e.g., Data Virtuality and Teiid, in alternative to its own XML). SQL/MED provides specialized SQL data definition language (SQL DDL) statements, such as CREATE FOREIGN TABLE, for defining the objects stored in the federated sources and how to access them. In place of SQL/MED, other systems (e.g., Apache Spark) use regular or customized versions of plain SQL DDL statements, such as CREATE TABLE with additional clauses, for the purpose of acquiring catalog metadata and without the intent of actually modifying the source itself.
Unified schema and mappings. We divide the federation techniques for this component into two families: the one where the virtual schema is simply a merge of all the source schemas, and the one where the virtual schema is fully customizable by the user. In Table 3, many examples of the former category are SPARQL-based systems that federate SPARQL endpoints, while most of the examples of the latter category are either systems such as PolyWeb that allow the definition of a flexible virtual schema through R2RML/RML mappings, or SQL-based systems that allow the definition of views over the source data, as well as constraints over such views (e.g., primary and foreign keys).
Source selection and query partition. A common approach for the identification of the sources of a query relies on the pre-computation of an index out of the information available in the metadata catalog. Another technique involves the evaluation of probing queries and is exemplified by many SPARQL-based systems in Table 3. One of them is FedX (RDF4J), which issues a probing SPARQL ASK query for each triple pattern in the input query, so as to dynamically identify non-empty sources for that pattern in a more precise, albeit slower, way than using the index. Some systems, like SPLENDID, combine these two approaches to gather their respective strengths. Other systems, like HiBISCuS, propose a refinement of the query-based strategy where the candidate sources identified by the probing queries are further pruned through an analysis based on the structure of the SPARQL query. For SQL-based systems, source selection is straightforward in the typical scenario where tables of the unified schema are mapped 1:1 to their respective sources, but becomes non-trivial when table data is contributed by multiple sources, as it occurs with data partitioning or replication. Teiid “multisource models”,19
for instance, support horizontal table partitioning across sources (e.g., an employee table partitioned across departments) by defining a source-denoting column (e.g., the department name) in the unified table schema, and exploiting WHERE conditions on that column to select a subset of sources to answer the query. For both SPARQL- and SQL-based systems, once sources are identified, query partitioning into sub-queries may involve the push down of query operators to those sources supporting them. An example is the push down of join operators to RDBMS sources [109, p. 326], a technique pioneered in the Garlic system [110].
Query optimization and query plan generation. Some systems rely on fully-fledged cost models for generating an optimized query plan, as per the traditional setting of query answering against a single relational datasource. This plan also indicates the evaluation order of sub-queries and the types of joins to be used to combine their results. In other systems, like Ontario, the optimization is purely driven by heuristics and optimization steps are performed according to a pre-defined set of deterministic rules, such as pushing down certain operators (e.g., selection, projection) as much as possible to reduce the size of intermediate results. Cost-based and rule-based optimization may be also combined to attempt generating better query execution plans, as done for instance by Data Virtuality and SAP HANA. Finally, a complementary technique is the creation of materialized views, which can be used in place of re-computing each time the result of expensive distributed operations, in those scenarios where the source data is expected not to change frequently.
Query execution. Apart from standard join techniques such as nested loop or hash join, data federation systems provide techniques for query plan execution that are specifically tailored towards the federated setting. A common trait of these techniques is that they aim at minimizing data movement across the different systems participating in the federation. In the bind join between two relations, the outer relation is sequentially scanned for join values, which are then used to “bind” the attributes in the inner relation. For each such bind, the matching tuples in the inner relation are transferred to the source of the outer relation and used to construct the result. This approach can be seen as multiple application of the semijoin technique, where one side of the join is first filtered with the matching values, and then this “reduced” relation is sent to the other source for performing the actual join. The broadcast join, instead, “broadcasts” the matching tuples of the inner relation to all sources in the federation, which is an effective strategy when the outer relation is spread across several sources and the inner relation is much smaller than the outer relation. Splitting relations into smaller chunks lies at the basis of the partitioned join, where relations are partitioned according to values of the join keys. This join technique works in combination with parallelization, where computation is performed in a distributed way across multiple nodes at the same time. Finally, caching of the intermediate results allows further diminishing the number of distributed operations performed, and is popular among industrial systems.
Data security dimension
Evaluation of the data security dimension. Academic systems in italics. “–” denotes feature/information not found in the systems’ official documentation, websites, or academic publications, to the best of our efforts. Subscript denotes the use of named graph-based solutions to hide (mask) sensitive information in selected graphs to certain users, and possibly (for AnzoGraph DB) expose sanitized named graph views
We evaluate here the data security dimension. The concrete investigation results are shown in Table 4, organized according to the sub-dimensions of authentication, authorization, auditing, encryption, and data masking. In particular, by analyzing the information we synthesized in the table, the following can be observed:
Almost all the considered industrial systems (31 out of 33, i.e., 94%) provide security mechanisms, such as authentication and authorization, to protect against unauthorized data access and leaking. This shows that the importance of data security is actually recognized by system providers in the data federation setting, where integrating multiple data sources via a unified virtual layer has the potential of making the private and sensitive data contained in federated sources more likely to be revealed.
Among the inspected mechanisms, authentication and authorization are definitely the most frequently adopted ones (see total counts in Table 4) and are implemented by almost all the industrial systems to identify users and control their access to data. For example, the Denodo Platform supports role-based authentication20
and enforces strict and fine-grained row and column level access control.
Besides authentication and authorization, the other three mechanisms, i.e., auditing, encryption, and data masking, are adopted by some industrial (only) systems to enhance security by auditing the actions of users and encoding and hiding sensitive information. Take again Denodo as an example. The Denodo Platform provides an audit trail of all the information about the queries and other actions executed on the system. It also supports the application of strategies on a per-view basis to guarantee secure access to sensitive data through encryption/decryption at different levels, and it masks (hides) sensitive data to ensure they are not accessed by unauthorized users. In SPARQL federation engines, data masking is provided by allowing hiding named graphs with sensitive information to certain users in AnzoGraph DB21
as a way to define (via SPARQL CONSTRUCT queries) sanitized/masked named graphs to be exposed in place of sensitive ones.
Data security has rarely been mentioned in the systems developed by academic and research institutions. Among the 18 systems we have evaluated in this category, just one system, i.e., SAFE, takes data security into consideration. SAFE is a SPARQL query federation engine that enables policy-aware access to sensitive, distributed statistical data sources represented as RDF data cubes.
Interface dimension
Table 5 reports on the evaluation of the interface dimension, which is used to qualitatively evaluate the usability of the systems from both the end-user and the developer perspectives. As mentioned in Section 4 and reflected in the table, this dimension comprises the graphical, command line, and application programming interface sub-dimensions. Here, we analyze which of these interfaces are made available to the users, further identifying the different types of exposed application programming interfaces (e.g., JDBC drivers, web APIs). We cover only documented (vs. hidden in the code) interfaces and we do not consider effectiveness and ease of use, whose evaluation is largely subjective as, for any given interface, user experience is affected by individual user’s preferences and habits. In summary, from Table 5 we can derive the following observations:
Evaluation of the interface dimension. Academic systems in italics. “–” denotes feature/information not found in the systems’ official documentation, websites, or academic publications, to the best of our efforts
Nearly all of the industrial systems (31 out of 33, i.e., 94%) provide graphical interfaces, which consist mainly in web consoles or web interfaces, and command line interfaces (all 33 industrial systems), which help users to deploy and manage data federation instances. For example, AllegroGraph provides the AllegroGraph Web View,24
https://allegrograph.com/products/agwebview/
which is a browser-based graphical interface for exploring, querying, and managing AllegroGraph databases, and Teiid provides users with Teiid Console,25
Besides graphical and command line interfaces, most industrial systems like Denodo and Teiid also provide JDBC and ODBC drivers (respectively, 23 and 18 systems out of 33, i.e., 70% and 55%) to enable users to access and interact with them as standard relational sources. Web APIs (mainly RESTful) are also very frequent among industrial systems (25 out of 33, i.e., 76%), while there is less support for ADO.NET and the SPARQL HTTP API. The latter is exclusively provided by systems supporting the SPARQL query language (see Table 2) that also directly implement the associated SPARQL HTTP query protocol (instead of relying on other non-standard means for receiving a SPARQL query and returning its results). Furthermore, few systems, such as AllegroGraph, Presto and Stardog, provide also multiple client libraries to help users in interfacing with these systems using the most popular programming languages, like C, Go, Java, Python, R, and Ruby.
The three systems not associated to any interface in the table are all academic (Fedra, HiBISCuS, SAFE). For these systems, the documentation only covers the experiments conducted and indicates, at most, the script (Fedra) or the code entry points (HiBISCuS) for reproducing the specific experiments.
Development dimension
Table 6 reports on the evaluation of the development dimension and its sub-dimensions, which all together deliver information relevant to developers for integrating the system with other applications or for patching, extending, or otherwise modifying the system itself, if possible. Note that for the industrial systems, the information of the first release, i.e., the year and version number of the first version made available, is actually the information of the oldest versions we have been able to gather from their official websites. Note also that the academic systems often do not follow well-defined release cycles with proper versioning, e.g., CostFed.26
https://github.com/dice-group/CostFed
In such situations, we leave their versions as blank, and fill the years from their commit histories on their GitHub projects. The following are the main insights we can get from Table 6:
Evaluation of development dimension. Academic systems in italics. “F.” and “L.” denote “First” and “Latest” respectively. Subscript letters further qualify available deployment options: n = native; c = containerized; a = Amazon AWS; m = Microsoft Azure; g = Google Cloud Platform. “–” denotes feature/information not found in the systems’ official documentation, websites, or academic publications, to the best of our efforts
Java is the most used programming language for both industrial and academic systems, even when accounting for the incomplete information of this sub-dimension (see counts in Table 6). Comparatively less used languages include C/C++ (AnzoGraph DB, SAP HANA and other systems), Python (RDFLib, Squerall, SAGE), Scala (Apache Spark, Ontario), JavaScript (Comunica) and Lisp (AllegroGraph, in combination with Java).
Excluding two SaaS industrial systems from Amazon (Athena, Neptune), on-premises deployment is always offered, represents the only available option for academic systems, and concerns software both in native form (n subscript, almost always possible) and containerized form (c subscript, e.g., via Docker images), the latter supported more in industrial systems (21 out of 33, i.e., 64%) than academic systems (4 out of 18, i.e., 22%). SaaS (6 industrial systems out of 33, i.e., 18%) is less frequent than IaaS/PaaS (12 out of 33, i.e., 36%), the latter always supporting Amazon AWS (a subscript), followed by Microsoft Azure (m subscript, 8 IaaS/PaaS cases out of 12, i.e., 67%) and Google Cloud Platform (g subscript, 5 IaaS/PaaS cases out of 12, i.e., 42%).
Among the industrial systems, the majority are closed source (21 out of 33, i.e., 64%), and most of these come with commercial support services (19 systems out of 21, i.e., 90%). Similarly, most of the open source industrial systems offer the option of commercial support (7 systems out of 12, i.e., 58%). Academic systems are all open source without commercial support.
In comparison with academic systems, it is easy to see that industrial ones typically feature a much more active development. Some of these industrial systems have been developed, maintained, and improved for many years, such as Denodo and Teiid. Unfortunately, for the academic systems, despite the fact that all of them are open source initiatives, it is common that they are not enhanced or maintained after the publication of the respective academic papers.
Overall discussion and analysis
Based on the above reported evaluation and analysis, and after having reviewed the official documentation and academic publications of each of the systems considered in this survey, in the following we summarize the most crucial and interesting lessons we learned.
Background theory and standards Data federation, especially over heterogeneous data sources, is currently a very active field in both industry and academia. However, the overall development of data federation systems still seems to lack background theory and standards. Let us note, for instance, that different systems force users to adopt their own dialects to develop and model the logical or meta-data layer of the target data sources. This strategy drastically hinders information reuse, as information produced for one system cannot be directly used in other systems.
Other capabilities Among the other capabilities beyond the data federation task itself (cf. Section 2.2), only data security was captured by our evaluation framework, which is based solely on the aspects of interest arising from applying the methodology of Section 3.2. This fact further remarks the importance of data security, especially among industrial systems, whereas data update and data quality have been less investigated in combination with the data federation. Nevertheless, some of the considered systems provide capabilities related to data update and data quality. Concerning data update over the federated data sources, Teiid27
supports SPARQL UPDATE over the federated SPARQL endpoints. Other systems mention data update, however it is unclear from the systems’ documentation whether these updates can be performed on the data sources in the federation, or on the data stored locally by the system itself (e.g., for database systems extended with federation facilities). Concerning data quality, SAS Federation Server30
supports methods and rules specified in a “SAS Quality Knowledge Base” (QKB), while Stardog31
https://docs.stardog.com/data-quality-constraints
supports data quality constraints expressed in SHACL [128]. Given the current steady growth of data scale and variety, we expect these aspects to become increasingly important in the context of data federation.
Ontology-based data access Ontologies, providing a shared abstraction of a domain of interest, can play a key role in handling the heterogeneity of concepts in data integration. The so-called Ontology-Based Data Access (OBDA) approach has been studied intensively [37,77,129–134] in the last two decades. In OBDA, a mediating ontology provides a high-level representation of the data contained in a relational source, as well as an encoding of domain knowledge. The link between the ontology and the source is realized through mappings, e.g., expressed using R2RML [135]. The distinctive characteristics of OBDA are that query answers are enriched through automated reasoning over the ontology, and that such process is carried out in a virtual mode: the data in the database is not materialized as a graph, but rather queries are rewritten on-the-fly and executed against the original source.
The virtual characteristic of OBDA makes it a potential candidate for incorporating mediating ontologies in the data federation framework. Still, this marriage has rarely been discussed or considered to its fullest extent, and it represents an open research line. For instance, Squerall [96] and PolyWeb [32,85] are virtual systems based on RML/R2RML mappings but both lack reasoning support, hence they do not qualify as fully-fledged OBDA systems as per their definition in the literature [129]. An exception is Obi-Wan [75,76], an OBDA system32
Although, based on GLAV mappings as opposed to GAV mappings usually applied in OBDA contexts.
able to integrate heterogeneous data sources, including relational, graph-based, and NoSQL ones. Its main idea follows the classical OBDA framework by first rewriting the original queries based on the ontology and the mappings, and then using the mediator system Tatooine [78] to evaluate the rewritten queries over multiple and heterogeneous data sources.
Obi-Wan is for the most part a proof-of-concept of a more general and insightful theoretical exercise. Hence, it does not present any optimization technique specific to the federated setting and is not tailored towards handling real-world, complex scenarios. Using domain ontologies to virtually integrate heterogeneous data sources combines the difficulties of ontology reasoning with the ones of integrating heterogeneous data, and this negatively affects performance. Further investigations and, possibly, innovative approaches are required to obtain systems that would exhibit a performance that is adequate to real-world application needs. A preliminary investigation towards this direction has been conducted by Gu et al. [136,137]. The use of ontology-based techniques – and, more generally, of Semantic Web methods and standards – to address data quality, update, and security aspects of data federation systems also appears promising and deserves further research.
Interrelationships between data sources Most of the time, the data sources that are subject to a data integration initiative are not fully independent from each other. Indeed, there may exist interrelationships among the integrated data sources, such as information overlapping, complementarity, and conflicts. Automatically discovering such interrelationships may help developing data federation systems of higher efficiency. As a simple example, if a data source is part of a data source with respect to the metadata layer (both schema and content), then in the query evaluation procedure may be sometimes ignored (e.g., when querying for the union of the content of and ) and the overall performance improved.
Most advanced methods and systems handle overlapping to some extent. BigDAWG exploits equivalence and containment information provided by data curators [138] to identify equivalent operations across different data models, so as to optimize its source-selection strategy. DAW [139], not considered in this survey, exploits a compact representation of data as vectors for which estimates on overlapping can be automatically found. This information is then used to prune, with high recall, redundant sources during source selection. FEDRA [64] does not require to encode data, but relies on fragment descriptions for its source selection, where each fragment essentially describe the triples that can be extracted out of a set of data sources.
Note that all approaches require a substantial amount of meta-information which might be hard or even impossible to produce automatically. It has been recently observed by Gu et al. [136,137] that this limitation is greatly reduced in OBDA settings, where one can exploit both the semantic information provided by the ontology and the URI construction rules encoded in the mappings. This fact allows for optimizations that are not specific to the source selection phase, such as the removal of redundant or empty operators, or the automatic leveraging of materialization of pre-computed results and on-the-fly access to the sources.
Combining systems The capabilities of a system can be extended through combination with other tools. We identify two mechanisms for combining a data federation system with a tool, the latter operating as adapter and possibly being a data federation system itself; these mechanisms can be iteratively applied to combine multiple components.
In the first mechanism, the tool acts as a source of the system and is used to add indirect support for some additional sources that cannot be natively connected to the system, by adapting them to one of the supported source types (e.g., JDBC or ODBC). For instance, the data sources directly supported by Querona Data Virtualization exclude MongoDB but include Denodo and Apache Drill, which instead support MongoDB and can be thus combined to add indirect support for MongoDB. As another example, one may extend SPARQL-based federation to relational sources through the combination with an OBDA engine, as successfully applied by Sima et al. [140] who use the OBDA system Ontop to expose biomedical data as RDF graphs, then federated through a SPARQL federation engine.
In the second mechanism, the tool acts as client of the system and is used to adapt or extend the unified schema, query language(s) or capabilities offered by the system. For example, one may deploy33
https://ontop-vkg.org/tutorial/federation/
an OBDA engine like Ontop over a SQL-based data federation system such as Dremio or Denodo, so to provide indirect support for an RDF/OWL unified schema and SPARQL as unified query language. From a complementary perspective, this combination mechanism can be also seen as adding federation capabilities to the employed tool (the OBDA engine in the example), effectively giving birth to a new data federation system.
As remarked in the text (Section 6.1), the “Data source” dimension of Table 2 and in general all the dimensions and tables of this survey do not account for the combination of systems, but rather focus solely on sources and capabilities that are directly supported by the data federation system. The reason is that it is very difficult to comprehensively assess which sources or capabilities a data federation system may acquire by carefully combining it with other tools, as combinations are possibly limitless and the assessment of the practical feasibility of each is non-trivial and not clearly defined, as there might be hard-to-quantify integration costs involved (e.g., to remove minor incompatibilities at the interface between combined tools).
Related work
In this survey, we have investigated and analyzed a total of 51 data federation systems. Considering data federation in the broader context of data integration, in the following we situate this survey among other works in the Database and the Semantic Web literature that review existing approaches, techniques, and systems for both virtual and materialized data integration.
Database community Halevy et al. [6] discuss some of the most important results in the data integration field before 2006, and outline some challenges for data integration research. The survey by Magnani and Montesi [141] reports on the techniques for managing uncertainty in data integration, and the survey by Bikakis et al. [142] investigates the approaches focusing on semi-structured data. Finally, the works by Arputhamary et al. [143–145] mostly address the issues emerging when techniques and systems are meant to be applied to integrate big data.
Readers that are interested in knowing more about existing approaches and implemented systems for integrating data virtually can refer to several related surveys [9,42,43,146]. In particular, the survey by Sheth and Larson [9] discusses data federation systems. The authors define terminology and a “reference architecture” for distributed database management systems with the main aim of providing a framework in which to understand, categorize, and compare different architectural options for developing federated database systems. Additionally, they introduce a methodology for developing tightly coupled federated database systems with multiple federations and processors (that is, software modules that manipulate commands and data). In a different survey, Bondiombouy and Valduriez [146] investigate multistore systems by first introducing the currently available cloud data management and query processing solutions, then describing and analyzing some representative multistore systems according to their architecture, data model, query languages, and query processing techniques. They finally classify these systems into three categories, i.e., loosely-coupled, tightly-coupled, and hybrid. The survey by Tan et al. [43] focuses on query processing over heterogeneous data sources by first introducing a taxonomy that categorizes the solutions into data federation systems, polyglot systems, multistore systems, and polystore systems. On top of this categorization, the authors propose an evaluation framework, largely inspired by the work by Sheth and Larson [9], incorporating the axes of “Heterogeneity”, “Autonomy”, “Transparency”, “Flexibility” and “Optimality”. The survey finally compares and analyzes four specific systems – BigDAWG, CloudMdsQL, Myria, and Apache Drill – according to the introduced evaluation framework. Azevedo et al. [42] focus on new generation data federation systems addressing the manipulation of structured and unstructured data, usually in high volume, over distributed and heterogeneous data sources. The authors first review the literature aiming at giving an overview of state-of-the-art modern data federation systems and then analyze the four aforementioned systems – BigDAWG, CloudMdsQL, Myria, and Apache Drill – by reporting on their “Definition”, “Owners”, “Goals”, “Query Specification and Execution”, “Main Components”, and other significant dimensions.
Semantic web community Wache, Noy, Ekaputra et al. [147–149] provide general surveys of those solutions for integrating data that are based on Semantic Web technologies and that follow the so-called Ontology-Based Data Integration (OBDI) approach. OBDI is a broader approach than OBDA, and differs from the latter for the fact of allowing for very expressive ontology languages while dropping the requirement of virtual access to data. Hence, OBDI approaches are not really suited to the federation setting considered in this survey. Other works focus instead on specific subdomains in which semantic technologies have been applied to integrate data. In particular, Buccella et al. [150] analyze and compare existing approaches for ontology-driven geographic information integration. An investigation of the approaches and techniques developed in the ontology community for integrating biological data is given by Hassan et al. [151]. The survey by Mountantonakis and Tzitzikas [152] investigates the works that have been done in the area of Linked Data integration, covering both materialized and virtual integration approaches. This work provides a concise overview of the issues, methods, tools, and systems for semantic integration of data, and gives emphasis on the methods that provide support for the integration of large numbers of datasets.
As for the virtual approach to data integration, some literature can be found [20,38–41] surveying, in particular, approaches and systems for federated SPARQL query answering. To summarize, the survey by Rakhmawati et al. [38] gives an overview of SPARQL federation frameworks – i.e., frameworks supporting (i) SPARQL 1.1 federation extension, (ii) federation over SPARQL 1.0 endpoints, and (iii) federation over SPARQL 1.1 endpoints – and classifies and analyzes 14 existing SPARQL federation approaches. Oguz et al. [20] evaluate 7 federation engines by first providing a detailed and clear insight on data source selection, join, and query optimization methods. They also introduce a qualitative comparison of these engines according to the following criteria: “No Preprocessing per Query”, “Unbound Predicate Queries”, “Parallelization”, and “Adaptive Query Processing”. Ngonga Ngomo and Saleem [39] provide an overview of current challenges and opportunities of federated query processing as well as summarize the results of recent state-of-the-art studies. Saleem et al. [40] first provide a survey of 14 federated SPARQL query engines according to: “Code Availability”, “Implementation Language”, “Licensing”, “Source Selection Type”, “Join Type”, “Cache”, and “Index/Catalog Update”. They then compare 5 SPARQL endpoint federation systems by using the performance evaluation framework FedBench [153] and by considering the dimensions of query runtime, number of sources selected, total number of SPARQL ASK requests used, completeness of answers, and source selection time. Finally, Qudus et al. [41] first propose some metrics to measure the errors in cardinality estimations of cost-based federation engines and the correlation of the values of these metrics with the overall query runtimes. Then, they present an empirical evaluation of 5 cost-based SPARQL federation engines on LargeRDFBench [154] according to the proposed metrics.
Comparison This survey builds on the aforementioned literature and is consistent with the terminology, concepts and key distinctions adopted therein. For instance, considering the foundational work by Sheth and Larson [9], their terminology can be related to several (sub-)dimensions of our evaluation framework as follows: (i) “Heterogeneity” is captured at different levels by our Data source, Query language, Federation technique and various Development sub-dimensions; (ii) “Query processing and optimization” is also captured by our Federation technique sub-dimension; (iii) “Access Control” is related to our Data security dimension (especially, its Authorization sub-dimension). (iv) “Transparency” is captured by the distinction between Transparent vs. Explicit federation.
The key difference between our work and the aforementioned surveys is mainly reflected in the following two aspects. First, we have analyzed and investigated a larger number of systems, including among them both industrial and academic initiatives and systems adopting different data models, i.e., SQL-based and SPARQL-based. Second, we have introduced here as a novel contribution a framework to inspect, analyze, and then classify the main characteristics of each system. The framework has been developed by taking into consideration the requirements of the end-users, as well as those of the developers and of the scholars, this way trying to deliver the information that they need when making choices for their respective data federation activities and projects. Our main motivation is to assess the techniques and capabilities of the existing systems for data federation, so as to reveal their strengths and weaknesses in relation to the plurality of evaluation dimensions we consider, rather than classifying the systems along one single dimension or according to the requirements of one single category of prototypical users.
Concluding remarks and future work
In this paper, we provided a systematic overview of 51 data federation systems, with the motivation of evaluating their capabilities as well as the strengths and weaknesses of the employed techniques for integrating heterogeneous data sources uniformly and virtually. To do so, we have proposed a framework with four major dimensions and additional sub-dimensions to classify systems from the end-user, the developer, and the scholar perspectives, in a uniform and qualitative way. We think that the evaluation framework we have proposed can be valuable for all these target personas: it helps end-users in finding the system that most suits their application requirements and, at the same time, it drives decision making by developers and researchers in further improving the currently available solutions and in designing more powerful federation systems. Besides that, our work also aims at providing up-to-date reference information for all those interested in dipping their toes in the data federation water.
Integrating and managing heterogeneous data “uniformly and virtually” still have a long way to go both at the theoretical and at the practical application levels. Our future work will mainly focus on the following two aspects. In our current evaluation, efficiency of the investigated systems remains an ignored dimension. Therefore, one direction for future work is to design extensive experiments to evaluate the performance and assess the restrictions of each system in integrating and managing heterogeneous data virtually. On the other hand, it is well known that the Semantic Web provides standards for both knowledge and data representation and management. However, integrating heterogeneous data virtually by relying on semantic technologies and Semantic Web standards still represents an open and promising research field. The second main direction we want to take is indeed to develop innovative approaches for ontology-based heterogeneous data integration and management, covering federated query answering, data updates, security, and data quality assurance, where automated logic-based reasoning techniques play a central role.
Footnotes
Acknowledgements
This research has been partially supported by the EU H2020 project INODE (grant agreement No. 863410), by the Italian PRIN project HOPE (2019-2022), by the European Regional Development Fund (ERDF) Investment for Growth and Jobs Programme 2014-2020 through the project IDEE (FESR1133), by the RTD DM 1062/2021 project OntoCRM, by the Free University of Bozen-Bolzano through the project MP4OBDA, and by the “Fusion Grant” project HIVE sponsored by Fondazione Cassa di Risparmio di Bolzano and Ontopic s.r.l. in coordination with NOI Techpark, Südtiroler Wirtschaftsring and Rete Economia Alto Adige. G. Xiao is supported by the Norwegian Research Council via the SIRIUS Centre for Research Based Innovation (grant No. 237898). D. Calvanese is supported by the Wallenberg AI, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg Foundation. We thank our colleagues, in particular Julien Corman, for their discussions and feedback. We would also like to thank the reviewers for their valuable feedback and comments on earlier versions of this article.
Specific data sources supported by the selected systems
Table 7 lists the specific sources supported by each investigated data federation system, obtained from available systems’ documentation and publications. Sources are classified on a local, per-system basis, along the source types defined in Section 6.1, with additional source information – such as the specific kind(s) of relational, graph-based or aggregate-oriented system – reported next to the source name via subscript letters (see table caption for legend). We remark the following:
Some sources correspond to data access interfaces that can be configured to connect additional systems beyond the ones explicitly listed in the table. In particular, companies such as CData34
commercialize connectors for the relational SQL-based JDBC, ODBC, ADO.NET and OLE DB interfaces that can be used to access a myriad of heterogeneous data sources, possibly different from the ones listed in Table 7 (e.g., GraphQL sources via specific connectors36
https://www.cdata.com/drivers/graphql/
) and possibly using a different data model that is transparently adapted to the relational one by the connector (e.g., , via flattening of nested data). In Table 7, besides the supported data access interfaces, we explicitly list only the sources that are directly and natively supported by a system without relying on such third party connectors / adapters.
Structured files are distinguished from other source types with the same data model (e.g., relational sources for CSV files, aggregate-oriented – specifically, document-based – for JSON files) by virtue of direct access to raw file contents by the data federation system. In some cases, however, access to stored structured files may require metadata services external to the filesystem (e.g., Hive Metadata Store) for locating and interpreting file contents, or may leverage processing services (e.g., from Hadoop) co-located with the nodes storing the file in a distributed filesystem (e.g., HDFS), for instance to push down data access operations and computations (e.g., filtering, sorting) close to where raw file data reside, this way reducing communication costs.
Some of the data federation systems investigated in this survey are also listed as supported sources (marked with ∗ subscript) of other systems in Table 7, reflecting the fact that the virtual data sources obtained through data federation can be used themselves in downstream federations. As a limit case (e.g., AllegroGraph), a system may list only itself as a supported data source, which occurs when the system offers both storage and data federation capabilities, and the latter are restricted to instances of the same system.
Test sources (e.g., emulating /dev/null) and system-specific connectors used to access configuration, performance or log data of the system itself are omitted in Table 7, for simplicity.
Selection of academic systems
We report further details about the academic systems selection process described in Section 3.1, providing: (i) the statistics of the 295 academic publications found in our literature search (Section B.1); (ii) the metadata and the considered systems and aspects for the 17 system comparison publications found among them (Section B.2); (iii) the inclusion criteria satisfied or violated by the 56 academic systems found, which support our selection of 18 academic systems (Section B.3); and (iv) the full bibliography of all the 295 collected academic publications (Section B.4).
References
1.
D.Reinsel, J.Gantz and J.Rydning, The Digitization of the World from Edge to Core, International Data Corporation, Framingham, MA, 2018, Technical Report.
2.
A.Labrinidis and H.V.Jagadish, Challenges and opportunities with big data, Proc. of VLDB Endowment5(12) (2012), 2032–2033. doi:10.14778/2367502.2367572.
3.
S.Sagiroglu and D.Sinanc, Big data: A review, in: Proc. of Int. Conf. on Collaboration Technologies and Systems (CTS), IEEE, 2013, pp. 42–47. doi:10.1109/CTS.2013.6567202.
4.
M.Lenzerini, Data integration: A theoretical perspective, in: Proc. of ACM Symp. on Principles of Database Systems (PODS), ACM, 2002, pp. 233–246. doi:10.1145/543613.543644.
5.
A.Doan, A.Y.Halevy and Z.G.Ives, Principles of Data Integration, Morgan Kaufmann, 2012. ISBN 978-0-12-416044-6. doi:10.1016/C2011-0-06130-6.
6.
A.Y.Halevy, A.Rajaraman and J.J.Ordille, Data integration: The teenage years, in: Proc. of Int. Conf. on Very Large Data Bases (VLDB), ACM, 2006, pp. 9–16.
7.
J.Widom, Research problems in data warehousing, in: Proc. of Int. Conf. on Information and Knowledge Management (CIKM), ACM, 1995, pp. 25–30. doi:10.1145/221270.221319.
8.
S.Chaudhuri and U.Dayal, An overview of data warehousing and OLAP technology, SIGMOD Record26(1) (1997), 65–74. doi:10.1145/248603.248616.
9.
A.P.Sheth and J.A.Larson, Federated database systems for managing distributed, heterogeneous, and autonomous databases, ACM Computing Surveys22(3) (1990), 183–236. doi:10.1145/96602.96604.
10.
L.M.Haas, E.T.Lin and M.T.Roth, Data integration through database federation, IBM Systems J.41(4) (2002), 578–596. doi:10.1147/sj.414.0578.
11.
C.J.Date and H.Darwen, A Guide to the SQL Standard, 4th edn, Addison-Wesley, 1996.
M.Krötzsch, P.Patel-Schneider, S.Rudolph, B.Parsia and P.Hitzler, OWL 2 Web Ontology Language Primer (Second Edition), W3C Recommendation, W3C, 2012, https://www.w3.org/TR/2012/REC-owl2-primer-20121211/.
16.
R.van der Lans, Data Virtualization for Business Intelligence Systems: Revolutionizing Data Integration for Data Warehouses, 1st edn, Morgan Kaufmann Publishers, San Francisco, CA, USA, 2012. ISBN 0123944252.
17.
A.Bogdanov, A.Degtyarev, N.Shchegoleva, V.Korkhov and V.Khvatov, Big data virtualization: Why and how? in: Proc. of 4th Int. Workshop on Data Life Cycle in Physics (DLC), CEUR Workshop Proceedings, Vol. 2679, 2020, pp. 11–21.
18.
H.Betz, F.Gropengießer, K.Hose and K.Sattler, Learning from the history of distributed query processing – A heretic view on linked data management, in: Proceedings of the Third International Workshop on Consuming Linked Data, COLD 2012, Boston, MA, USA, November 12, 2012, CEUR Workshop Proceedings, Vol. 905, CEUR-WS.org, 2012.
19.
O.Görlitz and S.Staab, Federated data management and query optimization for linked open data, in: New Directions in Web Data Management 1, Studies in Computational Intelligence, Vol. 331, 2011, pp. 109–137. doi:10.1007/978-3-642-17551-0_5.
20.
D.Oguz, B.Ergenc, S.Yin, O.Dikenelli and A.Hameurlain, Federated query processing on linked data: A qualitative survey and open challenges, Knowledge Engineering Review30(5) (2015), 545–563. doi:10.1017/S0269888915000107.
21.
A.Schwarte, P.Haase, K.Hose, R.Schenkel and M.Schmidt, FedX: A federation layer for distributed query processing on linked open data, in: Proc. of Extended Semantic Web Conference (ESWC), LNCS, Vol. 6644, Springer, 2011, pp. 481–486. doi:10.1007/978-3-642-21064-8_39.
22.
A.Schwarte, P.Haase, K.Hose, R.Schenkel and M.Schmidt, FedX: Optimization techniques for federated query processing on linked data, in: Proc. of Int. Semantic Web Conf (ISWC), LNCS, Vol. 7031, Springer, 2011, pp. 601–616. doi:10.1007/978-3-642-25073-6_38.
23.
Teiid, Accessed 16 November 2021. https://teiid.io/.
Data Virtuality, Accessed 17 November 2021, https://datavirtuality.com/.
26.
M.N.M.Nazri, S.A.Noah and Z.Hamid, Using lexical ontology for semi-automatic logical data warehouse design, in: Proc. of Int. Conf. on Rough Set and Knowledge Technology (RSKT), LNCS, Vol. 6401, Springer, 2010, pp. 257–264. doi:10.1007/978-3-642-16248-0_39.
27.
S.Bouarar, L.Bellatreche, S.Jean and M.Baron, Do rule-based approaches still make sense in logical data warehouse design? in: Proc. of East European Conf. on Advances in Databases and Information Systems (ADBIS), LNCS, Vol. 8716, Springer, 2014, pp. 83–96. doi:10.1007/978-3-319-10933-6_7.
28.
K.M.Endris, P.D.Rohde, M.-E.Vidal and S.Auer, Ontario: Federated query processing against a semantic data lake, in: Proc. of Int. Conf. on Database and Expert Systems Applications (DEXA), LNCS, Vol. 11706, Springer, 2019, pp. 379–395. doi:10.1007/978-3-030-27615-7_29.
29.
F.Ravat and Y.Zhao, Data lakes: Trends and perspectives, in: Proc. of Int. Conf. on Database and Expert Systems Applications (DEXA), LNCS, Vol. 11706, Springer, 2019, pp. 304–313. doi:10.1007/978-3-030-27615-7_23.
30.
R.Hai, S.Geisler and C.Quix, Constance: An intelligent data lake system, in: Proc. of ACM SIGMOD Int. Conf. on Management of Data (SIGMOD), ACM, 2016, pp. 2097–2100. doi:10.1145/2882903.2899389.
31.
R.Hai, C.Quix and C.Zhou, Query rewriting for heterogeneous data lakes, in: Proc. of European Conf. on Advances in Databases and Information Systems (ADBIS), LNCS, Vol. 11019, Springer, 2018, pp. 35–49. doi:10.1007/978-3-319-98398-1_3.
32.
Y.Khan, A.Zimmermann, A.Jha, V.Gadepally, M.d’Aquin and R.Sahay, One size does not fit all: Querying web polystores, IEEE Access7 (2019), 9598–9617. doi:10.1109/ACCESS.2018.2888601.
33.
J.Duggan, A.J.Elmore, M.Stonebraker, M.Balazinska, B.Howe, J.Kepner, S.Madden, D.Maier, T.Mattson and S.B.Zdonik, The BigDAWG polystore system, SIGMOD Record44(2) (2015), 11–16. doi:10.1145/2814710.2814713.
34.
J.Wang, T.Baker, M.Balazinska, D.Halperin, B.Haynes, B.Howe, D.Hutchison, S.Jain, R.Maas, P.Mehta, D.Moritz, B.Myers, J.Ortiz, D.Suciu, A.Whitaker and S.Xu, The Myria big data management and analytics system and cloud services, in: Proc. of Biennial Conf. on Innovative Data Systems Research (CIDR), 2017, www.cidrdb.org.
35.
B.Kolev, C.Bondiombouy, P.Valduriez, R.Jiménez-Peris, R.Pau and J.Pereira, The CloudMdsQL multistore system, in: Proc. of ACM SIGMOD Int. Conf. on Management of Data (SIGMOD), ACM, 2016, pp. 2113–2116. doi:10.1145/2882903.2899400.
36.
R.Alotaibi, B.Cautis, A.Deutsch, M.Latrache, I.Manolescu and Y.Yang, ESTOCADA: Towards scalable polystore systems, Proc. of VLDB Endowment13(12) (2020), 2949–2952. doi:10.14778/3415478.3415516.
37.
D.Calvanese, G.De Giacomo, D.Lembo, M.Lenzerini and R.Rosati, Tractable reasoning and efficient query answering in description logics: The DL-lite family, J. Automated Reasoning39(3) (2007), 385–429. doi:10.1007/s10817-007-9078-x.
38.
N.A.Rakhmawati, J.Umbrich, M.Karnstedt, A.Hasnain and M.Hausenblas, A comparison of federation over SPARQL endpoints frameworks, in: Proc. of 4th Int. Conf. on Knowledge Engineering and the Semantic Web (KESW), CCIS, Vol. 394, Springer, 2013, pp. 132–146. doi:10.1007/978-3-642-41360-5_11.
39.
A.-C.Ngonga Ngomo and M.Saleem, Federated query processing: Challenges and opportunities, in: Proc. of Int. Workshop on Dataset Profiling and Federated Search for Linked Data (PROFILES), CEUR Workshop Proceedings, Vol. 1597, CEUR-WS.org, 2016.
40.
M.Saleem, Y.Khan, A.Hasnain, I.Ermilov and A.-C.Ngonga Ngomo, A fine-grained evaluation of SPARQL endpoint federation systems, Semantic Web7(5) (2016), 493–518. doi:10.3233/SW-150186.
41.
U.Qudus, M.Saleem, A.-C.Ngonga Ngomo and Y.-k.Lee, An empirical evaluation of cost-based federated SPARQL query processing engines, Semantic Web0(1) (2019), 1–26. doi:10.3233/SW-200420.
42.
L.G.Azevedo, E.F.de Souza Soares, R.Souza and M.F.Moreno, Modern federated database systems: An overview, in: Proc. of 22nd Int. Conf. on Enterprise Information Systems (ICEIS), SCITEPRESS, 2020, pp. 276–283. doi:10.5220/0009795402760283.
43.
R.Tan, R.Chirkova, V.Gadepally and T.G.Mattson, Enabling query processing across heterogeneous data models: A survey, in: Proc. of Int. Conf. on Big Data (BigData), IEEE Computer Society, 2017, pp. 3211–3220. doi:10.1109/BigData.2017.8258302.
44.
Apache Drill, Accessed 18 November 2021, https://drill.apache.org/.
45.
B.Quilitz and U.Leser, Querying distributed RDF data sources with SPARQL, in: Proc. of European Semantic Web Conf (ESWC), LNCS, Vol. 5021, Springer, 2008, pp. 524–538. doi:10.1007/978-3-540-68234-9_39.
46.
O.Görlitz and S.Staab, SPLENDID: SPARQL endpoint federation exploiting VOID descriptions, in: Proc. of 2nd Int. Workshop on Consuming Linked Data (COLD), CEUR Workshop Proceedings, Vol. 782, CEUR-WS.org, 2011.
47.
AllegroGraph, Accessed 18 November 2021, https://allegrograph.com/.
48.
Amazon Athena, Accessed 18 November 2021, https://docs.aws.amazon.com/athena/latest/ug/work-with-data-stores.html.
49.
Presto, Accessed 18 November 2021, https://prestodb.io/.
50.
Amazon Neptune, Accessed 18 November 2021, https://aws.amazon.com/neptune/.
51.
Anzograph, Accessed 17 November 2021, https://www.cambridgesemantics.com/anzograph/.
52.
M.Hausenblas and J.Nadeau, Apache drill: Interactive ad-hoc analysis at scale, Big Data1(2) (2013), 100–104. doi:10.1089/big.2013.0011.
53.
S.Melnik, A.Gubarev, J.J.Long, G.Romer, S.Shivakumar, M.Tolton and T.Vassilakis, Dremel: Interactive analysis of web-scale datasets, Communications of the ACM54(6) (2011), 114–123. doi:10.1145/1953122.1953148.
54.
Jena, Accessed 18 November 2021, https://jena.apache.org/documentation/query/.
55.
Spark SQL, Accessed 18 November 2021, https://spark.apache.org/sql/.
56.
M.Armbrust, R.S.Xin, C.Lian, Y.Huai, D.Liu, J.K.Bradley, X.Meng, T.Kaftan, M.J.Franklin, A.Ghodsi and M.Zaharia, Spark SQL: Relational data processing in spark, in: Proc. of ACM SIGMOD Int. Conf. on Management of Data (SIGMOD), ACM, 2015, pp. 1383–1394. doi:10.1145/2723372.2742797.
57.
V.Gadepally, K.O’Brien, A.Dziedzic, A.J.Elmore, J.Kepner, S.Madden, T.Mattson, J.Rogers, Z.She and M.Stonebraker, BigDAWG version 0.1, in: Proc. of IEEE High Performance Extreme Computing Conf. (HPEC), IEEE, 2017, pp. 1–7. doi:10.1109/HPEC.2017.8091077.
58.
Blazegraph, Accessed 6 December 2021, https://blazegraph.com/.
59.
B.Kolev, P.Valduriez, C.Bondiombouy, R.Jiménez-Peris, R.Pau and J.Pereira, CloudMdsQL: Querying heterogeneous cloud data stores with a common language, Distributed Parallel Databases34(4) (2016), 463–503. doi:10.1007/s10619-015-7185-y.
60.
R.Taelman, J.V.Herwegen, M.V.Sande and R.Verborgh, Comunica: A modular SPARQL query engine for the web, in: The Semantic Web – ISWC 2018 – 17th International Semantic Web Conference, Proceedings, Part II, Monterey, CA, USA, October 8–12, 2018, Lecture Notes in Computer Science, Vol. 11137, Springer, 2018, pp. 239–255. doi:10.1007/978-3-030-00668-6_15.
61.
M.Saleem, A.Potocki, T.Soru, O.Hartig and A.-C.Ngonga Ngomo, CostFed: Cost-based query optimization for SPARQL endpoint federation, in: Proc. of Int. Conf. on Semantic Systems (SEMANTICS), Procedia Computer Science, Vol. 137, Elsevier, 2018, pp. 163–174. doi:10.1016/j.procs.2018.09.016.
Dremio, Accessed 17 November 2021, https://www.dremio.com/.
64.
G.Montoya, H.Skaf-Molli, P.Molli and M.-E.Vidal, Federated SPARQL queries processing with replicated fragments, in: Proc. of Int. Semantic Web Conf (ISWC), LNCS, Vol. 9366, Springer, 2015, pp. 36–51. doi:10.1007/978-3-319-25007-6_3.
65.
GraphDB, Accessed 17 November 2021, https://graphdb.ontotext.com/.
66.
M.Saleem and A.-C.Ngonga Ngomo, HiBISCuS: Hypergraph-based source selection for SPARQL endpoint federation, in: Proc. of European Semantic Web Conf (ESWC), LNCS, Vol. 8465, Springer, 2014, pp. 176–191. doi:10.1007/978-3-319-07443-6_13.
67.
IBM Cloud Pak for Data, Accessed 17 November 2021, https://www.ibm.com/products/cloud-pak-for-data.
68.
IBM Db2 Big SQL, Accessed 18 November 2021, https://www.ibm.com/products/db2-big-sql.
69.
IBM InfoSphere Federation Server, Accessed 18 November 2021, https://www.ibm.com/docs/en/iis/11.7?topic=components-infosphere-federation-server.
70.
JBoss Data Virtualization, Accessed 17 November 2021, https://developers.redhat.com/products/datavirt/overview.
71.
Metaphactory, Accessed 18 November 2021, https://metaphacts.com/product.
72.
P.Haase, D.M.Herzig, A.Kozlov, A.Nikolov and J.Trame, metaphactory: A platform for knowledge graph management, Semantic Web10(6) (2019), 1109–1125. doi:10.3233/SW-190360.
73.
Neo4j, Accessed 17 November 2021, https://neo4j.com/.
74.
N.Francis, A.Green, P.Guagliardo, L.Libkin, T.Lindaaker, V.Marsault, S.Plantikow, M.Rydberg, P.Selmer and A.Taylor, Cypher: An evolving query language for property graphs, in: Proc. of ACM SIGMOD Int. Conf. on Management of Data (SIGMOD), ACM, 2018, pp. 1433–1445. doi:10.1145/3183713.3190657.
75.
M.Buron, F.Goasdoué, I.Manolescu and M.-L.Mugnier, Ontology-based RDF integration of heterogeneous data, in: Proc. of 23rd Int. Conf. on Extending Database Technology (EDBT), OpenProceedings.org, 2020, pp. 299–310. doi:10.5441/002/edbt.2020.27.
76.
M.Buron, F.Goasdoué, I.Manolescu and M.-L.Mugnier, Obi-Wan: Ontology-based RDF integration of heterogeneous data, Proc. of VLDB Endowment13(12) (2020), 2933–2936. doi:10.14778/3415478.3415512.
77.
G.Xiao, D.Calvanese, R.Kontchakov, D.Lembo, A.Poggi, R.Rosati and M.Zakharyaschev, Ontology-based data access: A survey, in: Proc. of 27th Int. Joint Conf. on Artificial Intelligence (IJCAI), ijcai.org, 2018, pp. 5511–5519. doi:10.24963/ijcai.2018/777.
78.
R.Bonaque, T.D.Cao, B.Cautis, F.Goasdoué, J.Letelier, I.Manolescu, O.Mendoza, S.Ribeiro, X.Tannier and M.Thomazo, Mixed-instance querying: A lightweight integration architecture for data journalism, Proc. of VLDB Endowment9(13) (2016), 1513–1516. doi:10.14778/3007263.3007297.
79.
G.Montoya, H.Skaf-Molli and K.Hose, The odyssey approach for optimizing federated SPARQL queries, in: Proc. of Int. Semantic Web Conf (ISWC), LNCS, Vol. 10587, Springer, 2017, pp. 471–489. doi:10.1007/978-3-319-68288-4_28.
80.
K.M.Endris, M.Galkin, I.Lytra, M.N.Mami, M.-E.Vidal and S.Auer, Querying interlinked data by bridging RDF molecule templates, Trans. Large Scale Data Knowledge Centered Systems39 (2018), 1–42. doi:10.1007/978-3-662-58415-6_1.
81.
M.Masmoudi, S.B.A.B.Lamine, H.B.Zghal, B.Archimède and M.-H.Karray, Knowledge hypergraph-based approach for data integration and querying: Application to Earth observation, Future Generation Computer Systems115 (2021), 720–740. doi:10.1016/j.future.2020.09.029.
82.
Oracle Big Data SQL, Accessed 18 November 2021, https://www.oracle.com/database/technologies/datawarehouse-bigdata/bigdata-sql.html.
83.
Oracle Spatial and Graph, Accessed 16 November 2021, https://www.oracle.com/database/technologies/spatialandgraph.html.
Y.Khan, A.Zimmermann, A.Jha, D.Rebholz-Schuhmann and R.Sahay, Querying web polystores, in: Proc. of IEEE Int. Conf. on Big Data (IEEE BigData), IEEE Computer Society, 2017, pp. 3190–3195. doi:10.1109/BigData.2017.8258299.
86.
R.Sethi, M.Traverso, D.Sundstrom, D.Phillips, W.Xie, Y.Sun, N.Yegitbasi, H.Jin, E.Hwang, N.Shingte and C.Berner, Presto: SQL on everything, in: Proc. of 35th Int. Conf. on Data Engineering (ICDE), IEEE, 2019, pp. 1802–1813. doi:10.1109/ICDE.2019.00196.
87.
Querona Data Virtualization, Accessed 17 November 2021, https://www.querona.io/.
88.
RDFLib, Accessed 26 July 2022, https://rdflib.readthedocs.io/en/stable/.
89.
Y.Khan, M.Saleem, M.Mehdi, A.Hogan, Q.Mehmood, D.Rebholz-Schuhmann and R.Sahay, SAFE: SPARQL federation over RDF data cubes with access control, J. Biomedical Semantics8(1) (2017), 5–1522. doi:10.1186/s13326-017-0112-6.
90.
T.Minier, H.Skaf-Molli and P.Molli, SaGe: Web preemption for public SPARQL query services, in: The World Wide Web Conference, WWW 2019, San Francisco, CA, USA, May 13–17, 2019, L.Liu, R.W.White, A.Mantrach, F.Silvestri, J.J.McAuley, R.Baeza-Yates and L.Zia, eds, ACM, 2019, pp. 1268–1278. doi:10.1145/3308558.3313652.
91.
SAP HANA (Smart Data Access), Accessed 18 November 2021, https://help.sap.com/docs/SAP_HANA_PLATFORM/6b94445c94ae495c83a19646e7c3fd56/a07c7ff25997460bbcb73099fb59007d.html?locale=en-US&version=2.0.05.
92.
SAS Federation Server, Accessed 18 November 2021, https://support.sas.com/en/software/federation-server-support.html.
93.
A.Charalambidis, A.Troumpoukis and S.Konstantopoulos, SemaGrow: Optimizing federated SPARQL queries, in: Proc. of 11th Int. Conf. on Semantic Systems (SEMANTICS), ACM, 2015, pp. 121–128. doi:10.1145/2814864.2814886.
94.
K.Alexander, R.Cyganiak, M.Hausenblas and J.Zhao, Describing linked datasets, in: Proc. of Int. Workshop on Linked Data on the Web (LDOW), CEUR Workshop Proceedings, Vol. 538, CEUR-WS.org, 2009.
95.
SQL Server (PolyBase), Accessed 18 November 2021, https://docs.microsoft.com/en-us/sql/relational-databases/polybase/polybase-guide?view=sql-server-ver15.
96.
M.N.Mami, D.Graux, S.Scerri, H.Jabeen, S.Auer and J.Lehmann, squerall: Virtual ontology-based access to heterogeneous and large data sources, in: Proc. of Int. Semantic Web Conf (ISWC), LNCS, Vol. 11779, Springer, 2019, pp. 229–245. doi:10.1007/978-3-030-30796-7_15.
97.
Starburst, Accessed 18 December 2021, https://www.starburst.io/.
98.
Stardog, Accessed 17 November 2021, https://www.stardog.com/.
99.
TIBCO Data Virtualization, Accessed 17 November 2021, https://www.tibco.com/products/data-virtualization.
100.
Trino, Accessed 18 November 2021, https://trino.io/.
101.
Virtuoso, Accessed 17 November 2021, https://virtuoso.openlinksw.com/.
102.
O.Erling, Virtuoso, a hybrid RDBMS/graph column store, IEEE Data Engineering Bull.35(1) (2012), 3–8.
103.
O.Erling and I.Mikhailov, RDF support in the virtuoso DBMS, in: Proc. of Conf. on Social Semantic Web (CSSW), LNI, Vol. P-113, GI, 2007, pp. 59–68.
104.
P.J.Sadalage and M.Fowler, NoSQL Distilled: A Brief Guide to the Emerging World of Polyglot Persistence, Pearson Education, 2013.
105.
DB-Engines, Accessed 16 February 2022, https://db-engines.com/en/.
106.
Database of Databases, Accessed 16 February 2022, https://dbdb.io/.
107.
S.Konstantopoulos, A.Charalambidis, A.Troumpoukis, G.Mouchakis and V.Karkaletsis, The Sevod vocabulary for dataset descriptions for federated querying, in: Proceedings of the 4th International Workshop on Dataset PROFIling and fEderated Search for Web Data (PROFILES 2017) co-located with The 16th International Semantic Web Conference (ISWC 2017), Vienna, Austria, October 22, 2017, CEUR Workshop Proceedings, Vol. 1927, CEUR-WS.org, 2017.
108.
H.Stuckenschmidt, R.Vdovjak, G.-J.Houben and J.Broekstra, Index structures and algorithms for querying distributed RDF repositories, in: Proceedings of the 13th International Conference on World Wide Web, WWW’04, Association for Computing Machinery, New York, NY, USA, 2004, pp. 631–639. ISBN 158113844X. doi:10.1145/988672.988758.
109.
M.T.Özsu and P.Valduriez, Principles of Distributed Database Systems, 4th edn, Springer, 2020. ISBN 978-3-030-26252-5. doi:10.1007/978-3-030-26253-2.
110.
L.Haas, D.Kossmann, E.Wimmers and J.Yang, Optimizing queries across diverse data sources, in: 23rd International Conference on Very Large Data Bases (VLDB 1997), 1997.
111.
Data Virtuality (Push down), Accessed 17 November 2021, https://documentation.datavirtuality.com/24/reference-guide/federated-planning/federated-optimizations#FederatedOptimizations-Pushdown.
112.
IBM Db2 Big SQL: What’s new and changed in Data Virtualization, Accessed 18 November 2021, https://www.ibm.com/docs/en/cloud-paks/cp-data/4.5.x?topic=new-data-virtualization.
113.
SQL Server (PolyBase Push down), Accessed 18 November 2021, https://docs.microsoft.com/en-us/sql/relational-databases/polybase/polybase-pushdown-computation?view=sql-server-ver15.
114.
Starburst Push down, Accessed 18 December 2021, https://docs.starburst.io/latest/optimizer/pushdown.html.
115.
Trino (Pushdown), Accessed 18 November 2021, https://trino.io/docs/current/optimizer/pushdown.html.
116.
A.Silberschatz, H.F.Korth and S.Sudarshan, Database System Concepts, 7th edn, McGraw-Hill Book Company, 2020. ISBN 9780078022159.
117.
Teiid (Planning Overview), Accessed 16 November 2021, http://teiid.github.io/teiid-documents/16.0.x/content/reference/r_planning-overview.html.
118.
SAP HANA (SQL Optimizer), Accessed 18 November 2021, https://help.sap.com/docs/SAP_HANA_PLATFORM/9de0171a6027400bb3b9bee385222eff/d2948cc2209a407ea2b686c29e72ca50.html.
119.
S.Blanas, J.M.Patel, V.Ercegovac, J.Rao, E.J.Shekita and Y.Tian, A comparison of join algorithms for log processing in MaPreduce, in: Proceedings of the ACM SIGMOD International Conference on Management of Data, SIGMOD 2010, Indianapolis, Indiana, USA, June 6–10, 2010, ACM, 2010, pp. 975–986. doi:10.1145/1807167.1807273.
120.
Apache Drill (Broadcast Join), Accessed 18 November 2021, https://drill.apache.org/docs/join-planning-guidelines/.
121.
Starburst (Broadcast Join), Accessed 18 December 2021, https://docs.starburst.io/latest/admin/properties-general.html.
122.
A.M.Gupta, V.Gadepally and M.Stonebraker, Cross-engine query execution in federated database systems, in: 2016 IEEE High Performance Extreme Computing Conference, HPEC 2016, Waltham, MA, USA, September 13–15, 2016, IEEE, 2016, pp. 1–6. doi:10.1109/HPEC.2016.7761648.
123.
Presto (Choosing Presto Join and Sort Algorithms), Accessed 18 November 2021. https://docs.treasuredata.com/display/public/PD/Choosing+Presto+Join+and+Sort+Algorithms.
124.
Trino, Accessed 18 November 2021, https://trino.io/docs/current/optimizer/cost-based-optimizations.html#join-distribution-selection.
125.
Data Federation (Data Movement), Accessed 12 March 2021, https://community.denodo.com/docs/html/browse/8.0/vdp/administration/optimizing_queries/data_movement/data_movement.
126.
C.Bizer, T.Heath and T.Berners-Lee, Linked data – the story so far, Int. J. Semantic Web and Information Systems5(3) (2009), 1–22. doi:10.4018/jswis.2009081901.
127.
ISO/IEC JTC 1/SC 32 Data management and interchange, ISO/IEC 9075-9:2016 – Information technology – Database languages – SQL – Part 9: Management of External Data (SQL/MED), Technical Report, ISO/IEC, 2016.
128.
D.Kontokostas and H.Knublauch, Shapes Constraint Language (SHACL), W3C Recommendation, W3C, 2017, https://www.w3.org/TR/2017/REC-shacl-20170720/.
129.
A.Poggi, D.Lembo, D.Calvanese, G.De Giacomo, M.Lenzerini and R.Rosati, Linking data to ontologies, J. Data Semantics10 (2008), 133–173. doi:10.1007/978-3-540-77688-8_5.
130.
G.Xiao, L.Ding, B.Cogrel and D.Calvanese, Virtual knowledge graphs: An overview of systems and use cases, Data Intelligence1(3) (2019), 201–223. doi:10.1162/dint_a_00011.
131.
C.Civili, M.Console, G.De Giacomo, D.Lembo, M.Lenzerini, L.Lepore, R.Mancini, A.Poggi, R.Rosati, M.Ruzzi, V.Santarelli and D.F.Savo, MASTRO STUDIO: managing ontology-based data access applications, Proc. of VLDB Endowment6(12) (2013), 1314–1317. doi:10.14778/2536274.2536304.
132.
D.Lanti, G.Xiao and D.Calvanese, Cost-driven ontology-based data access, in: Proc. of Int. Semantic Web Conf. (ISWC), LNCS, Vol. 10587, Springer, 2017, pp. 452–470. doi:10.1007/978-3-319-68288-4_27.
G.Xiao, D.Lanti, R.Kontchakov, S.Komla-Ebri, E.G.Kalayci, L.Ding, J.Corman, B.Cogrel, D.Calvanese and E.Botoeva, The virtual knowledge graph system ontop, in: Proc. of Int. Semantic Web Conf. (ISWC), LNCS, Vol. 12507, Springer, 2020, pp. 259–277. doi:10.1007/978-3-030-62466-8_17.
135.
S.Das, R.Cyganiak and S.Sundara, R2RML: RDB to RDF Mapping Language, W3C, Recommendation W3C (2012), http://www.w3.org/TR/2012/REC-r2rml-20120927/.
136.
Z.Gu, D.Lanti, A.Mosca, G.Xiao, J.Xiong and D.Calvanese, Ontology-based data federation, in: Proc. of the 35th Int. Workshop on Description Logics (DL), CEUR Workshop Proceedings, 2022, To appear in proceedings.
137.
Z.Gu, D.Lanti, A.Mosca, G.Xiao, J.Xiong and D.Calvanese, Ontology-based data federation, in: The 11th International Joint Conference on Knowledge Graphs (IJCKG 2022), 2022, To appear in print.
138.
Z.She, S.Ravishankar and J.Duggan, BigDAWG polystore query optimization through semantic equivalences, in: 2016 IEEE High Performance Extreme Computing Conference, HPEC 2016, Waltham, MA, USA, September 13–15, 2016, IEEE, 2016, pp. 1–6. doi:10.1109/HPEC.2016.7761584.
139.
M.Saleem, A.N.Ngomo, J.X.Parreira, H.F.Deus and M.Hauswirth, DAW: Duplicate-AWare federated query processing over the web of data, in: The Semantic Web – ISWC 2013 – 12th International Semantic Web Conference, Proceedings, Part I, Sydney, NSW, Australia, October 21–25, 2013, Lecture Notes in Computer Science, Vol. 8218, Springer, 2013, pp. 574–590. doi:10.1007/978-3-642-41335-3_36.
140.
A.C.Sima, T.M.de Farias, E.Zbinden, M.Anisimova, M.Gil, H.Stockinger, K.Stockinger, M.Robinson-Rechavi and C.Dessimoz, Enabling semantic queries across federated bioinformatics databases, Database J. Biol. Databases Curation2019 (2019), 106. doi:10.1093/database/baz106.
141.
M.Magnani and D.Montesi, A survey on uncertainty management in data integration, J. Data Information Quality2(1) (2010), 5–1533. doi:10.1145/1805286.1805291.
142.
N.Bikakis, C.Tsinaraki, N.Gioldasis, I.Stavrakantonakis and S.Christodoulakis, The XML and Semantic Web Worlds: Technologies, Interoperability and Integration: A Survey of the State of the Art, Semantic Hyper/Multimedia Adaptation – Schemes and Applications, SCI, Vol. 418, Springer, 2013, pp. 319–360. doi:10.1007/978-3-642-28977-4_12.
143.
B.Arputhamary and L.Arockiam, A review on big data integration, Int. J. Computer Applications22(3) (2015), 21–26.
144.
X.L.Dong and D.Srivastava, Big Data Integration, Synthesis Lectures on Data Management, Morgan & Claypool Publishers, 2015. doi:10.2200/S00578ED1V01Y201404DTM040.
145.
J.Hui, L.Li and Z.Zhang, Integration of big data: A survey, in: Proc. of 4th Int. Conf. or Pioneering Computer Scientists, Engineers and Educators (ICPCSEE), CCIS, Vol. 901, Springer, 2018, pp. 101–121. doi:10.1007/978-981-13-2203-7_9.
146.
C.Bondiombouy and P.Valduriez, Query processing in multistore systems: An overview, Int. J. Cloud Computing5(4) (2016), 309–346. doi:10.1504/IJCC.2016.080903.
147.
H.Wache, T.Vögele, U.Visser, H.Stuckenschmidt, G.Schuster, H.Neumann and S.Hübner, Ontology-based integration of information – a survey of existing approaches, in: Proc. of Workshop on Ontologies and Information Sharing, CEUR Workshop Proceedings, Vol. 47, CEUR-WS.org, 2001.
148.
N.F.Noy, Semantic integration: A survey of ontology-based approaches, SIGMOD Record33(4) (2004), 65–70. doi:10.1145/1041410.1041421.
149.
F.J.Ekaputra, M.Sabou, E.Serral, E.Kiesling and S.Biffl, Ontology-based data integration in multi-disciplinary engineering environments: A review, Open J. Information Systems4(1) (2017), 1–26. doi:10.24167/sisforma.v4i1.1040.
150.
A.Buccella, A.Cechich and P.R.Fillottrani, Ontology-driven geographic information integration: A survey of current approaches, Computers and Geosciences35(4) (2009), 710–723. doi:10.1016/j.cageo.2008.02.033.
151.
B.Hassan, R.Fissoune and C.Messaoudi, A survey of semantic integration approaches in bioinformatics, Int. J. Computer, Electrical, Automation, Control and Information Engineering10(12) (2016), 1968–1973.
152.
M.Mountantonakis and Y.Tzitzikas, Large-scale semantic integration of linked data: A survey, ACM Computing Surveys52(5) (2019), 103–110340. doi:10.1145/3345551.
153.
M.Schmidt, O.Görlitz, P.Haase, G.Ladwig, A.Schwarte and T.Tran, FedBench: A benchmark suite for federated semantic data query processing, in: Proc. of Int. Semantic Web Conf. (ISWC), LNCS, Vol. 7031, Springer, 2011, pp. 585–600. doi:10.1007/978-3-642-25073-6_37.
154.
M.Saleem, A.Hasnain and A.-C.Ngonga Ngomo, LargeRDFBench: A billion triples benchmark for SPARQL endpoint federation, J. Web Semantics48 (2018), 85–125. doi:10.1016/j.websem.2017.12.005.
155.
N.A.Rakhmawati, Evaluating and benchmarking the performance of federated SPARQL endpoints and their partitioning using selected metrics and specific query types, PhD thesis, National University of Ireland, Galway, 2017.
156.
N.A.Rakhmawati, An holistic evaluation of federated SPARQL query engine, in: Proc. of Information Systems International Conference (ISICO), 2013.
157.
N.A.Rakhmawati, M.Saleem, S.Lalithsena and S.Decker, QFed: Query set for federated SPARQL query benchmark, in: Proceedings of the 16th International Conference on Information Integration and Web-Based Applications & Services, Hanoi, Vietnam, December 4–6, 2014, ACM, 2014, pp. 207–211. doi:10.1145/2684200.2684321.
158.
N.A.Rakhmawati, M.Karnstedt, M.Hausenblas and S.Decker, On metrics for measuring fragmentation of federation over SPARQL endpoints, in: WEBIST 2014 – Proceedings of the 10th International Conference on Web Information Systems and Technologies, Vol. 1, Barcelona, Spain, 3–5 April, 2014, SciTePress, 2014, pp. 119–126. doi:10.5220/0004760101190126.
159.
O.Görlitz, Distributed query processing for federated RDF data management, PhD thesis, University of Koblenz-Landau, 2015.
160.
M.Saleem, Efficient source selection and benchmarking for SPARQL endpoint query federation, PhD thesis, Leipzig University, Germany, 2018. ISBN 978-3-89838-732-3.
161.
S.M.A.Hasnain, Cataloguing and linking publicly available biomedical SPARQL endpoints for federation-addressing aPosteriori data integration, PhD thesis, National University of IReland, Galway, 2017.
162.
A.Hasnain, M.Saleem, A.N.Ngomo and D.Rebholz-Schuhmann, Extending LargeRDFBench for multi-source data at scale for SPARQL endpoint federation, in: Emerging Topics in Semantic Technologies – ISWC 2018 Satellite Events [Best Papers from 13 of the Workshops Co-Located with the ISWC 2018 Conference], Studies on the Semantic Web, Vol. 36, IOS Press, 2018, pp. 203–218. doi:10.3233/978-1-61499-894-5-203.
163.
K.M.Endris, Federated Query Processing over Heterogeneous Data Sources in a Semantic Data Lake, PhD thesis, University of Bonn, Germany, 2020.
164.
A.Valdestilhas, Identifying, Relating, Consisting and Querying Large Heterogeneous RDF Sources, PhD thesis, Leipzig University, Germany, 2021.
165.
H.Stuckenschmidt, R.Vdovjak, J.Broekstra and G.Houben, Towards distributed processing of RDF path queries, Int. J. Web Eng. Technol.2(2/3) (2005), 207–230. doi:10.1504/IJWET.2005.008484.
166.
J.Zemánek and S.Schenk, Optimizing SPARQL queries over disparate RDF data sources through distributed semi-joins, in: Proceedings of the Poster and Demonstration Session at the 7th International Semantic Web Conference (ISWC2008), Karlsruhe, Germany, October 28, 2008, CEUR Workshop Proceedings, Vol. 401, CEUR-WS.org, 2008.
167.
S.Schenk and S.Staab, Networked graphs: A declarative mechanism for SPARQL rules, SPARQL views and RDF data integration on the web, in: Proceedings of the 17th International Conference on World Wide Web, WWW 2008, Beijing, China, April 21–25, 2008, pp. 585–594, ACM, 2008. doi:10.1145/1367497.1367577.
168.
A.Langegger, W.Wöß and M.Blöchl, A semantic web middleware for virtual data integration on the web, in: The Semantic Web: Research and Applications, 5th European Semantic Web Conference, ESWC 2008, Proceedings, Tenerife, Canary Islands, Spain, June 1–5, 2008, Lecture Notes in Computer Science, Vol. 5021, Springer, 2008, pp. 493–507. doi:10.1007/978-3-540-68234-9_37.
169.
K.Cheung, H.R.Frost, M.S.Marshall, E.Prud’hommeaux, M.Samwald, J.Zhao and A.Paschke, A journey to semantic web query federation in the life sciences, BMC Bioinform.10(S–10) (2009), 10. doi:10.1186/1471-2105-10-S10-S10.
170.
Z.Kaoudi, M.Koubarakis, K.Kyzirakos, I.Miliaraki, M.Magiridou and A.Papadakis-Pesaresi, Atlas: Storing, updating and querying RDF(S) data on top of DHTs, J. Web Semant.8(4) (2010), 271–277. doi:10.1016/j.websem.2010.07.001.
171.
A.Harth, K.Hose, M.Karnstedt, A.Polleres, K.Sattler and J.Umbrich, Data summaries for on-demand queries over linked data, in: Proceedings of the 19th International Conference on World Wide Web, WWW 2010, Raleigh, North Carolina, USA, April 26–30, 2010, M.Rappa, P.Jones, J.Freire and S.Chakrabarti, eds, ACM, 2010, pp. 411–420. doi:10.1145/1772690.1772733.
172.
G.Ladwig and T.Tran, Linked data query processing strategies, in: The Semantic Web – ISWC 2010 – 9th International Semantic Web Conference, ISWC 2010, Revised Selected Papers, Part I, Shanghai, China, November 7–11, 2010, Lecture Notes in Computer Science, Vol. 6496, Springer, 2010, pp. 453–469. doi:10.1007/978-3-642-17746-0_29.
173.
M.Acosta, M.-E.Vidal, T.Lampo, J.Castillo and E.Ruckhaus, ANAPSID: An adaptive query processing engine for SPARQL endpoints, in: Proc. of Int. Semantic Web Conference, LNCS, Vol. 7031, Springer, 2011, pp. 18–34. doi:10.1007/978-3-642-25073-6_2.
174.
S.Lynden, I.Kojima, A.Matono and Y.Tanimura, Aderis: An adaptive query processor for joining federated sparql endpoints, in: OTM Confederated International Conferences “On the Move to Meaningful Internet Systems”, 2011, pp. 808–817.
175.
X.Wang, T.Tiropanis and H.C.Davis, Querying the Web of Data with Graph Theory-based Techniques, Technical Report, University of Southampton, 2011.
176.
X.Wang, T.Tiropanis and H.C.Davis, Evaluating graph traversal algorithms for distributed SPARQL query optimization, in: The Semantic Web – Joint International Semantic Technology Conference, JIST 2011, Proceedings, Hangzhou, China, December 4–7, 2011, J.Z.Pan, H.Chen, H.Kim, J.Li, Z.Wu, I.Horrocks, R.Mizoguchi and Z.Wu, eds, Lecture Notes in Computer Science, Vol. 7185, Springer, 2011, pp. 210–225. doi:10.1007/978-3-642-29923-0_14.
177.
G.Ladwig and T.Tran, SIHJoin: Querying remote and local linked data, in: The Semantic Web: Research and Applications – 8th Extended Semantic Web Conference, ESWC 2011, Proceedings, Part I, Heraklion, Crete, Greece, May 29–June 2, 2011, Lecture Notes in Computer Science, Vol. 6643, Springer, 2011, pp. 139–153. doi:10.1007/978-3-642-21034-1_10.
178.
C.B.Aranda, M.Arenas and Ó.Corcho, Semantics and optimization of the SPARQL 1.1 federation extension, in: The Semanic Web: Research and Applications – 8th Extended Semantic Web Conference, Proceedings, Part II, ESWC 2011, Heraklion, Crete, Greece, May 29–June 2, 2011, Lecture Notes in Computer Science, Vol. 6644, Springer, 2011, pp. 1–15. doi:10.1007/978-3-642-21064-8_1.
179.
F.Prasser, A.Kemper and K.A.Kuhn, Efficient distributed query processing for autonomous RDF databases, in: 15th International Conference on Extending Database Technology, EDBT’12, Proceedings, Berlin, Germany, March 27–30, 2012, ACM, 2012, pp. 372–383. doi:10.1145/2247596.2247640.
Z.Akar, T.G.Halaç, E.E.Ekinci and O.Dikenelli, Querying the web of interlinked datasets using VOID descriptions, in: WWW2012 Workshop on Linked Data on the Web, Lyon, France, 16 April, 2012, CEUR Workshop Proceedings, Vol. 937, CEUR-WS.org, 2012.
182.
A.Hasnain, M.R.Kamdar, P.Hasapis, D.Zeginis, C.N.Warren Jr., H.F.Deus, D.Ntalaperas, K.A.Tarabanis, M.Mehdi and S.Decker, Linked biomedical dataspace: Lessons learned integrating data for drug discovery, in: The Semantic Web – ISWC 2014 – 13th International Semantic Web Conference, Proceedings, Part I, Riva del Garda, Italy, October 19–23, 2014, P.Mika, T.Tudorache, A.Bernstein, C.Welty, C.A.Knoblock, D.Vrandecic, P.Groth, N.F.Noy, K.Janowicz and C.A.Goble, eds, Lecture Notes in Computer Science, Vol. 8796, Springer, 2014, p. 114–130. doi:10.1007/978-3-319-11964-9_8.
183.
A.Hasnain, S.S.e Zainab, M.R.Kamdar, Q.Mehmood, C.N.Warren Jr., Q.A.Fatimah, H.F.Deus, M.Mehdi and S.Decker, A roadmap for navigating the life sciences linked open data cloud, in: Semantic Technology – 4th Joint International Conference, JIST 2014, Chiang Mai, Thailand, November 9-11, 2014. Revised Selected Papers, T.Supnithi, T.Yamaguchi, J.Z.Pan, V.Wuwongse and M.Buranarach, eds, Lecture Notes in Computer Science, Vol. 8943, Springer, 2014, pp. 97–112. doi:10.1007/978-3-319-15615-6_8.
184.
A.Hasnain, R.Fox, S.Decker and H.F.Deus, Cataloguing and linking life sciences LOD cloud, in: Proc. of 1st Int. Workshop on Ontology Engineering in a Data-Driven World (OEDW), Co-Located with EKAW, 2012.
185.
X.Wang, T.Tiropanis and H.C.Davis, LHD: Optimising linked data query processing using parallelisation, in: Proceedings of the WWW2013 Workshop on Linked Data on the Web, Rio de Janeiro, Brazil, 14 May, 2013, CEUR Workshop Proceedings, Vol. 996, CEUR-WS.org, 2013.
186.
A.Nikolov, A.Schwarte and C.Hütter, FedSearch: Efficiently combining structured queries and full-text search in a SPARQL federation, in: The Semantic Web – ISWC 2013 – 12th International Semantic Web Conference, Proceedings, Part I, Sydney, NSW, Australia, October 21–25, 2013, Lecture Notes in Computer Science, Vol. 8218, Springer, 2013, pp. 427–443. doi:10.1007/978-3-642-41335-3_27.
187.
O.Hartig, SQUIN: A traversal based query execution system for the web of linked data, in: Proceedings of the ACM SIGMOD International Conference on Management of Data, SIGMOD 2013, New York, NY, USA, June 22–27, 2013, ACM, 2013, pp. 1081–1084. doi:10.1145/2463676.2465231.
188.
C.Basca and A.Bernstein, Avalanche: Putting the spirit of the web back into semantic web querying, in: Proceedings of the ISWC 2010 Posters & Demonstrations Track: Collected Abstracts, Shanghai, China, November 9, 2010, CEUR Workshop Proceedings, Vol. 658, CEUR-WS.org, 2010.
189.
C.Basca and A.Bernstein, Querying a messy web of data with Avalanche, J. Web Semant.26 (2014), 1–28. doi:10.1016/j.websem.2014.04.002.
190.
D.R.B.Cunha and B.F.Lóscio, An approach for query decomposition on federated SPARQL query systems, J. Inf. Data Manag.6(2) (2015), 106–117.
191.
B.Kolev, C.Bondiombouy, O.Levchenko, P.Valduriez, R.Jiménez-Peris, R.Pau and J.Pereira, Design and implementation of the CloudMdsQL multistore system, in: CLOSER 2016 – Proceedings of the 6th International Conference on Cloud Computing and Services Science, Vol. 1, Rome, Italy, April 23–25, 2016, SciTePress, 2016, pp. 352–359. doi:10.5220/0005923803520359.
192.
A.J.Elmore, J.Duggan, M.Stonebraker, M.Balazinska, U.Çetintemel, V.Gadepally, J.Heer, B.Howe, J.Kepner, T.Kraska, S.Madden, D.Maier, T.G.Mattson, S.Papadopoulos, J.Parkhurst, N.Tatbul, M.Vartak and S.Zdonik, A demonstration of the BigDAWG polystore system, Proc. VLDB Endow.8(12) (2015), 1908–1911. doi:10.14778/2824032.2824098.
193.
P.Fafalios, T.Yannakis and Y.Tzitzikas, Querying the web of data with SPARQL-LD, in: Research and Advanced Technology for Digital Libraries – 20th International Conference on Theory and Practice of Digital Libraries, TPDL 2016, Proceedings, Hannover, Germany, September 5–9, 2016, Lecture Notes in Computer Science, Vol. 9819, Springer, 2016, pp. 175–187. doi:10.1007/978-3-319-43997-6_14.
194.
P.Fafalios and Y.Tzitzikas, SPARQL-LD: A SPARQL extension for fetching and querying linked data, in: Proceedings of the ISWC 2015 Posters & Demonstrations Track Co-Located with the 14th International Semantic Web Conference (ISWC-2015), Bethlehem, PA, USA, October 11, 2015, CEUR Workshop Proceedings, Vol. 1486, CEUR-WS.org, 2015.
195.
T.Yannakis, P.Fafalios and Y.Tzitzikas, Heuristics-based query reordering for federated queries in SPARQL 1.1 and SPARQL-LD, in: Proceedings of the 2nd Workshop on Querying the Web of Data Co-Located with 15th Extended Semantic Web Conference (ESWC 2018), Heraklion, Greece, June 3, 2018, CEUR Workshop Proceedings, Vol. 2110, CEUR-WS.org, 2018, pp. 74–88.
196.
D.Collarana, C.Lange and S.Auer, FuhSen: A platform for federated, RDF-based hybrid search, in: Proceedings of the 25th International Conference on World Wide Web, WWW 2016, Companion Volume, Montreal, Canada, April 11–15, 2016, ACM, 2016, pp. 171–174. doi:10.1145/2872518.2890535.
197.
Y.Khan, M.Saleem, A.Iqbal, M.Mehdi, A.Hogan, A.N.Ngomo, S.Decker and R.Sahay, SAFE: Policy aware SPARQL query federation over RDF data cubes, in: Proceedings of the 7th International Workshop on Semantic Web Applications and Tools for Life Sciences, Berlin, Germany, December 9–11, 2014, CEUR Workshop Proceedings, Vol. 1320, CEUR-WS.org, 2014.
198.
I.Abdelaziz, E.Mansour, M.Ouzzani, A.Aboulnaga and P.Kalnis, Lusail: A system for querying linked data at scale, Proc. of VLDB Endowment11(4) (2017), 485–498. doi:10.1145/3186728.3164144.
199.
E.Mansour, I.Abdelaziz, M.Ouzzani, A.Aboulnaga and P.Kalnis, A demonstration of lusail: Querying linked data at scale, in: Proceedings of the 2017 ACM International Conference on Management of Data, SIGMOD, Conference 2017, Chicago, IL, USA, May 14–19, 2017, ACM, 2017, pp. 1603–1606. doi:10.1145/3035918.3058731.
200.
I.Abdelaziz, E.Mansour, M.Ouzzani, A.Aboulnaga and P.Kalnis, Query optimizations over decentralized RDF graphs, in: 33rd IEEE International Conference on Data Engineering, ICDE 2017, San Diego, CA, USA, April 19–22, 2017, IEEE Computer Society, 2017, 139–142. doi:10.1109/ICDE.2017.59.
201.
D.Halperin, V.T.de Almeida, L.L.Choo, S.Chu, P.Koutris, D.Moritz, J.Ortiz, V.Ruamviboonsuk, J.Wang, A.Whitaker, S.Xu, M.Balazinska, B.Howe and D.Suciu, Demonstration of the Myria big data management service, in: International Conference on Management of Data, SIGMOD 2014 (Demonstrations), Snowbird, UT, USA, June 22–27, 2014, ACM, 2014, pp. 881–884. doi:10.1145/2588555.2594530.
202.
A.Hasnain, Q.Mehmood, S.S.e Zainab, M.Saleem, C.N.Warren Jr., D.Zehra, S.Decker and D.Rebholz-Schuhmann, BioFed: Federated query processing over life sciences linked open data, J. Biomedical Semantics8(1) (2017), 13–11319. doi:10.1186/s13326-017-0118-0.
203.
R.Verborgh, M.V.Sande, O.Hartig, J.V.Herwegen, L.D.Vocht, B.D.Meester, G.Haesendonck and P.Colpaert, Triple pattern fragments: A low-cost knowledge graph interface for the web, J. Web Semant.37–38 (2016), 184–206. doi:10.1016/j.websem.2016.03.003.
204.
A.Potocki, M.Saleem, T.Soru, O.Hartig, M.Voigt and A.N.Ngomo, Federated SPARQL query processing via CostFed, in: Proceedings of the ISWC 2017 Posters & Demonstrations and Industry Tracks Co-Located with 16th International Semantic Web Conference (ISWC 2017), Vienna, Austria, October 23rd–25th, 2017, CEUR Workshop Proceedings, Vol. 1963, CEUR-WS.org, 2017.
205.
K.M.Endris, Z.Almhithawi, I.Lytra, M.Vidal and S.Auer, BOUNCER: Privacy-aware query processing over federations of RDF datasets, in: Database and Expert Systems Applications – 29th International Conference, DEXA 2018, Proceedings, Part I, Regensburg, Germany, September 3–6, 2018, Lecture Notes in Computer Science, Vol. 11029, Springer, 2018, pp. 69–84. doi:10.1007/978-3-319-98809-2_5.
206.
F.Michel, C.Faron-Zucker and F.Gandon, SPARQL micro-services: Lightweight integration of web APIs and linked data, in: Workshop on Linked Data on the Web Co-Located with the Web Conference 2018, LDOW@WWW 2018, Lyon, France, April 23rd, 2018, CEUR Workshop Proceedings, Vol. 2073, CEUR-WS.org, 2018.
207.
M.N.Mami, D.Graux, S.Scerri, H.Jabeen, S.Auer and J.Lehmann, How to feed the squerall with RDF and other data nuts? in: Proceedings of the ISWC 2019 Satellite Tracks (Posters & Demonstrations, Industry, and Outrageous Ideas) Co-Located with 18th International Semantic Web Conference (ISWC 2019), Auckland, New Zealand, October 26–30, 2019, CEUR Workshop Proceedings, Vol. 2456, CEUR-WS.org, 2019, pp. 293–296.
208.
M.N.Mami, D.Graux, S.Scerri, H.Jabeen and S.Auer, Querying Data Lakes using Spark and Presto, in: The World Wide Web Conference, WWW 2019, San Francisco, CA, USA, May 13-17, 2019, ACM, 2019, pp. 3574–3578. doi:10.1145/3308558.3314132.
209.
M.N.Mami, D.Graux, S.Scerri, H.Jabeen, S.Auer and J.Lehmann, Uniform access to multiform data lakes using semantic technologies, in: Proceedings of the 21st International Conference on Information Integration and Web-Based Applications & Services, IiWAS 2019, Munich, Germany, December 2–4, 2019, ACM, 2019, pp. 313–322. doi:10.1145/3366030.3366054.
210.
Q.Ge, P.Peng, Z.Xu, L.Zou and Z.Qin, FMQO: A federated RDF system supporting multi-query optimization, in: Web and Big Data – Third International Joint Conference, APWeb-WAIM 2019, Proceedings, Part II, Chengdu, China, August 1–3, 2019, Lecture Notes in Computer Science, Vol. 11642, Springer, 2019, pp. 397–401. doi:10.1007/978-3-030-26075-0_30.
211.
P.Peng, L.Zou, M.T.Özsu and D.Zhao, Multi-query optimization in federated RDF systems, in: Database Systems for Advanced Applications – 23rd International Conference, DASFAA 2018, Proceedings, Part I, Gold Coast, QLD, Australia, May 21–24, 2018, Lecture Notes in Computer Science, Vol. 10827, Springer, 2018, pp. 745–765. doi:10.1007/978-3-319-91452-7_48.
212.
A.Valdestilhas, T.Soru and M.Saleem, More complete resultset retrieval from large heterogeneous RDF sources, in: Proceedings of the 10th International Conference on Knowledge Capture, K-CAP 2019, Marina del Rey, CA, USA, November 19–21, 2019, ACM, 2019, pp. 223–230. doi:10.1145/3360901.3364436.
213.
B.Arsic, M.Ðokic-Petrovic, P.C.Spalevic, I.Z.Milentijevic, D.D.Rancic and M.Zivanovic, SpecINT: A framework for data integration over cheminformatics and bioinformatics RDF repositories, Semantic Web10(4) (2019), 795–813. doi:10.3233/SW-180327.
214.
P.Fafalios and Y.Tzitzikas, How many and what types of SPARQL queries can be answered through zero-knowledge link traversal? in: Proceedings of the 34th ACM/SIGAPP Symposium on Applied Computing, SAC 2019, Limassol, Cyprus, April 8–12, 2019, ACM, 2019, pp. 2267–2274. doi:10.1145/3297280.3297505.
215.
R.Singhal, N.Zhang, L.Nardi, M.Shahbaz and K.Olukotun, Polystore++: Accelerated polystore system for heterogeneous workloads, in: 39th IEEE International Conference on Distributed Computing Systems, ICDCS 2019, Dallas, TX, USA, July 7–10, 2019, IEEE, 2019, pp. 1641–1651. doi:10.1109/ICDCS.2019.00163.
216.
K.S.Aggour, V.S.Kumar, P.Cuddihy, J.W.Williams, V.Gupta, L.Dial, T.Hanlon, J.Gambone and J.Vinciquerra, Federated multimodal big data storage & analytics platform for additive manufacturing, in: 2019 IEEE International Conference on Big Data (IEEE BigData), Los Angeles, CA, USA, December 9–12, 2019, C.K.Baru, J.Huan, L.Khan, X.Hu, R.Ak, Y.Tian, R.S.Barga, C.Zaniolo, K.Lee and Y.F.Ye, eds, IEEE, 2019, pp. 1729–1738. doi:10.1109/BigData47090.2019.9006495.
217.
B.Moreau and P.Serrano-Alvarado, Ensuring license compliance in federated query processing, in: 36ème Conférence sur la Gestion de Données – Principes, Technologies et Applications (BDA 2020), 2020.