Dissertations / Theses on the topic 'Linked Open Dynamic Data'

To see the other types of publications on this topic, follow the link: Linked Open Dynamic Data.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Linked Open Dynamic Data.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Jain, Prateek. "Linked Open Data Alignment & Querying." Wright State University / OhioLINK, 2012. http://rave.ohiolink.edu/etdc/view?acc_num=wright1345575500.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Pfeffer, Magnus, and Kai Eckert. "Linked Open Projects." Universitätsbibliothek Chemnitz, 2011. http://nbn-resolving.de/urn:nbn:de:bsz:ch1-qucosa-64786.

Full text
Abstract:
Semantic Web und Linked Data sind in aller Munde. Nach fast einem Jahrzehnt der Entwicklung der Technologien und Erforschung der Möglichkeiten des Semantic Webs rücken nun die Daten in den Mittelpunk, denn ohne diese wäre das Semantic Web nicht mehr als ein theoretisches Konstrukt. Fast wie das World Wide Web ohne Websites. Bibliotheken besitzen mit Normdaten (PND, SWD) und Titelaufnahmen eine Fülle Daten, die sich zur Befüllung des Semantic Web eignen und teilweise bereits für das Semantic Web aufbereitet und zur Nutzung freigegeben wurden. Die Universitätsbibliothek Mannheim hat sich in zwei verschiedenen Projekten mit der Nutzung solcher Daten befasst – allerdings standen diese zu diesem Zeitpunkt noch nicht als Linked Data zur Verfügung. In einem Projekt ging es um die automatische Erschließung von Publikationen auf der Basis von Abstracts, im anderen Projekt um die automatische Klassifikation von Publikationen auf der Basis von Titeldaten. Im Rahmen dieses Beitrags stellen wir die Ergebnisse der Projekte kurz vor, möchten aber im Schwerpunkt auf einen Nebenaspekt eingehen, der sich erst im Laufe dieser Projekte herauskristallisiert hat: Wie kann man die gewonnenen Ergebnisse dauerhaft und sinnvoll zur Nachnutzung durch Dritte präsentieren? Soviel vorweg: Beide Verfahren können und wollen einen Bibliothekar nicht ersetzen. Die Einsatzmöglichkeiten der generierten Daten sind vielfältig. Konkrete Einsätze, zum Beispiel das Einspielen in einen Verbundkatalog, sind aber aufgrund der Qualität und mangelnden Kontrolle der Daten umstritten. Die Bereitstellung dieser Daten als Linked Data im Semantic Web ist da eine naheliegende Lösung – jeder, der die Ergebnisse nachnutzen möchte, kann das tun, ohne dass ein bestehender Datenbestand damit kompromittiert werden könnte. Diese Herangehensweise wirft aber neue Fragen auf, nicht zuletzt auch nach der Identifizierbarkeit der Ursprungsdaten über URIs, wenn diese (noch) nicht als Linked Data zur Verfügung stehen. Daneben erfordert die Bereitstellung von Ergebnisdaten aber auch weitere Maßnahmen, die über die gängige Praxis von Linked Data hinaus gehen: Die Bereitstellung von Zusatzinformationen, die die Quelle und das Zustandekommen dieser Daten näher beschreiben (Provenienzinformationen), aber auch weitere Informationen, die über das zugrunde liegende Metadatenschema meist hinausgehen, wie Konfidenzwerte im Falle eines automatischen Verfahrens der Datenerzeugung. Dazu präsentieren wir Ansätze auf Basis von RDF Reification und Named Graphs und schildern die aktuellen Entwicklungen auf diesem Gebiet, wie sie zum Beispiel in der Provenance Incubator Group des W3C und in Arbeitsgruppen der Dublin Core Metadaten-Initiative diskutiert werden.
APA, Harvard, Vancouver, ISO, and other styles
3

Felix, Juan Manuel. "Esplorando i Linked Open Data con RSLT." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2020. http://amslaurea.unibo.it/20925/.

Full text
Abstract:
Se è vero che i Linked Open Data accessibili sul Web sono ormai tantissimi, è vero anche che la fruizione di queste informazioni da parte di un pubblico umano è una questione quantomai spinoza. Questo lavoro si propone di esplorare le possibilità di creare applicazioni Web dinamiche con l'ausilio di RSLT, per la visualizzazione di Linked Open Data in formato RDF.
APA, Harvard, Vancouver, ISO, and other styles
4

Le, Xuan Dung. "Webové aplikace s využitím Linked Open Data." Master's thesis, Vysoká škola ekonomická v Praze, 2014. http://www.nusl.cz/ntk/nusl-198433.

Full text
Abstract:
This thesis deals with the issue of open data. The aim is to introduce to reader the currently very popular topic. Linking these data together gives us more advantages and opportuni-ties, however a large number of open data datasets are published in the format that cannot be linked together. Therefore, the author put great emphasis into his work on Linked Data. Emphasis is not placed only on the emergence, current status and future development, but also on the technical aspect. First, readers will be familiar with theoretical concepts, principles of Linked Open Data, expansion of open government data in the Czech Republic and abroad. In the next chapter, the author aimed at the data formats RDF, SPARQL language, etc. In the last section, the author introduce to readers the tools to work with Linked Open Data and design sample application using the Linked Open Data. The benefit of the whole work is a comprehensive view of the Linked Open Data both from a theoretical and from a practical part. The main goal is to provide to readers quality introduction to the issue.
APA, Harvard, Vancouver, ISO, and other styles
5

Sánchez, Adam. "Big Data, Linked Data y Web semántica." Universidad Peruana de Ciencias Aplicadas (UPC), 2016. http://hdl.handle.net/10757/620705.

Full text
Abstract:
Conferencia realizada en el marco de la Semana del Acceso Abierto Perú, llevada a cabo del 24 al 26 de Octubre de 2016 en Lima, Peru. Las instituciones organizadoras: Universidad Peruana de Ciencias aplciadasd (UPC), Pontificia Universidad Católica del Perú (PUCP) y Universidad Peruana Cayetano Heredia (UPCH).
Conferencia que aborda aspectos del protocolo Linked Data, temas de Big Data y Web Semantica,
APA, Harvard, Vancouver, ISO, and other styles
6

Gu, Chen. "Ontology Alignment Techniques for Linked Open Data Ontologies." Miami University / OhioLINK, 2013. http://rave.ohiolink.edu/etdc/view?acc_num=miami1386858615.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Minier, Thomas. "Web preemption for querying the linked data open." Thesis, Nantes, 2020. http://www.theses.fr/2020NANT4047.

Full text
Abstract:
En suivant les principes du Linked Open Data, les fournisseurs de données ont publié des milliards de documents RDF via des services publics d’évaluation de requêtes SPARQL. Pour garantir la disponibilité et la stabilité de ces services, ils appliquent des politiques de quotas sur l’utilisation des serveurs. Les requêtes qui excèdent ces quotas sont interrompues et ne renvoient que des résultats partiels. Cette interruption n’est pas un problème s’il est possible de reprendre l’exécution des requêtes ultérieurement, mais il n’existe aucun modèle de préemption le permettant. Dans cette thèse, nous proposons de résoudre le problème relatif à la construction des services qui permettent à n’importe quel utilisateur d’exécuter n’importe quelle requête SPARQL en obtenant des résultats complets. Nous proposons la préemption Web, un nouveau modèle d’exécution qui permet l’interruption de requêtes SPARQL après un quantum de temps, ainsi que leur reprise sur demande des clients. Nous proposons également SaGe, un moteur d’évaluation de requêtes SPARQL qui implémente la préemption Web tout en garantissant un surcoût de préemption minimal. Nos résultats expérimentaux démontrent que SaGe est plus performant que les approches existantes, en termes de temps moyen d’exécution des requêtes et d’obtention des premiers résultats
Following the Linked Open Data principles, data providers have published billions of RDF documents using public SPARQL query services. To ensure these services remains stable and responsive, they enforce quotas on server usage. Queries which exceed these quotas are interrupted and deliver partial results. Such interruption is not an issue if it is possible to resume queries execution afterward. Unfortunately, there is no preemption model for the Web that allows for suspending and resuming SPARQL queries. In this thesis, we propose to tackle the issue of building public SPARQL query servers that allow any data consumer to execute any SPARQL query with complete results. First, we propose a new query execution model called Web Preemption. It allows SPARQL queries to be suspended by the Web server after a fixed time quantum and resumed upon client request. Web preemption is tractable only if its cost in time is negligible compared to the time quantum. Thus, we propose SaGe: a SPARQL query engine that implements Web Preemption with minimal overhead. Experimental results demonstrate that SaGe outperforms existing SPARQL query processing approaches by several orders of magnitude in term of the average total query execution time and the time for first results
APA, Harvard, Vancouver, ISO, and other styles
8

Budka, Michal. "Návrh postupu tvorby aplikace pro Linked Open Data." Master's thesis, Vysoká škola ekonomická v Praze, 2014. http://www.nusl.cz/ntk/nusl-193918.

Full text
Abstract:
This thesis deals with the issue of Linked Open Data. The goal of this thesis is to introduce the reader to this issue as a whole and to the possibility of using Linked Open Data for developing useful applications by proposing a new development process focusing on such applications. The theoretical part offers an insight into the issue of Open Data, Linked Open Data and the NoSQL database systems and their usability in this field. It focuses mainly on graph database systems and compares them with relational database systems using predefined criteria. Additionally, the goal of this thesis is to develop an application using the proposed development process, which provides a tool for data presentation and statistical visualisation for open data sets published by the Supreme Audit Office and the Czech Trade Inspection. The application is mainly developed for the purpose of verifying the proposed development process and to demonstrate the connectivity of open data published by two different organizations.The thesis includes the process of selecting a development methodology, which is then used for optimising work on the implementation of the resulting application and the process of selecting a graph database system, that is used to store and modify open data for the purposes of the application.
APA, Harvard, Vancouver, ISO, and other styles
9

Passerini, Serena. "Analisi Empirica delle Distinzioni Fondazionali in Linked Open Data." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2019. http://amslaurea.unibo.it/18442/.

Full text
Abstract:
Analisi delle distinzioni fondazionali nei Linked Open Data. Sono state analizzate 4,5 milioni di entità DBPedia, successivamente classificate tramite metodi di classificazione basati sull'apprendimento automatico. I risultati sono stati validati tramite Crowdsourcing. I risultati sono stati confrontati con quelli della ricerca presentata nell'articolo "Empirical Analysis of Foundational Distinctions in Linked Open Data” scritto da Luigi Asprino, Valerio Basile, Paolo Ciancarini e Valentina Presutti.
APA, Harvard, Vancouver, ISO, and other styles
10

RODRIGUES, LIVIA COUTO RUBACK. "ENRICHING AND ANALYZING SEMANTIC TRAJECTORIES WITH LINKED OPEN DATA." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2017. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=33109@1.

Full text
Abstract:
PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO
COORDENAÇÃO DE APERFEIÇOAMENTO DO PESSOAL DE ENSINO SUPERIOR
PROGRAMA DE EXCELENCIA ACADEMICA
Os últimos anos testemunharam o uso crescente de dispositivos que rastreiam objetos móveis: equipamentos com GPS e telefones móveis, veículos ou outros sensores da Internet das Coisas, além de dados de localização de check-ins de redes sociais. Estes dados de mobilidade são representados como trajetórias, e armazenam a sequência de posições de um objeto móvel. Porém, estas sequências representam somente os dados de posição originais, que precisam ser semanticamente enriquecidos para permitir tarefas de análise e apoiar um entendimento profundo sobre o comportamento do movimento. Um outro espaço de dados global sem precedentes tem crescido rapidamente, a Web de Dados, graças à iniciativa de Dados Interligados. Estes dados semânticos ricos e livremente disponíveis fornecem uma nova maneira de enriquecer dados de trajetória. Esta tese apresenta contribuições para os desafios que surgem considerando este cenário. Em primeiro lugar, a tese investiga como dados de trajetória podem se beneficiar da iniciativa de dados interligados, guiando todo o processo de enriquecimento semântico utilizando fontes de dados externas. Em segundo lugar, aborda o tópico de computação de similaridade entre entidades representadas como dados interligados com o objetivo de computar a similaridade entre trajetórias semanticamente enriquecidas. A novidade da abordagem apresentada nesta tese consiste em considerar as características relevantes das entidades como listas ranqueadas. Por último, a tese aborda a computação da similaridade entre trajetórias enriquecidas comparando a similaridade entre todas as entidades representadas como dados interligados que representam as trajetórias enriquecidas.
The last years witnessed a growing number of devices that track moving objects: personal GPS equipped devices and GSM mobile phones, vehicles or other sensors from the Internet of Things but also the location data deriving from the Social Networks check-ins. These mobility data are represented as trajectories, recording the sequence of locations of the moving object. However, these sequences only represent the raw location data and they need to be semantically enriched to be meaningful in the analysis tasks and to support a deep understanding of the movement behavior. Another unprecedented global space that is also growing at a fast pace is the Web of Data, thanks to the emergence of the Linked Data initiative. These freely available semantic rich datasets provide a novel way to enhance trajectory data. This thesis presents a contribution to the many challenges that arise from this scenario. First, it investigates how trajectory data may benefit from the Linked Data Initiative by guiding the whole trajectory enrichment process with the use of external datasets. Then, it addresses the pivotal topic of the similarity computation between Linked Data entities with the final objective of computing the similarity between semantically enriched trajectories. The novelty of our approach is that the thesis considers the relevant entity features as a ranked list. Finally, the thesis targets the computation of the similarity between enriched trajectories by comparing the similarity of the Linked Data entities that represent the enriched trajectories.
APA, Harvard, Vancouver, ISO, and other styles
11

Franchini, Giulia. "Associazioni non profit e linked open data: un esperimento." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2015. http://amslaurea.unibo.it/8350/.

Full text
Abstract:
Le Associazioni Non Profit giocano un ruolo sempre più rilevante nella vita dei cittadini e rappresentano un'importante realtà produttiva del nostro paese; molto spesso però risulta difficile trovare informazioni relative ad eventi, attività o sull'esistenza stessa di queste associazioni. Per venire in contro alle esigenze dei cittadini molte Regioni e Province mettono a disposizione degli elenchi in cui sono raccolte le informazioni relative alle varie organizzazioni che operano sul territorio. Questi elenchi però, presentano spesso grossi problemi, sia per quanto riguarda la correttezza dei dati, sia per i formati utilizzati per la pubblicazione. Questi fattori hanno portato all'idea e alla necessità di realizzare un sistema per raccogliere, sistematizzare e rendere fruibili le informazioni sulle Associazioni Non Profit presenti sul territorio, in modo che questi dati possano essere utilizzati liberamente da chiunque per scopi diversi. Il presente lavoro si pone quindi due obiettivi principali: il primo consiste nell'implementazione di un tool in grado di recuperare le informazioni sulle Associazioni Non Profit sfruttando i loro Siti Web; questo avviene per mezzo dell'utilizzo di tecniche di Web Crawling e Web Scraping. Il secondo obiettivo consiste nel pubblicare le informazioni raccolte, secondo dei modelli che ne permettano un uso libero e non vincolato; per la pubblicazione e la strutturazione dei dati è stato utilizzato un modello basato sui principi dei linked open data.
APA, Harvard, Vancouver, ISO, and other styles
12

Fusetti, Camilla <1984&gt. "Dati bibliografici e linked (open) data: prospettive e criticità." Master's Degree Thesis, Università Ca' Foscari Venezia, 2017. http://hdl.handle.net/10579/9479.

Full text
Abstract:
La tesi si propone di analizzare l’applicazione dei linked open data ai dati bibliografici gestiti e creati dalle biblioteche. La necessità da parte delle biblioteche di integrare le proprie risorse informative con il web ha portato al ripensamento della struttura dei dati stessi e degli strumenti di accesso all’informazione. I linked open data rappresentano una delle tecnologie possibili per rispondere alle esigenze di un nuovo contesto di produzione e fruizione della conoscenza. La tesi si sviluppa a partire dalla contestualizzazione dei dati bibliografici prendendo in esame le criticità e le rinnovate esigenze dei sistemi catalografici, che portano al ripensamento delle procedure di strutturazione dei dati e delle funzionalità dei cataloghi. Il secondo capitolo definisce il quadro teorico e tecnologico dei linked open data, con riferimento ai principi e ai linguaggi che ne sono alla base. Vengono poi analizzate le relazioni tra queste tecnologie e i dati bibliografici, con riguardo alle linee guida proposte dal Library Linked Data Incubator Group e alle ontologie e schemi di dati necessari per la rappresentazione delle risorse informative. Per concludere, tre casi di studio illustrano come istituzioni legate al mondo delle biblioteche (Europeana, Digital Public Library of America e OCLC) abbiano convertito e pubblicato i propri dati come linked open data. L’obiettivo della tesi è duplice: offrire un quadro teorico di riferimento riguardo i dati bibliografici e i linked open data e presentare delle applicazioni pratiche attraverso l’analisi dei casi di studio.
APA, Harvard, Vancouver, ISO, and other styles
13

Giorgi, Riccardo. "Semantic publishing: analisi di Linked Open Dataset." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2014. http://amslaurea.unibo.it/7238/.

Full text
Abstract:
Il presente lavoro si occupa di fare una rassegna esaustiva di alcuni Linked Open Dataset nel contesto delle pubblicazioni scientifiche, cercando di inquadrare la loro eterogeneità ed identificando i principali pregi e difetti di ciascuno. Inoltre, descriviamo il nostro prototipo GReAT (Giorgi's Redundant Authors Tool), creato per il corretto riconoscimento e disambiguazione degli autori.
APA, Harvard, Vancouver, ISO, and other styles
14

Ahmad, Hussien, and Salah Dowaji. "Linked-OWL: A new approach for dynamic linked data service workflow composition." Webology, 2013. http://hdl.handle.net/10150/299603.

Full text
Abstract:
The shift from Web of Document into Web of Data based on Linked Data principles defined by Tim Berners-Lee posed a big challenge to build and develop applications to work in Web of Data environment. There are several attempts to build service and application models for Linked Data Cloud. In this paper, we propose a new service model for linked data "Linked-OWL" which is based on RESTful services and OWL-S and copes with linked data principles. This new model shifts the service concept from functions into linked data things and opens the road for Linked Oriented Architecture (LOA) and Web of Services as part and on top of Web of Data. This model also provides high level of dynamic service composition capabilities for more accurate dynamic composition and execution of complex business processes in Web of Data environment.
APA, Harvard, Vancouver, ISO, and other styles
15

Musyaffa, Fathoni Arief [Verfasser]. "Comparative Analysis of Open Linked Fiscal Data / Fathoni Arief Musyaffa." Bonn : Universitäts- und Landesbibliothek Bonn, 2021. http://d-nb.info/123868744X/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
16

Bedeschi, Luca. "Analisi sulla crescita e sulle funzioni dei Linked Open Data - LODStories." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2014. http://amslaurea.unibo.it/7733/.

Full text
Abstract:
L'Open Data, letteralmente “dati aperti”, è la corrente di pensiero (e il relativo “movimento”) che cerca di rispondere all'esigenza di poter disporre di dati legalmente “aperti”, ovvero liberamente re-usabili da parte del fruitore, per qualsiasi scopo. L’obiettivo dell’Open Data può essere raggiunto per legge, come negli USA dove l’informazione generata dal settore pubblico federale è in pubblico dominio, oppure per scelta dei detentori dei diritti, tramite opportune licenze. Per motivare la necessità di avere dei dati in formato aperto, possiamo usare una comparazione del tipo: l'Open Data sta al Linked Data, come la rete Internet sta al Web. L'Open Data, quindi, è l’infrastruttura (o la “piattaforma”) di cui il Linked Data ha bisogno per poter creare la rete di inferenze tra i vari dati sparsi nel Web. Il Linked Data, in altre parole, è una tecnologia ormai abbastanza matura e con grandi potenzialità, ma ha bisogno di grandi masse di dati tra loro collegati, ossia “linkati”, per diventare concretamente utile. Questo, in parte, è già stato ottenuto ed è in corso di miglioramento, grazie a progetti come DBpedia o FreeBase. In parallelo ai contributi delle community online, un altro tassello importante – una sorta di “bulk upload” molto prezioso – potrebbe essere dato dalla disponibilità di grosse masse di dati pubblici, idealmente anche già linkati dalle istituzioni stesse o comunque messi a disposizione in modo strutturato – che aiutino a raggiungere una “massa” di Linked Data. A partire dal substrato, rappresentato dalla disponibilità di fatto dei dati e dalla loro piena riutilizzabilità (in modo legale), il Linked Data può offrire una potente rappresentazione degli stessi, in termini di relazioni (collegamenti): in questo senso, Linked Data ed Open Data convergono e raggiungono la loro piena realizzazione nell’approccio Linked Open Data. L’obiettivo di questa tesi è quello di approfondire ed esporre le basi sul funzionamento dei Linked Open Data e gli ambiti in cui vengono utilizzati.
APA, Harvard, Vancouver, ISO, and other styles
17

Mastria, Maurizio. "Disambiguazione di dati da fonti eterogenee in Linked Open Data: un'esperienza." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2015. http://amslaurea.unibo.it/8850/.

Full text
Abstract:
La pubblicazione si incentra sulla descrizione di un programma generico di disambiguazione di IRI e letterali, in Linked Open Data, fortemente configurabile, quindi applicabile in più contesti. CALID è la sigla di "Customizable Application for Literal and IRI's Disambiguation". Esso è stato creato per risolvere la disambiguazione degli autori di pubblicazioni scientifiche, e in questo articolo viene descritta la parte progettuale, il modo in cui si utilizza e i valori di performance e precisione ottenuti testandolo su diversi datasets.
APA, Harvard, Vancouver, ISO, and other styles
18

CARABALLO, ALEXANDER ARTURO MERA. "CLUSTERING AND DATASET INTERLINKING RECOMMENDATION IN THE LINKED OPEN DATA CLOUD." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2017. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=30656@1.

Full text
Abstract:
PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO
COORDENAÇÃO DE APERFEIÇOAMENTO DO PESSOAL DE ENSINO SUPERIOR
FUNDAÇÃO DE APOIO À PESQUISA DO ESTADO DO RIO DE JANEIRO
PROGRAMA DE EXCELENCIA ACADEMICA
BOLSA NOTA 10
O volume de dados RDF publicados na Web aumentou consideravelmente, o que ressaltou a importância de seguir os princípios de dados interligados para promover a interoperabilidade. Um dos princípios afirma que todo novo conjunto de dados deve ser interligado com outros conjuntos de dados publicados na Web. Esta tese contribui para abordar este princípio de duas maneiras. Em primeiro lugar, utiliza algoritmos de detecção de comunidades e técnicas de criação de perfis para a criação e análise automática de um diagrama da nuvem da LOD (Linked Open Data), o qual facilita a localização de conjuntos de dados na nuvem da LOD. Em segundo lugar, descreve três abordagens, apoiadas por ferramentas totalmente implementadas, para recomendar conjuntos de dados a serem interligados com um novo conjunto de dados, um problema conhecido como problema de recomendação de interligação de conjunto de dados. A primeira abordagem utiliza medidas de previsão de links para produzir recomendações de interconexão. A segunda abordagem emprega algoritmos de aprendizagem supervisionado, juntamente com medidas de previsão de links. A terceira abordagem usa algoritmos de agrupamento e técnicas de criação de perfil para produzir recomendações de interconexão. Essas abordagens são implementadas, respectivamente, pelas ferramentas TRT, TRTML e DRX. Por fim, a tese avalia extensivamente essas ferramentas, usando conjuntos de dados do mundo real. Os resultados mostram que estas ferramentas facilitam o processo de criação de links entre diferentes conjuntos de dados.
The volume of RDF data published on the Web increased considerably, which stressed the importance of following the Linked Data principles to foster interoperability. One of the principles requires that a new dataset should be interlinked with other datasets published on the Web. This thesis contributes to addressing this principle in two ways. First, it uses community detection algorithms and profiling techniques for the automatic creation and analysis of a Linked Open Data (LOD) diagram, which facilitates locating datasets in the LOD cloud. Second, it describes three approaches, backed up by fully implemented tools, to recommend datasets to be interlinked with a new dataset, a problem known as the dataset interlinking recommendation problem. The first approach uses link prediction measures to provide a list of datasets recommendations for interlinking. The second approach employs supervised learning algorithms, jointly with link prediction measures. The third approach uses clustering algorithms and profiling techniques to produce dataset interlinking recommendations. These approaches are backed up, respectively, by the TRT, TRTML and DRX tools. Finally, the thesis extensively evaluates these tools, using real-world datasets, reporting results that show that they facilitate the process of creating links between disparate datasets.
APA, Harvard, Vancouver, ISO, and other styles
19

Meymandpour, Rouzbeh. "Semantic analysis of linked open data: an information content-based approach." Thesis, The University of Sydney, 2014. http://hdl.handle.net/2123/12848.

Full text
Abstract:
The Semantic Web is a collection of standards and technologies that makes Web documents ready to be consumed, reused and shared by applications. Linked Open Data (LOD) is a recent community-driven effort to provide access to a large amount of structured data in diverse domains using semantic technologies and through open standards and liberal licences. This not only offers unprecedented opportunities for developing novel and innovative applications but also makes the application development more efficient and cost-effective. LOD is a complex semantic network of information resources interlinked via meaningful, semantic relations. A wide range of entities such as movies, artists, books, etc. are represented as resources in LOD and are semantically linked to other related entities. These entities are described using billions of statements with various levels of informativeness which can significantly affect the quality of LOD-based semantic applications. A primary challenge in semantic analysis is to systematically define what is considered to be useful information. This thesis addresses the problem of reliable and valid measurement of LOD informativeness based on the concept of information content (IC): defining information as a measurable mathematical quantity. We extend the notion of IC measurement to LOD, and develop, evaluate and experiment with several measures of informativeness. By building on a valid mathematical definition of LOD which complies with accepted standards and principles, we ensure that our proposed measures are robust and reliable. This is supported by experimental evaluations using well-established benchmark data and evaluation metrics. These experiments also demonstrate the applicability and value of the proposed measures in diverse applications and domains. First, we propose partitioned information content (PIC) which is a measure of the information content of entities in LOD. As a fundamental application area, PIC is applied to entity ranking problem. The PIC-based approach for ranking universities shows a high degree of correlation with international, well-established ranking systems. Second, we develop the generated information content (GIC) measure that assesses the informativeness of relations in LOD. It has a wide range of applications in semantic navigation, faceted browsing and visualisation. Third, this thesis presents a novel, PIC-based semantic similarity measure of resources, called PICSS. We apply PICSS to develop a hybrid recommender system. The experimental evaluation of the proposed approach shows that it outperforms the comparable recommender systems, especially, in situations where there is a lack of information on newly added items. Finally, PICSS-based measures are applied to address the problem of lack of diversity in recommendations in order to better satisfy users’ requirements and to increase the average diversity of the recommendations, while preserving the overall accuracy.
APA, Harvard, Vancouver, ISO, and other styles
20

PALMERO, APROSIO ALESSIO. "EXTENDING LINKED OPEN DATA RESOURCES EXPLOITING WIKIPEDIA AS SOURCE OF INFORMATION." Doctoral thesis, Università degli Studi di Milano, 2014. http://hdl.handle.net/2434/233327.

Full text
Abstract:
DBpedia is a project aiming to represent Wikipedia content in RDF triples. It plays a central role in the Semantic Web, due to the large and growing number of resources linked to it. Currently, the information contained in DBpedia is mainly collected from Wikipedia infoboxes, a set of subject-attribute-value triples that represent a summary of the Wikipedia page. The extraction procedure requires to manually map Wikipedia infoboxes into the DBpedia ontology. Thanks to crowdsourcing, a large number of infoboxes in the English Wikipedia has been mapped to the corresponding classes in DBpedia. Subsequently, the same procedure has been applied to other languages to create the localized versions of DBpedia. However, (i) the number of accomplished mappings is still small and limited to most frequent infoboxes, (ii) mappings need maintenance due to the constant and quick changes of Wikipedia articles, and (iii) infoboxes are manually compiled by the Wikipedia contributors, therefore in more than 50% of the Wikipedia articles the infobox is missing. As a demonstration of these issues, only 1.7M Wikipedia pages are “deeply” classified in the DBpedia ontology, although the English Wikipedia contains almost 4M pages. This shows a clear problem of coverage, and this issue is even worse in other languages (like French and Spanish). The objective of this thesis is to define a methodology to increase the coverage of DBpedia in different languages, using various techniques to reach two different goals: automatic mapping generation and DBpedia dataset completion. A key aspect of our research is multi-linguality in Wikipedia: we bootstrap the available information through cross-language links, starting from the available mappings in some pivot languages, and then extending the existing DBpedia datasets (or create new ones from scratch) comparing the classifications in different languages. When the DBpedia classification is missing, we train a supervised classifier using the original DBpedia as training. We also use the Distant Supervision paradigm to extract the missing properties directly from the Wikipedia articles. We evaluated our system using a manually annotated test set and some existing DBpedia mappings excluded from the training. The results demonstrate the suitability of the approach in extending the DBpedia resource. Finally, the resulting resources are made available through a SPARQL endpoint and a downloadable package.
APA, Harvard, Vancouver, ISO, and other styles
21

Pääkkölä, Jonas. "OPEN LINKED DATA : HOW TO CONVERT AND PUBLISH STRUCTURED DATA AND DEMONSTRATIONS OF POSSIBLE USES." Thesis, Umeå universitet, Institutionen för fysik, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-104930.

Full text
Abstract:
The goal with this project was to convert open structured data, e.g. csv-files, to Linked Open Data and to demonstrate possible uses of such data. The project was also supposed to inspire and lay a foundation for future work in the area. The conversion was done using the D2RQ platform, and the chosen dataset contains air quality measurements from Umeå Municipality. The resulting data was then published on the internet with D2R server. For demonstration purposes two tasks were done. A visualization of the converted data was published on the web, together with traffic and weather data. Secondly a physical city model was built of carton with eight photodiodes, a Raspberry Pi and visualized with a local webserver. The goals for the project were fulfilled and it has also inspired Knowit to continue with commercial projects in the area of Linked Open Data. Future effort should be put in converting more data to Linked Open Data and to create full scale sensor networks in a city.
APA, Harvard, Vancouver, ISO, and other styles
22

Ianniello, Raffaele. "Linked Open Data per la pubblica amministrazione: conversione e utilizzo dei dati." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2013. http://amslaurea.unibo.it/5776/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
23

Bazzali, Denis. "Un'applicazione mobile di guida turistica context-aware basata su linked open data." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2014. http://amslaurea.unibo.it/7446/.

Full text
Abstract:
In questo elaborato viene presentata Semantic City Guide, un'applicazione mobile di guida turistica basata su Linked Open Data. Si vogliono presentare i principali vantaggi e svantaggi derivati dall'interazione tra sviluppo nativo di applicazioni mobili e tecnologie del Semantic Web. Il tutto verrà contestualizzato esaminando alcuni progetti di aziende ed enti statali operativi nel settore turistico e dell'informatica.
APA, Harvard, Vancouver, ISO, and other styles
24

Bischof, Stefan, Andreas Harth, Benedikt Kämpgen, Axel Polleres, and Patrik Schneider. "Enriching integrated statistical open city data by combining equational knowledge and missing value imputation." Elsevier, 2017. http://dx.doi.org/10.1016/j.websem.2017.09.003.

Full text
Abstract:
Several institutions collect statistical data about cities, regions, and countries for various purposes. Yet, while access to high quality and recent such data is both crucial for decision makers and a means for achieving transparency to the public, all too often such collections of data remain isolated and not re-useable, let alone comparable or properly integrated. In this paper we present the Open City Data Pipeline, a focused attempt to collect, integrate, and enrich statistical data collected at city level worldwide, and re-publish the resulting dataset in a re-useable manner as Linked Data. The main features of the Open City Data Pipeline are: (i) we integrate and cleanse data from several sources in a modular and extensible, always up-to-date fashion; (ii) we use both Machine Learning techniques and reasoning over equational background knowledge to enrich the data by imputing missing values, (iii) we assess the estimated accuracy of such imputations per indicator. Additionally, (iv) we make the integrated and enriched data, including links to external data sources, such as DBpedia, available both in a web browser interface and as machine-readable Linked Data, using standard vocabularies such as QB and PROV. Apart from providing a contribution to the growing collection of data available as Linked Data, our enrichment process for missing values also contributes a novel methodology for combining rule-based inference about equational knowledge with inferences obtained from statistical Machine Learning approaches. While most existing works about inference in Linked Data have focused on ontological reasoning in RDFS and OWL, we believe that these complementary methods and particularly their combination could be fruitfully applied also in many other domains for integrating Statistical Linked Data, independent from our concrete use case of integrating city data.
APA, Harvard, Vancouver, ISO, and other styles
25

Rafes, Karima. "Le Linked Data à l'université : la plateforme LinkedWiki." Thesis, Université Paris-Saclay (ComUE), 2019. http://www.theses.fr/2019SACLS032/document.

Full text
Abstract:
Le Center for Data Science de l’Université Paris-Saclay a déployé une plateforme compatible avec le Linked Data en 2016. Or, les chercheurs rencontrent face à ces technologies de nombreuses difficultés. Pour surmonter celles-ci, une approche et une plateforme appelée LinkedWiki, ont été conçues et expérimentées au-dessus du cloud de l’université (IAAS) pour permettre la création d’environnements virtuels de recherche (VRE) modulaires et compatibles avec le Linked Data. Nous avons ainsi pu proposer aux chercheurs une solution pour découvrir, produire et réutiliser les données de la recherche disponibles au sein du Linked Open Data, c’est-à-dire du système global d’information en train d’émerger à l’échelle du Web. Cette expérience nous a permis de montrer que l’utilisation opérationnelle du Linked Data au sein d’une université est parfaitement envisageable avec cette approche. Cependant, certains problèmes persistent, comme (i) le respect des protocoles du Linked Data et (ii) le manque d’outils adaptés pour interroger le Linked Open Data avec SPARQL. Nous proposons des solutions à ces deux problèmes. Afin de pouvoir vérifier le respect d’un protocole SPARQL au sein du Linked Data d’une université, nous avons créé l’indicateur SPARQL Score qui évalue la conformité des services SPARQL avant leur déploiement dans le système d’information de l’université. De plus, pour aider les chercheurs à interroger le LOD, nous avons implémenté le démonstrateur SPARQLets-Finder qui démontre qu’il est possible de faciliter la conception de requêtes SPARQL à l’aide d’outils d’autocomplétion sans connaissance préalable des schémas RDF au sein du LOD
The Center for Data Science of the University of Paris-Saclay deployed a platform compatible with Linked Data in 2016. Because researchers face many difficulties utilizing these technologies, an approach and then a platform we call LinkedWiki were designed and tested over the university’s cloud (IAAS) to enable the creation of modular virtual search environments (VREs) compatible with Linked Data. We are thus able to offer researchers a means to discover, produce and reuse the research data available within the Linked Open Data, i.e., the global information system emerging at the scale of the internet. This experience enabled us to demonstrate that the operational use of Linked Data within a university is perfectly possible with this approach. However, some problems persist, such as (i) the respect of protocols and (ii) the lack of adapted tools to interrogate the Linked Open Data with SPARQL. We propose solutions to both these problems. In order to be able to verify the respect of a SPARQL protocol within the Linked Data of a university, we have created the SPARQL Score indicator which evaluates the compliance of the SPARQL services before their deployments in a university’s information system. In addition, to help researchers interrogate the LOD, we implemented a SPARQLets-Finder, a demonstrator which shows that it is possible to facilitate the design of SPARQL queries using autocompletion tools without prior knowledge of the RDF schemas within the LOD
APA, Harvard, Vancouver, ISO, and other styles
26

Nohejl, Petr. "Transformace a publikace otevřených a propojitelných dat." Master's thesis, Vysoká škola ekonomická v Praze, 2013. http://www.nusl.cz/ntk/nusl-198076.

Full text
Abstract:
The principle of Open Data and Linked data is in growing interest of many organizations, developers and even government institutions. This work is aimed on providing actual information about development of Open and Linked data, further there are introduced featured tools for creating, manipulating, transformation and other operations regarding Open and Linked Data. Finally, there is description of development of Linked Data application based on universal visualization system Payola.
APA, Harvard, Vancouver, ISO, and other styles
27

Schaible, Johann [Verfasser]. "TermPicker: Recommending Vocabulary Terms for Reuse When Modeling Linked Open Data / Johann Schaible." Kiel : Universitätsbibliothek Kiel, 2017. http://d-nb.info/1127044257/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
28

Escobar, Esteban María Pilar. "Un enfoque multidimensional basado en RDF para la publicación de Linked Open Data." Doctoral thesis, Universidad de Alicante, 2020. http://hdl.handle.net/10045/109950.

Full text
Abstract:
Cada vez hay disponibles más datos de manera pública en Internet y surgen nuevas bases de conocimiento conocidas como Knowledge Graph, basadas en conceptos de Linked Open Data (datos abiertos enlazados), como DBPedia, Wikidata, YAGO o Google Knowledge Graph, que cubren un amplio abanico de campos del conocimiento. Además, se incorporan los datos que provienen de diversas fuentes como dispositivos inteligentes o las redes sociales. Sin embargo, que estos datos estén públicos y accesibles no garantiza que sean útiles para los usuarios, no siempre se garantiza que sean confiables ni que puedan ser reutilizados de manera eficiente. Actualmente, siguen existiendo barreras que dificultan la reutilización de los datos, porque los formatos son poco adecuados para el procesamiento automático y publicación de la información, por falta de metadatos descriptivos y de semántica, duplicidades, ambigüedad o incluso errores en los propios datos. A todos estos problemas hay que añadir la complejidad del proceso de explotación de la información de un repositorio de datos abiertos enlazados. El trabajo y conocimientos técnicos que requiere el acceso, recolección, normalización y preparación de los datos para que puedan ser reutilizados supone una carga extra para los usuarios y organizaciones que quieran utilizarlos. Para garantizar una eficiente explotación de los mismos, resulta fundamental dotarlos de más valor estableciendo conexiones con otros repositorios que permitan enriquecerlos; garantizar su valor, evaluando y mejorando la calidad de lo que se publica; y asimismo ofrecer los mecanismos necesarios que faciliten su explotación. En este trabajo de tesis se ha propuesto un modelo para la publicación de Linked Open Data que, a partir de un conjunto de datos obtenidos de diversas fuentes, facilita la publicación, enriquecimiento y validación de los datos, generando información útil y de calidad orientada a usuarios expertos y no expertos.
APA, Harvard, Vancouver, ISO, and other styles
29

Persello, Mara <1974&gt. "Un archivio sottoculturale con i linked open data: aspetti etici e aspetti pratici." Master's Degree Thesis, Università Ca' Foscari Venezia, 2019. http://hdl.handle.net/10579/15872.

Full text
Abstract:
Un primo capitolo dedicato alla teoria della cultura espone il modello sociosemiotico di Juri Lotman e Umberto Eco, che definiscono rispettivamente la cultura come semiosfera, e il progresso della conoscenza come forma di enciclopedia. Confrontando questo modello teorico con i sistemi di gestione delle informazioni, sia quelli del web che quelli più strettamente biblioteconomici, emerge una serie di criticità etiche, dato che i modi di organizzazione del sapere influenzano la percezione culturale delle informazioni accessibili e d'altronde l'orizzonte culturale stabilisce la pertinenza dei documenti elegibili. Si propone attraverso un esempio pratico di archivio dedicato alle sottoculture italiane, da costruirsi con i linked open data, una possibile soluzione alle criticità sia etiche che pratiche rilevate. Si espone un modello di archiviazione che tenga presente da una parte le possibilità fornite dagli strumenti tecnici, e dall'altra la struttura socialmente significativa della sottocultura, cercando un modello di partecipazione alla raccolta della documentazione che rispetti i principi dell'accessibilità e della cooperazione orizzontale che la sottocultura prevede.
APA, Harvard, Vancouver, ISO, and other styles
30

Eduards, Rasmus. "Current and Potential Use of Linked Geodata." Thesis, KTH, Geoinformatik, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-210784.

Full text
Abstract:
As of Today (2017) Geographic Information (GI) is a vital part of our daily life. With different applications like Google Maps it is hard to not get in contact with these platforms. Applications like Google are becoming more than just maps for us to find our way in the real world, they contain important data. As of now some of these datasets are kept by authorities and institutes with no connection to each other. One way to link this information to each other is by using Linked Data and more specifically when it comes to GI, Linked Geodata. By linking data together, information becomes connected, which can help the structure of Open Data and other data collaborates. It also enables ways to query the data to for example in search engines. This Bachelor of Science thesis has been conducted at KTH Royal Institute of Technology, in cooperation with Digpro AB. This thesis purpose is to examine whether the Linked Geodata is something to invest in. This was done by investigating current use to understand how Linked Geodata is implemented, as well to describe challenges and possibilities in respect to Linked Geodata. This is done by literature review and through interviews with personnel working with implementation of Linked Geodata. The result showed some implementations in the Netherlands and in Finland, also a private initiative from the University of Leipzig called LinkedGeoData. In Sweden authorities had explored the topic of Linked Geodata without any actual attempts to implement it. The biggest challenges was that queries did not supported all kind of spatial data, maintain the Linked Geodata consistent and find a way to fund the workload. The biggest possibilities were to create cooperation between authorities, integration and discoverability of data in search engines and to improve the environment for publishing open data, which could lead to an improved social and economic situation. After evaluation this thesis concludes that there is a lot of potential use for Linked Geodata. The most considerable possible use is for authorities with a larger amount of geodata especially regarding their publishing of Open Data and integrating their data to search engines to provide more advanced queries. The technology seems to have some problems, mainly the lack of support for spatial data and also problems with maintaining the connections. However the problems are not too severe in order to not invest in the technology. The technology just needs some improvements and more initiatives.
Idag (2017) är Geografisk Information (GI) en viktigt del av vårt dagliga liv. Med olika applikationer som Google Maps så är det svårt att inte komma i kontakt med sådana plattformar. Dem börjar bli mer än bara kartor för att hitta vart man ska. Idag är informationen i många fall inte knuten till varandra vilket betyder att informationen skulle kunna utnyttjas bättre om det var länkat. Ett sätt att länka sådan information och länka objekt till varandra är med Länkade Data och mer specifikt när det kommer till GI Länkade Geodata. Länkade Data kan sedan användas vid publicering av öppen data, för att berika mängden information. Det kan också användas för att förberedd webben för maskin läsning. Med detta menas att datorer ska kunna läsa av webben. Detta är en Kandidat examensarbete som har varit dirigerat av Kungliga Tekniska Högskolan, i samarbete med företaget Digpro AB. Syftet med denna uppsats är att ta reda på om Länkade Geodata är något att investera i. Detta var utfört genom att ta reda på hur dagsläget ser ut i olika länder samt hur det är implementerat. Samt beskriva utmaningar och möjligheter med Länkade Geodata. Detta är utfört genom litteraturstudier och intervjuer med behörig personal som antingen arbetar inom Geodata sektorn eller med Länkade Geodata. Resultatet visade några implementationer i Nederländer och Finland samt ett privat initiativ av ett Universitet kallat LinkedGeoData. I Sverige hade institut utvärderat möjligheterna för Länkade Geodata samt kommit fram med riktlinjer, men ingen storskalig implementation har blivit gjord. De störst utmaningarna var hitta tillräckligt med stöd för alla typer av spatial data, underhålla det så kallade Semantiska Molnet samt fördela arbetsbörda och hitta en finansiär. Det största möjligheterna var att kunna skapa en samverkan mellan olika instituts data, integration och upptäckbarhet av information i sökmotorer samt en förbättrad miljö för publicering av öppen data som kan leda till social och ekonomiska förbättringar. Den här uppsatsen drar slutsatsen att det finns stor potential användande för Länkade Geodata. Det störst användningsområdet är för institut med stor mängder geodata speciellt när det kommer till användandet av att publicera Öppen Data och integrera information till sökmotorer för att möjligöra svårare frågeställningar. Tekniken har en del problem t.ex. med att bearbeta spatial data och att den är svår att underhålla. Dock är dessa problem inte graverande nog att stoppa investeringar i den. Tekniken behöver förbättringar och mer initiativ till bearbetning av den.
APA, Harvard, Vancouver, ISO, and other styles
31

Masini, Fabrizio. "Progettazione e realizzazione di un'applicazione per l'estrazione, l'analisi e l'integrazione di social data e linked open data." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2015. http://amslaurea.unibo.it/8620/.

Full text
Abstract:
Studio ed analisi delle principali tecniche in ambito di Social Data Analysis. Progettazione e Realizzazione di una soluzione software implementata con linguaggio Java in ambiente Eclipse. Il software realizzato permette di integrare differenti servizi di API REST, per l'estrazione di dati sociali da Twitter, la loro memorizzazione in un database non-relazionale (realizzato con MongoDB), e la loro gestione. Inoltre permette di effettuare operazioni di classificazione di topic, e di analizzare dati complessivi sulle collection di dati estratti. Infine permette di visualizzare un albero delle "ricondivisioni", partendo da singoli tweet selezionati, ed una mappa geo-localizzata, contenente gli utenti coinvolti nella catena di ricondivisioni, e i relativi archi di "retweet".
APA, Harvard, Vancouver, ISO, and other styles
32

Heuss, Timm. "A framework to support developers in the integration and application of linked and open data." Thesis, University of Plymouth, 2016. http://hdl.handle.net/10026.1/8088.

Full text
Abstract:
In the last years, the number of freely available Linked and Open Data datasets has multiplied into tens of thousands. The numbers of applications taking advantage of it, however, have not. Thus, large portions of potentially valuable data remain unexploited and are inaccessible for lay users. Therefore the upfront investment in releasing data in the first place is hard to justify. The lack of applications needs to be addressed in order not to undermine efforts put into Linked and Open Data. In existing research, strong indicators can be found that the dearth of applications is due to a lack of pragmatic, working architectures supporting these applications and guiding developers. In this thesis, a new architecture for the integration and application of Linked and Open Data is presented. Fundamental design decisions are backed up by two studies: firstly, based on real-world Linked and Open Data samples, characteristic properties are identified. A key finding is the fact that large amounts of structured data display tabular structures, do not use clear licensing and involve multiple different file formats. Secondly, following on from that study, a comparison of storage choices in relevant query scenarios is made. It includes the de-facto standard storage choice in this domain, Triples Stores, as well as relational and NoSQL approaches. Results show significant performance deficiencies of some technologies in certain scenarios. Consequently, when integrating Linked and Open Data in scenarios with application-specific entities, the first choice of storage is relational databases. Combining these findings and related best practices of existing research, a prototype framework is implemented using Java 8 and Hibernate. As a proof-of-concept it is employed in an existing Linked and Open Data integration project. Thereby, it is shown that a best practice architectural component is introduced successfully, while development effort to implement specific program code can be simplified. Thus, the present work provides an important foundation for the development of semantic applications based on Linked and Open Data and potentially leads to a broader adoption of such applications.
APA, Harvard, Vancouver, ISO, and other styles
33

Evangelisti, Giacomo. "Sviluppo e utilizzo degli Open Data nella scuola." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2017.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
34

Zapilko, Benjamin [Verfasser], and York [Akademischer Betreuer] Sure-Vetter. "Methods for Matching of Linked Open Social Science Data / Benjamin Zapilko. Betreuer: York Sure-Vetter." Mannheim : Universitätsbibliothek Mannheim, 2014. http://d-nb.info/1067055312/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
35

De, Wilde Max. "From Information Extraction to Knowledge Discovery: Semantic Enrichment of Multilingual Content with Linked Open Data." Doctoral thesis, Universite Libre de Bruxelles, 2015. http://hdl.handle.net/2013/ULB-DIPOT:oai:dipot.ulb.ac.be:2013/218774.

Full text
Abstract:
Discovering relevant knowledge out of unstructured text in not a trivial task. Search engines relying on full-text indexing of content reach their limits when confronted to poor quality, ambiguity, or multiple languages. Some of these shortcomings can be addressed by information extraction and related natural language processing techniques, but it still falls short of adequate knowledge representation. In this thesis, we defend a generic approach striving to be as language-independent, domain-independent, and content-independent as possible. To reach this goal, we offer to disambiguate terms with their corresponding identifiers in Linked Data knowledge bases, paving the way for full-scale semantic enrichment of textual content. The added value of our approach is illustrated with a comprehensive case study based on a trilingual historical archive, addressing constraints of data quality, multilingualism, and language evolution. A proof-of-concept implementation is also proposed in the form of a Multilingual Entity/Resource Combiner & Knowledge eXtractor (MERCKX), demonstrating to a certain extent the general applicability of our methodology to any language, domain, and type of content.
Découvrir de nouveaux savoirs dans du texte non-structuré n'est pas une tâche aisée. Les moteurs de recherche basés sur l'indexation complète des contenus montrent leur limites quand ils se voient confrontés à des textes de mauvaise qualité, ambigus et/ou multilingues. L'extraction d'information et d'autres techniques issues du traitement automatique des langues permettent de répondre partiellement à cette problématique, mais sans pour autant atteindre l'idéal d'une représentation adéquate de la connaissance. Dans cette thèse, nous défendons une approche générique qui se veut la plus indépendante possible des langues, domaines et types de contenus traités. Pour ce faire, nous proposons de désambiguïser les termes à l'aide d'identifiants issus de bases de connaissances du Web des données, facilitant ainsi l'enrichissement sémantique des contenus. La valeur ajoutée de cette approche est illustrée par une étude de cas basée sur une archive historique trilingue, en mettant un accent particulier sur les contraintes de qualité, de multilinguisme et d'évolution dans le temps. Un prototype d'outil est également développé sous le nom de Multilingual Entity/Resource Combiner & Knowledge eXtractor (MERCKX), démontrant ainsi le caractère généralisable de notre approche, dans un certaine mesure, à n'importe quelle langue, domaine ou type de contenu.
Doctorat en Information et communication
info:eu-repo/semantics/nonPublished
APA, Harvard, Vancouver, ISO, and other styles
36

Triperina, Evangelia. "Visual interactive knowledge management for multicriteria decision making and ranking in linked open data environments." Thesis, Limoges, 2020. http://www.theses.fr/2020LIMO0010.

Full text
Abstract:
Le doctorat impliqués la recherche dans le domaine des représentations visuelles assistées par des technologies sémantiques et des ontologies afin de soutenir les décisions et les procédures d'élaboration des politiques, dans le cadre de la recherche et des systèmes d'information académique. Les visualisations seront également prises en charge par l'exploration de données et les processus d'extraction de connaissances dans l'environnement de données liées. Pour élaborer, les techniques d'analyse visuelle seront utilisées pour l'organisation des visualisations afin de présenter l'information de manière à utiliser les capacités perceptuelles humaines et aideront éventuellement les procédures de prise de décision et de prise de décision. En outre, la représentation visuelle et, par conséquent, les processus décisionnels et décisionnels seront améliorés au moyen des technologies sémantiques basées sur des modèles conceptuels sous forme d'ontologies. Ainsi, l'objectif principal de la thèse de doctorat proposée consiste en la combinaison des technologies sémantiques clés et des techniques de visualisation interactive basées principalement sur la perception du graphique afin de rendre les systèmes de prise de décision plus efficaces. Le domaine de la demande sera le système de recherche et d'information académique
The dissertation herein involves research in the field of the visual representations aided by semantic technologies and ontologies in order to support decisions and policy making procedures, in the framework of research and academic information systems. The visualizations will be also supported by data mining and knowledge extraction processes in the linked data environment. To elaborate, visual analytics’ techniques will be employed for the organization of the visualizations in order to present the information in such a way that will utilize the human perceptual abilities and that will eventually assist the decision support and policy making procedures. Furthermore, the visual representation and consequently the decision and policy making processes will be ameliorated by the means of the semantic technologies based on conceptual models in the form of ontologies. Thus, the main objective of the proposed doctoral thesis consists the combination of the key semantic technologies with interactive visualisations techniques based mainly on graph’s perception in order to make decision support systems more effective. The application field will be the research and academic information systems
APA, Harvard, Vancouver, ISO, and other styles
37

Marcelli, Fulvio. "Estrazione automatica di informazioni da articoli scientifici in formato PDF e pubblicazione in Linked Open Data." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2016. http://amslaurea.unibo.it/10892/.

Full text
Abstract:
La tesi ha lo scopo di introdurre Investiga, un'applicazione per l'estrazione automatica di informazioni da articoli scientifici in formato PDF e pubblicazione di queste informazioni secondo i principi e i formati Linked Open Data, creata per la tesi. Questa applicazione è basata sul Task 2 della SemPub 2016, una challenge che ha come scopo principale quello di migliorare l'estrazione di informazioni da articoli scientifici in formato PDF. Investiga estrae i capitoli di primo livello, le didascalie delle figure e delle tabelle da un dato articolo e crea un grafo delle informazioni così estratte collegate adeguatamente tra loro. La tesi inoltre analizza gli strumenti esistenti per l'estrazione automatica di informazioni da documenti PDF e dei loro limiti.
APA, Harvard, Vancouver, ISO, and other styles
38

Sangiorgi, Luca. "Analisi e progettazione di un sistema di link discovery semi-automatico nell'ambito dei Linked Open Data." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2015. http://amslaurea.unibo.it/8194/.

Full text
Abstract:
Lo scopo di questo elaborato è di analizzare e progettare un sistema in grado di supportare la definizione dei dati nel formato utilizzato per definire in modo formale la semantica dei dati, ma soprattutto nella complessa e innovativa attività di link discovery. Una attività molto potente che, tramite gli strumenti e le regole del Web Semantico (chiamato anche Web of Data), permette data una base di conoscenza sorgente ed altre basi di conoscenza esterne e distribuite nel Web, di interconnettere i dati della base di conoscenza sorgente a quelli esterni sulla base di complessi algoritmi di interlinking. Questi algoritmi fanno si che i concetti espressi sulla base di dati sorgente ed esterne vengano interconnessi esprimendo la semantica del collegamento ed in base a dei complessi criteri di confronto definiti nel suddetto algoritmo. Tramite questa attività si è in grado quindi di aumentare notevolmente la conoscenza della base di conoscenza sorgente, se poi tutte le basi di conoscenza presenti nel Web of Data seguissero questo procedimento, la conoscenza definita aumenterebbe fino a livelli che sono limitati solo dalla immensa vastità del Web, dando una potenza di elaborazione dei dati senza eguali. Per mezzo di questo sistema si ha l’ambizioso obiettivo di fornire uno strumento che permetta di aumentare sensibilmente la presenza dei Linked Open Data principalmente sul territorio nazionale ma anche su quello internazionale, a supporto di enti pubblici e privati che tramite questo sistema hanno la possibilità di aprire nuovi scenari di business e di utilizzo dei dati, dando una potenza al dato che attualmente è solo immaginabile.
APA, Harvard, Vancouver, ISO, and other styles
39

Owusu-Banahene, Wiafe. "Evaluating different options of integrating linked open data into standard geospatial web services for thematic mapping." Thesis, University of Pretoria, 2015. http://hdl.handle.net/2263/53549.

Full text
Abstract:
The open data movement has resulted in many datasets on the web to be freely available for anyone to freely access, use, modify and share for any purpose (subject, at most, to requirements that preserve provenance and openness). The Linked Open Data (LOD) cloud is an example of such an open data repository of attribute data in the form of billions of Resource Description Framework triples on the Web. Visualising such open data in thematic maps provides a powerful spatial analysis tool for planning and decision-making. In this research, several styles of creating web thematic maps by integrating the attributes from the LOD cloud with geometry in a spatial database server were investigated and evaluated. Requirements for a specialised geospatial web service that combines linked data with geospatial data to create thematic maps were specified. Standard technologies were used, motivated by the widespread deployment of standardised web map services in the geospatial community and the widespread publication of alphanumeric data (by statistical agencies) in the LOD cloud. A specialisation of an Open Geospatial Consortium Web Map Service (WMS) that creates web thematic maps by integrating alphanumeric linked data from the LOD cloud with geometry in a spatial database server was conceptualised. Three integration styles (referred to as design options) for this specialised web service were designed and implemented. The first style integrates linked data with spatial data by an importer. The second and third styles use a middleware and extension of a spatial database server respectively to integrate linked data with spatial data. In each of the three styles, attributes are retrieved from the LOD cloud through semantic queries and only the results of the semantic query are visualised on the thematic map. In this way, the benefits of semantic queries are exploited in the Semantic Web itself and the WMS mapping capabilities are used to visualise the semantic query results on a thematic map by integrating these with geospatial data. The three integration styles are critically evaluated against the specified requirements. This research contributes to understanding the pros and cons of incorporating semantic (linked) data models into standard geospatial web service models to create cartographic products (web thematic maps). This research contributes to bridging the gap between linked data and web thematic mapping.
Thesis (PhD)--University of Pretoria, 2015.
Geography, Geoinformatics and Meteorology
PhD
Unrestricted
APA, Harvard, Vancouver, ISO, and other styles
40

Ferdan, Ondřej. "Možnosti zpracování a využití otevřených dat." Master's thesis, Česká zemědělská univerzita v Praze, 2016. http://www.nusl.cz/ntk/nusl-259894.

Full text
Abstract:
Main goal of this diploma thesis is characterization of open data, standards and analyzation of adoption and utilization of open principles in the public sector of the Czech Republic. And comparison with European Union and chosen countries. Identifies technology and tools for linked data, used for deployment of highest rating of data openness. Defines geographical data, its standards and INSPIRE directive for spatial information in Europe. The goal of practical part of thesis is to analyze adoption of open principles for geographical data between Czech institutions. Focusing on what data are available, if open principles are applied and on what circumstances are data available. Foreign countries are also covered for the comparison.
APA, Harvard, Vancouver, ISO, and other styles
41

Lanzarone, Lorenzo Biagio. "Teoria e pratica degli open data: rassegna ed implementazione in un portale web turistico." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2018. http://amslaurea.unibo.it/16720/.

Full text
Abstract:
I dati sono sempre più importanti nella società e rappresentano una fondamentale risorsa. Da sempre l’essere umano ha la necessità di organizzare le informazioni e utilizzarle per creare nuova conoscenza. Nella presente tesi si approfondisce la tematica degli open data, cioè dati messi a disposizione da enti di vario genere, per essere utilizzati liberamente da chiunque per qualsiasi scopo. Nella prima parte dell’elaborato si approfondiscono origini e aspetti teorici degli open data, mentre nella seconda gli aspetti pratici. Quanto descritto in quest’ultima parte è stato svolto durante un tirocinio presso l’agenzia web Magic di Funo di Argelato (Bologna). Il progetto riguarda la realizzazione di un esempio pratico di applicazione degli open data, in particolare la progettazione e implementazione degli stessi all’interno del portale web turistico OkBed&Breakfast.
APA, Harvard, Vancouver, ISO, and other styles
42

Klimek, Bettina, Natanael Arndt, Sebastian Krause, and Timotheus Arndt. "Creating Linked Data morphological language resources with MMoOn: the Hebrew Morpheme Inventory." Universität Leipzig, 2016. https://ul.qucosa.de/id/qucosa%3A15774.

Full text
Abstract:
The development of standard models for describing general lexical resources has led to the emergence of numerous lexical datasets of various languages in the Semantic Web. However, there are no models that describe the domain of Morphology in a similar manner. As a result, there are hardly any language resources of morphemic data available in RDF to date. This paper presents the creation of the Hebrew Morpheme Inventory from a manually compiled tabular dataset comprising around 52.000 entries. It is an ongoing effort of representing the lexemes, word-forms and morphologigal patterns together with their underlying relations based on the newly created Multilingual Morpheme Ontology (MMoOn). It will be shown how segmented Hebrew language data can be granularly described in a Linked Data format, thus, serving as an exemplary case for creating morpheme inventories of any inflectional language with MMoOn. The resulting dataset is described a) according to the structure of the underlying data format, b) with respect to the Hebrew language characteristic of building word-forms directly from roots, c) by exemplifying how inflectional information is realized and d) with regard to its enrichment with external links to sense resources.
APA, Harvard, Vancouver, ISO, and other styles
43

Kovář, Jiří. "Využití „Open Dynamics Engine“ pro modelování mobilních robotů." Master's thesis, Vysoké učení technické v Brně. Fakulta strojního inženýrství, 2008. http://www.nusl.cz/ntk/nusl-227991.

Full text
Abstract:
This diploma thesis deals with the problems of virtual physical modelling of mobile robots for the needs of their real-time control. To create a virtual physical world, an open-source project OPEN DYNAMICS ENGINE (ODE) was used, the results were displayed facilitating DirectX graphical interface. Simulated systems in ODE were written in C# on Microsoft.NET platform. The properites and qualities in ODE were verified by simulation in several types of simple systems and on a simplified robot model "Kracmera I.". Subsequently, the usability of ODE for its control was being verified.
APA, Harvard, Vancouver, ISO, and other styles
44

Frank, Matthias T. [Verfasser], and Y. [Akademischer Betreuer] Sure-Vetter. "Knowledge-Driven Harmonization of Sensor Observations: Exploiting Linked Open Data for IoT Data Streams / Matthias T. Frank ; Betreuer: Y. Sure-Vetter." Karlsruhe : KIT Scientific Publishing, 2021. http://d-nb.info/1235140989/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
45

Gonano, Ciro Mattia. "Un'esperienza di rappresentazione di dati di cataloghi digitali in Linked Open Data: il caso della Fondazione Zeri." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2014. http://amslaurea.unibo.it/7257/.

Full text
Abstract:
Il documento presenta il caso dell'archivio fotografico della Fondazione Zeri come caso reale di conversione di un catalogo ricco di informazioni ma povero di interconnessioni nel dominio dei Linked Open Data, basandosi sull'ontologia CIDOC-CRM per il cultural heritage.
APA, Harvard, Vancouver, ISO, and other styles
46

Nishanbaev, Ikrom. "A Web GIS-based Integration of 3D Digital Models with Linked Open Data for Cultural Heritage Exploration." Thesis, Curtin University, 2021. http://hdl.handle.net/20.500.11937/88578.

Full text
Abstract:
This PhD project explores how geospatial semantic web concepts, 3D web-based visualisation, digital interactive map, and cloud computing concepts could be integrated to enhance digital cultural heritage exploration; to offer long-term archiving and dissemination of 3D digital cultural heritage models; to better interlink heterogeneous and sparse cultural heritage data. The research findings were disseminated via four peer-reviewed journal articles and a conference article presented at GISTAM 2020 conference (which received the ‘Best Student Paper Award’).
APA, Harvard, Vancouver, ISO, and other styles
47

Wenige, Lisa [Verfasser], Johannes M. [Gutachter] Ruhland, and Nils [Gutachter] Boysen. "On-the-fly recommendation retrieval from linked open data repositories / Lisa Wenige ; Gutachter: Johannes M. Ruhland, Nils Boysen." Jena : Friedrich-Schiller-Universität Jena, 2019. http://d-nb.info/1177387026/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
48

Sherif, Mohamed Ahmed Mohamed. "Automating Geospatial RDF Dataset Integration and Enrichment." Doctoral thesis, Universitätsbibliothek Leipzig, 2016. http://nbn-resolving.de/urn:nbn:de:bsz:15-qucosa-215708.

Full text
Abstract:
Over the last years, the Linked Open Data (LOD) has evolved from a mere 12 to more than 10,000 knowledge bases. These knowledge bases come from diverse domains including (but not limited to) publications, life sciences, social networking, government, media, linguistics. Moreover, the LOD cloud also contains a large number of crossdomain knowledge bases such as DBpedia and Yago2. These knowledge bases are commonly managed in a decentralized fashion and contain partly verlapping information. This architectural choice has led to knowledge pertaining to the same domain being published by independent entities in the LOD cloud. For example, information on drugs can be found in Diseasome as well as DBpedia and Drugbank. Furthermore, certain knowledge bases such as DBLP have been published by several bodies, which in turn has lead to duplicated content in the LOD . In addition, large amounts of geo-spatial information have been made available with the growth of heterogeneous Web of Data. The concurrent publication of knowledge bases containing related information promises to become a phenomenon of increasing importance with the growth of the number of independent data providers. Enabling the joint use of the knowledge bases published by these providers for tasks such as federated queries, cross-ontology question answering and data integration is most commonly tackled by creating links between the resources described within these knowledge bases. Within this thesis, we spur the transition from isolated knowledge bases to enriched Linked Data sets where information can be easily integrated and processed. To achieve this goal, we provide concepts, approaches and use cases that facilitate the integration and enrichment of information with other data types that are already present on the Linked Data Web with a focus on geo-spatial data. The first challenge that motivates our work is the lack of measures that use the geographic data for linking geo-spatial knowledge bases. This is partly due to the geo-spatial resources being described by the means of vector geometry. In particular, discrepancies in granularity and error measurements across knowledge bases render the selection of appropriate distance measures for geo-spatial resources difficult. We address this challenge by evaluating existing literature for point set measures that can be used to measure the similarity of vector geometries. Then, we present and evaluate the ten measures that we derived from the literature on samples of three real knowledge bases. The second challenge we address in this thesis is the lack of automatic Link Discovery (LD) approaches capable of dealing with geospatial knowledge bases with missing and erroneous data. To this end, we present Colibri, an unsupervised approach that allows discovering links between knowledge bases while improving the quality of the instance data in these knowledge bases. A Colibri iteration begins by generating links between knowledge bases. Then, the approach makes use of these links to detect resources with probably erroneous or missing information. This erroneous or missing information detected by the approach is finally corrected or added. The third challenge we address is the lack of scalable LD approaches for tackling big geo-spatial knowledge bases. Thus, we present Deterministic Particle-Swarm Optimization (DPSO), a novel load balancing technique for LD on parallel hardware based on particle-swarm optimization. We combine this approach with the Orchid algorithm for geo-spatial linking and evaluate it on real and artificial data sets. The lack of approaches for automatic updating of links of an evolving knowledge base is our fourth challenge. This challenge is addressed in this thesis by the Wombat algorithm. Wombat is a novel approach for the discovery of links between knowledge bases that relies exclusively on positive examples. Wombat is based on generalisation via an upward refinement operator to traverse the space of Link Specifications (LS). We study the theoretical characteristics of Wombat and evaluate it on different benchmark data sets. The last challenge addressed herein is the lack of automatic approaches for geo-spatial knowledge base enrichment. Thus, we propose Deer, a supervised learning approach based on a refinement operator for enriching Resource Description Framework (RDF) data sets. We show how we can use exemplary descriptions of enriched resources to generate accurate enrichment pipelines. We evaluate our approach against manually defined enrichment pipelines and show that our approach can learn accurate pipelines even when provided with a small number of training examples. Each of the proposed approaches is implemented and evaluated against state-of-the-art approaches on real and/or artificial data sets. Moreover, all approaches are peer-reviewed and published in a conference or a journal paper. Throughout this thesis, we detail the ideas, implementation and the evaluation of each of the approaches. Moreover, we discuss each approach and present lessons learned. Finally, we conclude this thesis by presenting a set of possible future extensions and use cases for each of the proposed approaches.
APA, Harvard, Vancouver, ISO, and other styles
49

MANZELLA, Michele. "Materialising Resilience. Open data management, flood risk assessment and operative actions for dynamic cities." Doctoral thesis, Università degli studi di Ferrara, 2016. http://hdl.handle.net/11392/2403299.

Full text
Abstract:
Obiettivo di questa ricerca è indagare nuove possibilità in architettura e progettazione urbana consentite dal progresso nella mappatura digitale. La consapevolezza dei cambiamenti ambientali e la crescente capacità di utilizzare dati in tempo reale facilitano soluzioni tramite il processo di progettazione. Data l’importanza delle nuove tecnologie nell’aiutare i processi di pianificazione, è sorprendente che gli enti governativi non abbiano investito in maniera sufficiente per tentare di connettere la realtà digitale a quella fisica. L’architettura della città dipende solo dal considerare il territorio nel suo insieme, e le mappe come piattaforme su cui studiare future azioni. L’avanzamento dell’analisi di dati comporta la progressiva facilitazione dell’estrazione ed utilizzo di informazioni; di conseguenza, le mappature tradizionali non soddisfano più i bisogni attuali. Se la correlazione tra i dati è cambiata—e di conseguenza sono cambiate anche le mappe—le politiche di gestione spaziale devono essere alterate come diretta conseguenza. La capacità di produrre cambiamenti istantanei sulle mappe riduce il divario tra mondo fisico e digitale e, dall’unione di queste due realtà, è possibile cambiare la comprensione delle dinamiche urbane. Inoltre, la direzione che la pianificazione strategica dovrebbe intraprendere può essere meglio analizzata. Il fine è di fornire agli architetti una piattaforma multi-scalare che consenta l’impiego di un unico sistema per l’accesso alla molteplicità delle informazioni. La Telematic Map of Risk svela nuove relazioni fra gli edifici e il loro contesto. Come mappa basata su sistemi GIS, è uno strumento di gestione di ogni tipo di rischio finalizzato a supportare il processo decisionale delle Autorità Locali, per la formazione di società che resistano ad eventi catastrofici. L’uso della mappa digitale consente la progettazione in un ambiente non più statico e garantisce lo sviluppo infrastrutturale in un contesto urbano che cambia col tempo. Creando un sistema ibrido che opera tra la pianificazione territoriale e la crescita spontanea della città, le mappature consentono agli architetti di creare un quadro coerente—basato sui Cluster Plans e le Guidelines for Territory Planning—all’interno del quale gli edifici possono essere progettati nella loro complessità.
The aim of this research is to explore new possibilities in architecture and urban design allowed by advancements in digital mapping. Our awareness of changes to the environment and increasing ability to utilise real-time data allow us to respond to these changes through the design process. Given the importance of new technologies in aiding the planning process, it is surprising that Governmental Agencies have been doing very little work trying to connect virtual and physical realities. The architecture of the city depends only on considering the territory as a whole, and maps are the platform on which it is possible to study future actions.   As big data analytics become more advanced the information becomes easier to extract and utilise. Therefore, traditional maps cannot satisfy current needs any longer. If the correlations between data have changed, and consequentially maps too, politics of spatial management must be modified as a direct consequence. The ability to make instantaneous changes to maps reduces the gap between physical and digital domains. Through the interweaving of these two realms, one could change the awareness of cities’ dynamics. Moreover, the direction that strategic planning should take can be better understood.   The goal is to provide architects a multi-scalar platform allowing them to use one single dashboard to access multiple layers of information. The Telematic Map of Risk unveils new relationships and connections between buildings and their context. As GIS-based map, it shall be considered as a multi-hazard tool aimed to form and support decision-making processes of Local Planning Authorities, to foster disaster resilient societies. Using the digital map, architects no longer design within a fixed environment, but create infrastructures that allow the urban fabric to change over time. By creating a hybrid system of design that positions itself between top down urban planning and bottom up city growth, the maps will allow architects to create an organised framework—based on the Cluster Plans and the Guidelines for Territory Planning—within which buildings can be designed.
APA, Harvard, Vancouver, ISO, and other styles
50

Haller, Armin, Javier D. Fernández, Maulik R. Kamdar, and Axel Polleres. "What are Links in Linked Open Data? A Characterization and Evaluation of Links between Knowledge Graphs on the Web." Department für Informationsverarbeitung und Prozessmanagement, WU Vienna University of Economics and Business, 2019. http://epub.wu.ac.at/7193/1/20191002ePub_LOD_link_analysis.pdf.

Full text
Abstract:
Linked Open Data promises to provide guiding principles to publish interlinked knowledge graphs on the Web in the form of findable, accessible, interoperable and reusable datasets. We argue that while as such, Linked Data may be viewed as a basis for instantiating the FAIR principles, there are still a number of open issues that cause significant data quality issues even when knowledge graphs are published as Linked Data. Firstly, in order to define boundaries of single coherent knowledge graphs within Linked Data, a principled notion of what a dataset is, or, respectively, what links within and between datasets are, has been missing. Secondly, we argue that in order to enable FAIR knowledge graphs, Linked Data misses standardised findability and accessability mechanism, via a single entry link. In order to address the first issue, we (i) propose a rigorous definition of a naming authority for a Linked Data dataset (ii) define different link types for data in Linked datasets, (iii) provide an empirical analysis of linkage among the datasets of the Linked Open Data cloud, and (iv) analyse the dereferenceability of those links. We base our analyses and link computations on a scalable mechanism implemented on top of the HDT format, which allows us to analyse quantity and quality of different link types at scale.
Series: Working Papers on Information Systems, Information Business and Operations
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography