Paper accepted at EPIA 2019

We are very pleased to announce that our group got a paper accepted for presentation at the EPIA 2019: 19th EPIA Conference on Artificial Intelligence, which will be held on Sep 3 – 6, 2019 Vila Real, Portugal.  

The EPIA Conference on Artificial Intelligence is a well-established European conference in the field of AI. The 19th edition, EPIA 2019, will take place at UTAD University, Vila Real in September 3rd-6th, 2019. As in previous editions, this international conference is hosted with the patronage of the Portuguese Association for Artificial Intelligence (APPIA). The purpose of this conference is to promote research in all Artificial Intelligence (AI) areas, covering both theoretical/foundational issues and applications, as well as the scientific exchange among researchers, engineers and practitioners in related disciplines. 

Here is the pre-print of the accepted paper with its abstract: 

Abstract: The information on the internet suffers from noise and corrupt knowledge that may arise due to human and mechanical errors. To further exacerbate this problem, an ever-increasing amount of fake news on social media, or the internet in general, has created another challenge to drawing correct information from the web. This huge sea of data makes it difficult for human fact checkers and journalists to assess all the information manually. In recent years automated Fact-Checking has emerged as a branch of natural language processing devoted to achieving this feat. In this work, we give an overview of recent work, emphasizing on the key challenges faced during the development of such frameworks. We benchmark existing solutions to perform claim classification and introduce a new model dubbed SimpleLSTM, which outperforms baseline by 11%, 10.2% and 18.7% on FEVER-Support, FEVER-Reject, and 3-Class datasets respectively. The data, metadata, and code are released as open-source and are available at https://github.com/DeFacto/SimpleLSTM.

Acknowledgment

This work was partially funded by the European Union Marie Curie ITN Cleopatra project (GA no. 812997). 

Looking forward to seeing you at The EPIA 2019

Paper accepted at DEXA 2019

dexa2019-logoWe are very pleased to announce that our group got a paper accepted for presentation at the DEXA 2019: 30th International Conference on Database and Expert Systems Applications, which will be held on August 26 – 29, 2019, Linz, Austria.  

DEXA provides a forum to present research results and to examine advanced applications in the field. The conference and its associated workshops offer an opportunity for developers, scientists, and users to extensively discuss requirements, problems, and solutions in database, information, and knowledge systems. 

Here is the pre-print of the accepted paper with its abstract: 

Abstract: Context-specific description of entities –expressed in RDF– poses challenges during data-driven tasks, e.g., data integration, and context-aware entity matching represents a building-block for these tasks. However, existing approaches only consider inter-schema mapping of data sources, and are not able to manage several contexts during entity matching. We devise COMET, an entity matching technique that relies on both the knowledge stated in RDF vocabularies and context-based similarity metrics to match \textit{contextually equivalent} entities. COMET executes a novel 1-1 perfect matching algorithm for matching contextually equivalent entities based on the combined scores of semantic similarity and context similarity. COMET employs the Formal Concept Analysis algorithm in order to compute the context similarity of RDF entities. We empirically evaluate the performance of COMET on a testbed from DBpedia. The experimental results suggest that COMET is able to accurately match equivalent RDF graphs in a context-dependent manner.

Acknowledgment
This work was partially funded by the European project QualiChain (GA~822404).

Looking forward to seeing you at The DEXA 2019.

Workshop paper accepted at ICAIL/AIAS 2019

aias19We are very pleased to announce that our group got a paper accepted for presentation at the AIAS 2019: The First Workshop on Artificial Intelligence and the Administrative State, which was held in conjunction with the 17th International Conference on AI and Law (ICAIL 2019) on Monday, 17 June 2019 in Montréal, Québec, Canada. 

Recent advances in AI, Machine Learning, Human Language Technology, Network Science, and Human Factors analysis offer promising new approaches to improving the ability of all stakeholders, including agencies themselves, to operate within this complex regulatory environment. The scale of administrative states means that the benefits of automation have very high potential impact, both in improvements to government processes and in the delivery of services and benefits to citizens. At the same time, the black-box nature of many automated decision-making systems, particularly sub-symbolic AI components such as those generated by machine learning algorithms, can create considerable tension with the norms of transparency, accountability, and reason-giving that typically govern administrative action. Explainable, responsible, and trustworthy AI is vital for addressing these factors. 

Here is the pre-print of the accepted paper with its abstract: 

Abstract: The ubiquitous availability of online services and mobile apps results in a rapid proliferation of contractual agreements in the form of privacy policies. Despite the importance of such consent forms, the majority of users tend to ignore them due to their content length and complexity. Thus, users might be consenting policies that are not aligned to regulations in laws such as the GDPR from the EU law. In this study, we propose a hybrid approach which measures a privacy policy’s risk factor applying both supervised deep learning and rule-based information extraction. Benefiting from an annotated dataset of 115 privacy policies, a deep learning component is first able to predict high-level categories for each paragraph. Then, a rule-based module extracts pre-defined attributes and their values, based on high-level classes. Finally, a privacy policy’s risk factor is computed based on these attribute values.

SDA at ESWC 2019 and a Best Demo Award

eswc2019The Extended Semantic Web Conference (ESWC) is a major venue for discussing the latest scientific results and technology innovations around semantic technologies. Building on its past success, ESWC is seeking to broaden its focus to span other relevant related research areas in which Web semantics plays an important role. ESWC 2019 will present the latest results in research, technologies and applications in its field. Besides the technical program organized over twelve tracks, the conference will feature a workshop and tutorial program, a dedicated track on Semantic Web challenges, system descriptions and demos, a posters exhibition and a doctoral symposium.

We are very pleased to announce that we got 2 papers accepted at ESWC 2019 for presentation at the main conference. Additionally, we also had 1 Workshop, 2 Tutorials and 1 workshop paper and 3 Posters/Demo papers accepted. In addition, we participated in the Networking Session.

Furthermore, we are very happy to announce that we won the Best Demo Award for the “MantisTable: a Tool for Creating Semantic Annotations on Tabular Data” by Marco Cremaschi, Anisa Rula, Alessandra Siano, and Flavio De Paoli.

Here are some further pointers in case you want to know more about MantisTable tool:

Among the other presentations, our colleagues presented the following presentations:

Workshops & Tutorials: 

 

ESWC’19 was a great venue to meet the community, create new connections, talk about current research challenges, share ideas and settle new collaborations. We look forward to the next ESWC conference.
Until then, meet us at SDA!

Papers accepted at ISWC 2019

iswc2019We are very pleased to announce that our group got 14 papers accepted for presentation at ISWC 2019 – the 18th International Semantic Web Conference, which will be held on October 26 – 30 2019 in Auckland, New Zealand. ISWC is an A-ranked conference (CORE ranking) and currently 11th in Google Scholar in the category “Databases & Information Systems” with an h5-index of 41 as well as 4th in terms WWW related conferences in MS Academic Search.

The International Semantic Web Conference (ISWC) is the premier international forum where Semantic Web / Linked Data / Knowledge Graph researchers, practitioners, and industry specialists come together to discuss, advance, and shape the future of semantic technologies on the web, within enterprises and in the context of the public institution.

Here is the list of the accepted papers with their abstract: 

Abstract: Over the last years, Linked Data has grown continuously.  Today, we count more than 10,000 datasets being available online following Linked Data standards.  These standards allow data to be machine readable and inter-operable. Nevertheless, many applications, such as data integration, search, and interlinking, cannot take full advantage of Linked Data if it is of low quality. There exist a few approaches for the quality assessment of Linked Data, but their performance degrades with the increase in data size and quickly grows beyond the capabilities of a single machine. In this paper, we present DistQualityAssessment — an open source implementation of quality assessment of large RDF datasets that can scale out to a cluster of machines. This is the first distributed, in-memory approach for computing different quality metrics for large RDF datasets using Apache Spark. We also provide a quality assessment pattern that can be used to generate new scalable metrics that can be applied to big data. The work presented here is integrated with the SANSA framework and has been applied to at least three use cases beyond the SANSA community. The results show that our approach is more generic, efficient, and scalable as compared to previously proposed approaches.

 

Abstract: One of the key traits of Big Data is its complexity in terms of representation, structure, or formats. One existing way to deal with it is offered by Semantic Web standards. Among them, RDF –which proposes to model data with triples representing edges in a graph– has received a large success and the semantically annotated data has grown steadily towards a massive scale. Therefore, there is a need for scalable and efficient query engines capable of retrieving such information. In this paper, we propose \emph{Sparklify}: a scalable software component for efficient evaluation of SPARQL queries over distributed RDF datasets. It uses Sparqlify as a SPARQL-to-SQL rewriter for translating SPARQL queries into Spark executable code. Our preliminary results demonstrate that our approach is more extensible, efficient, and scalable as compared to state-of-the-art approaches. Sparklify is integrated into a larger SANSA framework and it serves as a default query engine and has been used by at least three external use scenarios.

Abstract: The  last two  decades witnessed  a remarkable evolution  in terms of data formats,  modalities, and storage capabilities.  Instead of having to adapt one’s application needs to the, earlier limited, available storage options, today there is a wide array of options to choose from to best meet an application’s needs. This has resulted in vast amounts of data available in a variety of forms and formats which, if interlinked and jointly  queried, can generate valuable knowledge and insights. In this article, we describe Squerall: a framework that builds on the principles of Ontology-Based Data Access (OBDA) to enable the querying of disparate heterogeneous sources using a unique query language, SPARQL. In Squerall, original data is queried on-the-fly without prior data materialization or transformation. In particular, Squerall allows the aggregation and joining of large data in a distributed manner. Squerall supports out-of-the-box five data sources and moreover, it can be programmatically extended to cover more sources and incorporate new query engines. The framework provides user interfaces for the creation of necessary inputs, as well as guiding non-SPARQL experts to write SPARQL queries. Squerall is integrated into the popular SANSA stack and available as open-source software via GitHub and as a Docker image

Abstract: Relation linking is an important problem for knowledge graph-based Question Answering. Given a natural language question and a knowledge graph, the task is to identify relevant relations from the given knowledge graph. Since existing techniques for entity extraction and linking are more stable compared to relation linking, our idea is to exploit entities extracted from the question to support relation linking. In this paper, we propose a novel approach, based on DBpedia entities, for computing relation candidates. We have empirically evaluated our approach on different standard benchmarks. Our evaluation shows that our approach significantly outperforms existing baseline systems in both recall, precision and runtime.

Abstract: Over the last years, a number of Linked Data-based Question Answering (QA) systems have been developed. Consequently, the series of Question Answering Over Linked Data (QALD1–QALD9) challenges and other datasets have been proposed to evaluate these systems. However, the QA datasets contain a fixed number of natural language questions and do not allow users to generate micro benchmarks tailored towards specific use-cases. We propose QaldGen, a natural language benchmark generation framework for Knowledge Graphs which is able to generate customised QA benchmark from existing QA repositories. The framework is flexible enough to generate benchmarks of varying sizes and according to the user-defined criteria on the most important features to be considered for QA benchmarking. This is achieved using different clustering algorithms. We compare state-of-the-art QA systems over knowledge graphs by using different QA benchmarks. The observed results show that specialised micro-benchmarking is important to pinpoint the limitations of the various components of QA systems.

Abstract: Knowledge graphs are composed of different elements: entity nodes, relation edges, and literal nodes. Each literal node contains an entity’s attribute value (e.g. the height of an entity of type person) and thereby encodes information which in general cannot be represented by relations between entities alone. However, most of the existing embedding or latent-feature-based methods for knowledge graph analysis only consider entity nodes and relation edges, and thus do not take the information provided by literals into account. In this paper, we extend existing latent feature methods for link prediction by a simple portable module for incorporating literals, which we name LiteralE. Unlike in concurrent methods where literals are incorporated by adding a literal-dependent term to the output of the scoring function and thus only indirectly affect the entity embeddings, LiteralE directly enriches these embeddings with information from literals via a learnable parameterized function. This function can be easily integrated into the scoring function of existing methods and learned along with the entity embeddings in an end-to-end manner. In an extensive empirical study over three datasets, we evaluate LiteralE-extended versions of various state-of-the-art latent feature methods for link prediction and demonstrate that LiteralE presents an effective way to improve their performance. For these experiments, we augmented standard datasets with their literals, which we publicly provide as testbeds for further research. Moreover, we show that LiteralE leads to a qualitative improvement of the embeddings and that it can be easily extended to handle literals from different modalities.

Abstract: The growing interest in free and open-source software which occurred  over the last decades has accelerated the usage of versioning systems to help developers collaborating together in the same projects. As a consequence, specific tools such as git and specialized open-source on-line  platforms gained importance. In this study, we introduce and share SemanGit which provides a resource at the crossroads of both Semantic Web and git web-based version control systems. SemanGit is actually the first collection of linked data extracted from GitHub based on a git ontology we designed and extended to include specific GitHub features. In this article, we present the dataset, describe the extraction process according to the ontology, show some promising analyses of the data and outline how SemanGit could be linked with external datasets or enriched with new sources to allow for more complex analyses.

Abstract: Scientific events have become a key factor of scholarly com- munication for many scientific domains. They are considered as the focal point for establishing scientific relations between scholarly objects such as people (e.g., chairs, participants), places (e.g., location), actions (e.g., roles  of participants), and artifacts (e.g., proceedings) in the scholarly communication domain. Metadata of scientific events have been made available in unstructured or semi-structured formats, which hides the interconnected and complex relationships between them and prevents transparency. To facilitate the management of such metadata, the repres- entation of event-related information in an interoperable form requires a  uniform conceptual modeling. The Scientific Events Ontology (OR-SEO) has been engineered to represent metadata of scientific events. We describe a systematic redesign of the information model that is used as a schema for the event pages of the OpenResearch.org community wiki, reusing well-known vocabularies to make OR-SEO interoperable in different contexts. OR-SEO is now in use on thousands of Open- Research.org events pages, which enables users to represent structured knowledge about events without tackling technical implementation challenges and ontology development.

Abstract: There is an emerging trend of embedding knowledge graphs (KGs) in continuous vector spaces in order to use those for machine learning tasks. Recently, many knowledge graph embedding (KGE) models have been proposed that learn low dimensional representations while trying to maintain the structural properties of the KGs such as the similarity of nodes depending on their edges to other nodes. KGEs can be used to address tasks within KGs such as the prediction of novel links and the disambiguation of entities. They can also be used for downstream tasks like question answering and fact-checking. Overall, these tasks are relevant for the semantic web community. Despite their popularity, the reproducibility of KGE experiments and the transferability of proposed KGE models to research fields outside the machine learning community can be a major challenge. Therefore, we present the KEEN Universe, an ecosystem for knowledge graph embeddings that we have developed with a strong focus on reproducibility and transferability. The KEEN Universe currently consists of the Python packages PyKEEN (Python KnowlEdge Graph EmbeddiNgs), BioKEEN (Biological KnowlEdge Graph EmbeddiNgs), and the KEEN Model Zoo for sharing trained KGE models with the community.

Abstract: The data quality improvement of DBpedia has been in the focus of many publications in the past years with topics covering both knowledge enrichment techniques such as type learning, taxonomy generation, interlinking as well as error detection strategies such as property or value outlier detection, type checking, ontology constraints, or unit-tests, to name just a few. The concrete innovation of the DBpedia FlexiFusion workflow, leveraging the novel DBpedia PreFusion dataset, which we present in this paper, is to massively cut down the engineering workload to apply any of the vast methods available in shorter time and also make it easier to produce customized knowledge graphs or DBpedias. While FlexiFusion is flexible to accommodate other use cases, our main use case in this paper is the generation of richer, language-specific DBpedias for the 20+ DBpedia chapters, which we demonstrate on the Catalan DBpedia. In this paper, we define a set of quality metrics and evaluate them for Wikidata and DBpedia datasets of several language chapters. Moreover, we show that an implementation of FlexiFusion, performed on the proposed PreFusion dataset, increases data size, richness as well as quality in comparison to the source datasets.

Abstract: Answering simple questions over knowledge graphs is a well-studied problem in question answering. Previous approaches for this task built on recurrent and convolutional neural networks (RNNs and CNNs) based architectures that use pretrained word embeddings. It was recently shown that a pretrained transformer network (BERT) can outperform RNN- and CNN based approaches on various natural language processing tasks. In this work, we investigate how well network BERT performs on the entity span prediction and relation prediction subtasks of simple QA. In addition, we provide an evaluation of both BERT and BiLSTM-based models in limited data scenarios.

Abstract: Providing machines with the capability of exploring knowledge graphs and answering natural language questions has been an active area of research over the past decade. In this direction translating natural language questions to formal queries has been one of the key approaches. To advance the research area, several datasets like WebQuestions, QALD and LCQuAD have been published in the past. The biggest data set available for complex questions (LCQuAD) over  knowledge graphs contains five thousand questions. We now provide LC-QuAD 2.0 (Large-Scale Complex Question Answering Dataset) with 30,000 questions, their paraphrases and their corresponding SPARQL queries. LC-QuAD 2.0 is compatible with both Wikidata and DBpedia 2018 knowledge graphs. In this article, we explain how the dataset was created and the variety of questions available with corresponding examples. We further provide a statistical analysis of the dataset. 

Abstract: Non-goal oriented, generative dialogue systems lack the ability to generate answers with grounded facts. A knowledge graph can be considered an abstraction of the real world consisting of well-grounded facts. This paper addresses the problem of generating well grounded responses by integrating knowledge graphs into the dialogue systems response generation process, in an end-to-end manner. A dataset for non-goal oriented dialogues is proposed in this paper in the domain of soccer, conversing on different clubs and national teams along with a knowledge graph for each of these teams. A novel neural network architecture is also proposed as a baseline on this dataset, which can integrate knowledge graphs into the response generation process, producing well articulated, knowledge grounded responses. Empirical evidence suggests that the proposed model performs better than other state-of-the-art models for knowledge graph integrated dialogue systems.

Abstract: In this paper, we conduct an empirical investigation of neural query graph ranking approaches for the task of complex question answering over knowledge graphs. We propose a novel self-attention based slot matching model which exploits the inherent structure of query graphs, our logical form of choice. Our proposed model generally outperforms other ranking models on two QA datasets over the DBpedia knowledge graph, evaluated in different settings. We also show that domain adaption and pre-trained language model based transfer learning yield improvements, effectively offsetting the general lack of training data.

Acknowledgment

This work was partly supported by the EU Horizon2020 projects BigDataOcean (GA no. 732310), Boost4.0 (GA no. 780732), QROWD (GA no. 723088), SLIPO (GA no. 731581), BETTER (GA 776280), QualiChain (GA 822404), CLEOPATRA (GA no. 812997), LIMBO (Grant no. 19F2029I), OPAL (no. 19F2028A), KnowGraphs (no. 860801), SOLIDE (no. 13N14456), Bio2Vec (grant no. 3454), LAMBDA (#809965), FAIRplus (#802750), the ERC project ScienceGRAPH (#819536), “Industrial Data Space Plus” (GA 01IS17031), Fraunhofer Cluster of Excellence “Cognitive Internet Technologies” (CCIT),  “InclusiveOCW” (grant no. 01PE17004D), the German national funded BmBF project MLwin, the National Natural Science Foundation of China (61673304) and the Key Projects of National Social Science Foundation of China(11&ZD189), EPSRC grant EP/M025268/1, WWTF grant VRG18-013, WMF-funded GlobalFactSync project, and by the ADAPT Centre for Digital Content Technology funded under the SFI Research Centres Programme (Grant 13/RC/2106) and co-funded under the European Regional Development Fund.


Looking forward to seeing you at The ISWC 2019.

Poster and Workshop papers accepted at ESWC 2019

eswc2019We are very pleased to announce that our group got 2 poster papers accepted for presentation at the ESWC 2019 : The 16th edition of the The Extended Semantic Web Conference, which will be held on June 2-6, 2019 in Portorož, Slovenia.

The ESWC is a major venue for discussing the latest scientific results and technology innovations around semantic technologies. Building on its past success, ESWC is seeking to broaden its focus to span other relevant related research areas in which Web semantics plays an important role. ESWC 2019 will present the latest results in research, technologies and applications in its field. Besides the technical program organized over twelve tracks, the conference will feature a workshop and tutorial program, a dedicated track on Semantic Web challenges, system descriptions and demos, a posters exhibition and a doctoral symposium.

Here is the pre-print of the accepted papers with their abstract:

Abstract: Among the various domains involved in large RDF graphs, applications may rely on geographical information which are often carried and presented via Points Of Interests. In particular, one challenge aims at extracting patterns from POIs sets to discover Areas Of Interest. To tackle it, an usual method is to aggregate various points according to specific distances (e.g. geographical) via clustering algorithms. In this study, we present a flexible architecture to design pipelines able to aggregate POIs from contextual to geographical dimensions in a single run. This solution allows any kind of state-of-the-art clustering algorithm combinations to compute AOIs and is built on top of a Semantic Web stack which allows multiple-source querying and filtering through SPARQL.

Abstract: Due to the rapid expansion of multilingual data on the web, developing ontology enrichment approaches has become an interesting and active subject of research. In this paper, we propose a cross-lingual matching approach for ontology enrichment (OECM) in order to enrich an ontology using another one in a different natural language. A prototype for the proposed approach has been implemented and evaluated using the MultiFarm benchmark. Evaluation results are promising and show high precision and recall compared to state-of-the-art approaches.

Furthermore, we got a paper accepted at LASCAR Workshop: 1st Workshop on Large Scale RDF Analytics, co-located with the ESWC 2019.

LASCAR Workshop seeks original “articles and posters” describing theoretical and practical methods as well as techniques for performing scalable analytics on knowledge graphs.

Here is the pre-print of the accepted paper with its abstract:

Abstract: Controlling the usage of business-critical data is essential for every company. While the upcoming age of Industry 4.0 propagates a seamless data exchange between all participating devices, facilities and companies along the production chain, the required data control mechanisms are lacking behind. We claim that for an effective protection, both access and usage control enforcement is a must-have for organizing Industry 4.0 collaboration networks. Formalized and machine-readable policies are one fundamental building block to achieve the needed trust level for real data-driven collaborations. We explain the current challenges of specifying access and usage control policies and outline respective approaches relying on Semantic Web of Things practices. We analyze the requirements and implications of existing technologies and discuss their shortcomings. Based on our experiences from the specification of the International Data Spaces Usage Control Language, the necessary next steps towards automatically monitored and enforced policies are outlined and research needs formulated.

 

Acknowledgment

This work was partially funded by the European H2020 SLIPO project (GA. 731581).

Looking forward to seeing you at The ESWC 2019.

Workshop proposal accepted at ECML 2019

ecml2019We are pleased to announce a new workshop on “New Trends in Representation Learning with Knowledge Graphs” which will be included in the program of ECML 2019: The European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Database. ECML 2019 will take place in Würzburg, Germany, from the 16th to the 20th of September 2019.

This conference is the premier European machine learning and data mining conference and builds upon over 17 years of successful events and conferences held across Europe.

Title: New Trends in Representation Learning with Knowledge Graphs

Organizers:

  • Volker Tresp (Ludwig-Maximilians University and Siemens, Germany)
  • Jens Lehmann (Bonn University and Fraunhofer IAIS, Germany)
  • Aditya Mogadala (Saarland University, Germany)
  • Achim Rettinger (Trier University, Germany)
  • Afshin Sadeghi (Fraunhofer IAIS and University of Bonn, Germany)
  • Mehdi Ali (University of Bonn and Fraunhofer IAIS, Germany)

Website: https://sites.google.com/view/kgrlfr-workshop/home

Abstract Knowledge Graphs are becoming the standard for storing, retrieving and querying structured data. In academia and in industry, they are increasingly used to provide background knowledge. Over the last years, several research contributions are made to show machine learning especially representation learning is successfully applied to knowledge graphs enabling inductive inference about facts with unknown truth values. In this workshop we intend to focus on the most exciting new developments in Knowledge Graph learning, bridging the gap to recent developments in different fields. Also, we want to bring together researchers from different disciplines but united by their adoption of earlier mentioned techniques from machine learning.

Invited speakers:

  • Maximilian Nickel (Facebook AI Research, USA)
  • Mathias Niepert (NEC Labs Europe, Germany)

 

Looking forward to seeing you at The ECML/PKDD 2019.

Paper accepted at ICWE 2019

cropped-ICWE2019-LOGO-1We are very pleased to announce that our group got a paper accepted for presentation at the ICWE 2019: The 19th International Conference on Web Engineering, which will be held on June 11 – 14, 2019 / Daejeon Convention Center (DCC), Daejeon, Korea. The ICWE is the prime yearly international conference on the different aspects of designing, building, maintaining and using Web applications.

The conference will cover the different aspects of Web Engineering, including the design, creation, maintenance and usage of Web applications. ICWE aims to bring together researchers and practitioners from various disciplines in academia and industry to tackle the emerging challenges in the engineering of Web applications and in the problems of its associated technologies, as well as the impact of those technologies on society, media and culture.

Here is the accepted paper with its abstract:

Abstract: ver the last decades the Web has evolved from a human–human communication network to a network of complex human–machine interactions. An increasing amount of data is available as Linked Data which allows machines to “understand” the data, but RDF is not meant to be understood by humans. With Jekyll RDF we present a method to close the gap between structured data and human accessible exploration interfaces by publishing RDF datasets as customizable static HTML sites. It consists of an RDF resource mapping system to serve the resources under their respective IRI, a template mapping based on schema classes, and a markup language to define templates to render customized resource pages. Using the template system, it is possible to create domain specific browsing interfaces for RDF data next to the Linked Data resources. This enables content management and knowledge management systems to serve datasets in a highly customizable, low effort, and scalable way to be consumed by machines as well as humans.

 

Looking forward to seeing you at The ICWE 2019.

Dr. Philipp Senger and Dr. Daniel Burkow from Bayer Crop Science visit SDA

naming_logo_cropscienceDr. Philipp Senger and Dr. Daniel Burkow from Bayer Crop Science visited the SDA group on April  17, 2019.

On Wednesday, SDA had visitors from Bayer: Philipp Senger is the Head of CLS (Computational Life Science) Translational R&D at Bayer Crop Science. His main research interests comprise data science, semantic modeling, knowledge graphs (KGs), natural language processing and machine learning. He together with his team apply their research results to develop digital products and services. Philipp earned a Ph.D. in computer science from the Eberhard-Karls-Universität Tübingen in cooperation with the Robert Bosch GmbH where he investigated data-based methods for automatic generation of electrical behavior models in the automotive industry.

Daniel Burkow is a Post-Doctoral researcher at CLS Translational R&D at Bayer Crop Science.
His main research interests comprise applied mathematics for life sciences, machine learning,  data modeling and knowledge graphs. At Translational R&D at Bayer Crop Science, he works on the development of machine learning applications and data modeling approaches.
Daniel earned his Ph.D. from Arizona State University in the field of Applied Mathematics for Life Sciences in which he examined intramyocellular lipids and the progression of muscular insulin resistance.

Dr. Senger and Dr. Burkow were invited at the bi-weekly “SDA colloquium presentations” where they presented Bayer Crop Science and their main research topics. The goal of the visit was to exchange experience and ideas on combining machine learning approaches with knowledge graphs. In particular, they explained how field trials are modeled at Bayer Crop Science and presented a recent use case in which knowledge graph embeddings have been applied to get more insights about their experimental data.

During the meeting, SDA’s core research topics and main research projects have been presented and future collaborations between Bayer Crop Science R&D and SDA has been discussed. Especially, the application of reasoning and inference methods on large scale KGs describing field experiments, the development of KGE embedding models that incorporate literals (e.g., numerical values) and the application of question answering have been discussed.

We are looking forward to future collaborations with Bayer Crop Science.

Paper accepted at NAACL 2019

naclWe are very pleased to announce that our group got a paper accepted for presentation at The 2019 edition of The NAACL conference, which will be held on June 2–7, 2019 Minneapolis, USA.

NAACL aims to bring together researchers interested in the design and study of natural language processing technology as well as its applications to new problem areas. With this goal in mind, the 2019 edition invites the submission of long and short papers on creative, substantial and unpublished research in all aspects of computational linguistics. It covers a diverse technical program–in addition to traditional research results, papers may present negative findings, survey an area, announce the creation of a new resource, argue a position, report novel linguistic insights derived using existing techniques, and reproduce, or fail to reproduce, previous results.

Here is the pre-print of the accepted paper with its abstract:

Abstract: Short texts challenge NLP tasks such as named entity recognition, disambiguation, linking and relation inference because they do not provide sufficient context or are partially malformed (e.g. wrt. capitalization, long tail entities, implicit relations). In this work, we present the Falcon approach which effectively maps entities and relations within a short text to its mentions of a background knowledge graph. Falcon overcomes the challenges of short text using a light-weight linguistic approach relying on a background knowledge graph. Falcon performs joint entity and relation linking of a short text by leveraging several fundamental principles of English morphology (e.g. compounding, headword identification) and utilizes an extended knowledge graph created by merging entities and relations from various knowledge sources. It uses the context of entities for finding relations and does not require training data. Our empirical study using several standard benchmarks and datasets show that Falcon significantly outperforms state-of-the-art entity and relation linking for short text query inventories.

 

Acknowledgment

This work was partially funded by the Fraunhofer IAIS, and EU H2020 project IASIS.

 

Looking forward to seeing you at The NAACL 2019 conference.