=Paper=
{{Paper
|id=Vol-3337/smart-preface
|storemode=property
|title=None
|pdfUrl=https://ceur-ws.org/Vol-3337/smart-preface.pdf
|volume=Vol-3337
}}
==None==
Preface - SMART 2022 SMART 2022 was the third edition of SMART, the subtasks for Question An- swering over Knowledge Graph, which is part of the ISWC 2022 Semantic Web Challenge. It was co-located with the 21st International Semantic Web Con- ference (ISWC 2022)1 . The first edition SMART2020 [1] was in ISWC 2020 and the second edition SMART2020 [2]. Given a question in natural lan- guage, the task of the SMART challenge is, to predict the answer type, en- tities and relations using a target ontology. In the current (third) edition, the SMART challenge had three tracks i.e. answer type prediction, entity link- ing and relation prediction. These three tasks are based on two popular KBs, one using the DBpedia ontology and the other using Wikidata. This volume contains peer-reviewed system description papers of all the systems that par- ticipated in the challenge. More details about the challenge can be found at https://smart-task.github.io/2022/. Program and Keynote This year’s edition had two talks and one keynote. First Philipp Christmann spoke on behalf of his co-authors Rishiraj Saha Roy, Gerhard Weikum about ’Entity and Relation Linking using CLOCQ’ Second the paper ’Contribution to SMART task 2022: Answer Type Prediction, Relation Linking and Entity Linking’ by Azanzi Jiomekong, Vadel Tsague, Brice Foko, Uriel Melie, Gaoussou Camara was presented. SMART also invited Dennis Diefenbach to give a keynote. He is the CEO of The QA Company. Dennis holds a Ph.D. from the University of Lyon. He talked about the ’Challenges of Domain Specific Question Answering Systems’. We would like to thank all presenters again. Challenge Description This challenge is focused on answer type prediction, entity linking and relation prediction, which play an important role in KGQA (Question Answering over Knowledge Graphs) systems. Answer Type Prediction Given a natural language question, the task is to produce a ranked list of answer types of a given target ontology. Previ- ous such answer type classifications in literature are performed as a short-text classification task using a set of coarse-grained types, for instance, either six types [3, 4, 5, 6] or 50 types [7] with TREC QA task2 . We propose a more granular answer type classification using popular Semantic Web ontologies such as DBpedia and Wikidata. 1 https://iswc2022.semanticweb.org/ 2 https://trec.nist.gov/data/qamain.html 1 Relation Prediction Given a natural language question, the task is to iden- tify the relation and link to the relations in KG. Depending on the number of relations in the KG, the number of relation types to be linked varies. Entity Linking Given a natural language question, the task is to identify the entity mentions and link to the entities in KG. Table 1, Table 2 and Table 3 illustrate some examples. The participating systems can be either supervised (training data is provided) or unsupervised. The systems can utilize a wide range of approaches; from rule-based to neural approaches. Table 1: Example questions and answer types. Answer Type Question DBpedia Wikidata Give me all actors starring in dbo:Actor wd:Q33999 movies directed by and star- ring William Shatner. Which programming lan- dbo:ProgrammingLanguage wd:Q9143 guages were influenced by Perl? Who is the heaviest player of dbo:BasketballPlayer wd:Q3665646 the Chicago Bulls? How many employees does xsd:integer xsd:integer Google have? Table 2: Example questions and relation types. Relation Type Question DBpedia Wikidata Which languages were in- dbo:influencedBy wdt:P737 fluenced by Perl? Give me all actors star- dbo:starring, dbo:director wdt:P161, wdt:P57 ring in movies directed by and starring William Shatner. How many employees dbo:numberOfEmployees wdt:P1128 does IBM have? Organisation In this section, we list the people who organised and contributed to the success of this event. 2 Table 3: Example questions and entity linking. Entity Linking Question DBpedia Wikidata Which languages were influenced by dbr:Perl wd:Q42478 Perl? Give me all actors starring in movies dbr:William_Shatner wd:Q16297 directed by and starring William Shatner. How many employees does IBM dbr:IBM wd:Q37156 have? Challenge Chairs • Nandana Mihindukulasooriya (IBM Research AI) • Mohnish Dubey (Philips Research, Netherlands) • Alfio Gliozzo (IBM Research AI) • Jens Lehmann (Amazon) • Axel-Cyrille Ngonga Ngomo (Universität Paderborn) • Ricardo Usbeck (University of Hamburg) • Gaetano Rossiello (IBM Research AI) • Uttam Kumar (University of Bonn) • Debayan Banerjee (Universität Hamburg) Challenge Programme Committee Members The challenge program committee helped to peer-review the eight system pa- pers. Each paper received 2 or 3 reviews from the program committee members and authors took those feedback into account when preparing the camera-ready versions. The organizers would like to thank them for their valuable time. • Ahmad Alobaid (Universidad Politécnica de Madrid) • Debayan Banerjee (Universität Hamburg) • Mohnish Dubey (Philips Research, Netherlands) • Longquan Jiang (University of Hamburg) • Marcos Martinez-Galindo (IBM Research) • Nandana Mihindukulasooriya (IBM Research AI) • Cedric Möller (University of Hamburg) • Ricardo Usbeck (University of Hamburg) 3 Acknowledgements We would like to thank the ISWC Semantic Web Challenge chairs, Daniele DellAglio, Catia Pesquita and the whole ISWC organising committee for their invaluable support to make this event a success. We would also like to thank the challenge participants for their interest, quality of work, and informative presentations during the event which made it attractive to the ISWC audience. References [1] Nandana Mihindukulasooriya, Mohnish Dubey, Alfio Gliozzo, Jens Lehmann, Axel-Cyrille Ngonga Ngomo, and Ricardo Usbeck. SeMantic An- sweR Type prediction task (SMART) at ISWC 2020 Semantic Web Chal- lenge. CoRR/arXiv, abs/2012.00555, 2020. [2] Nandana Mihindukulasooriya, Mohnish Dubey, Alfio Gliozzo, Jens Lehmann, Axel-Cyrille Ngonga Ngomo, Ricardo Usbeck, Gaetano Rossiello, and Uttam Kumar. Semantic Answer Type and Relation Prediction Task (SMART 2021). CoRR/arXiv, abs/2112.07606, 2022. [3] Han Zhao, Zhengdong Lu, and Pascal Poupart. Self-adaptive hierarchical sentence model. In Twenty-Fourth International Joint Conference on Arti- ficial Intelligence, 2015. [4] Chunting Zhou, Chonglin Sun, Zhiyuan Liu, and Francis Lau. A C-LSTM neural network for text classification. arXiv preprint arXiv:1511.08630, 2015. [5] Yoon Kim. Convolutional neural networks for sentence classification. In Pro- ceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP 2014), 2014. [6] Nal Kalchbrenner, Edward Grefenstette, and Phil Blunsom. A convolutional neural network for modelling sentences. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 655–665, Baltimore, Maryland, June 2014. Association for Computational Linguistics. [7] Xin Li and Dan Roth. Learning Question Classifiers: the Role of Semantic Information. Natural Language Engineering, 12(3):229–249, 2006. 4