Workshop at COLING 2025 on
Generative AI and Knowledge Graphs (GenAIK)
Abu Dhabi, UAE
January 19/20, 2025
More Details!

About

Generative Artificial Intelligence (GenAI) is a branch of artificial intelligence capable of creating seemingly new, meaningful content, including text, images, and audio. It utilizes deep learning models, such as Large Language Models (LLMs), to recognize and replicate data patterns, enabling the generation of human-like content. Notable families of LLMs include GPT (GPT-3.5, GPT-3.5 Turbo, and GPT-4), LLaMA (LLaMA and LLaMA-2), and Mistral (Mistral and Mixtral). GPT, which stands for Generative Pretrained Transformer, is especially popular for text generation and is widely used in applications like ChatGPT. GenAI has taken the world by storm and revolutionized various industries, including healthcare, finance, and entertainment. However, GenAI models have several limitations, including biases from training data, generating factually incorrect information, and difficulty in understanding complex content. Additionally, their performance can vary based on domain specificity.

In recent times, Knowledge Graphs (KGs) have attracted considerable attention for their ability to represent structured and interconnected information, and adopted by many companies in various domains. KGs represent knowledge by depicting relationships between entities, known as facts, usually based on formal ontological models. Consequently, they enable accuracy, decisiveness, interpretability, domain-specific knowledge, and evolving knowledge in various AI applications. The intersection between GenAI and KG has ignited significant interest and innovation in Natural Language Processing (NLP). For instance, by integrating LLMs with KGs during pre-training and inference, external knowledge can be incorporated for enhancing the model’s capabilities and improving interpretability. When integrated, they offer a robust approach to problemsolving in diverse areas such as information enrichment, representation learning, conversational AI, cross-domain AI transfer, bias, content generation, and semantic understanding. This workshop, aims to reinforce the relationships between Deep Learning, Knowledge Graphs, and NLP communities and foster inter-disciplinary research in the area of GenAI.

Topics of Interest

  • Enhancing KG construction and completion with GenAI
    • Multimodal KG generation
    • Text-to-KG using LLMs
    • Multilingual KGs
  • GenAI for KG embeddings
  • GenAI for Temporal KGs
  • Dialogue systems enhanced by KG and GenAI
  • Cross-domain knowledge transfer with GenAI
  • Bias mitigation using KGs in GenAI
  • Explainability with KGs and GenAI
  • Natural language querying of KGs via GenAI
  • NLP tasks using KGs and GenAI
  • Prompt Engineering using KGs
  • GenAI for Ontology learning and schema induction in KGs
  • Hybrid QA systems combining KGs and GenAI
  • Recommendation systems and KGs with GenAI
  • Creating benchmark datasets relevant for tasks combining KGs and GenAI
  • Real-world applications on scholarly data, biomedical domain, etc.
  • Knowledge Graph Alignment
  • Applying to real-world scenarios


Keynote Speakers

Kang Liu

Could We Locate Knowledge in Large Language Models?

Abstract
Large language models (LLMs) could learn knowledge directly from massive data. Most research currently discusses the relationship between implicit knowledge in LLMs and symbolic knowledge in knowledge graphs. How to understand the knowledge in a LLM and where the knowledge is stored in a LLM have attracted more attention and introduced new research tasks of knowledge localization and knowledge editing. However, could we locate knowledge in a LLM? This talk will introduce our latest research work on this question.
Bio
Kang Liu is a full professor at Institute of Automation, Chinese Academy of Sciences. He is also a youth scientist of Beijing Academy of Artificial Intelligence and a professor of University of Chinese Academy of Sciences. His research interests include Knowledge Graphs, Natural Languase Processing and Large Language Models. He has published over 80 research papers in AI conferences and journals, like ACL, EMNLP, NAACL, COLING, et al. His work has over 20,000 citations on Google Scholar. He received the Best Paper Award at COLING-2014, Best Poster&Demo Award at ISWC-2023, and the Google Focused Research Award in 2015 and 2016.

Wenya Wang

Can LLMs function as, enhance and benefit from Knowledge Bases?

Abstract
The remarkable advancements in large language models (LLMs), including their capacity to store extensive knowledge and perform complex reasoning, have sparked significant interest in their integration with knowledge bases (KBs). This talk delves into our recent research exploring the interplay between LLMs and KBs, addressing whether LLMs can function as, enhance, and benefit from KBs. First, we examine the feasibility of LLMs serving as large-scale knowledge bases by assessing their ability to store knowledge, handle natural language queries, and infer new facts. Next, we introduce an LLM-driven framework designed to augment knowledge bases through reliable reasoning and knowledge base completion. Finally, we present a verification language model that leverages a large-scale corpus, including two knowledge bases, as training data to validate the truth of statements. The talk will conclude with key insights and challenges towards the evolving relationship between LLMs and KBs.
Bio
Wenya Wang is an Assistant Professor with the College of Computing and Data Science, Nanyang Technological University, Singapore. Prior to joining NTU, she worked as a Postdoc with the NLP group at the University of Washington. She obtained the “Lee Kuan Yew Postdoctoral Fellowship” after obtaining her PhD degree at NTU. Her research interests lie in Natural Language Processing, particularly investigating and utilizing generative models in structured knowledge reasoning, model explainability, and trustworthiness.

Workshop Program and Proceedings

The times below are all in UAE’s local time. GenAIK-2025 is an in-person (face to face) only event and will not be broadcasted online.
8:45 - 8:55 Welcome and Opening
8:55 - 10:30 Session 1
  • 8:55 - 9:35 Keynote - Kang Liu - Could We Locate Knowledge in Large Language Models?
  • 9:35 - 9:55 Effective Modeling of Generative Framework for Document-level Relational Triple Extraction - Pratik Saini, Tapas Nayak (paper 6)
  • 9:55 - 10:10 Learn Together: Joint Multitask Finetuning of Pretrained KG-enhanced LLM for Downstream Tasks - Anastasia Martynova, Vladislav Tishin, Natalia Semenova (paper 31)
  • 10:10 - 10:25 GNET-QG: Graph Network for Multi-hop Question Generation - Samin Jamshidi, Yllias Chali (paper 22)

10:30 - 11:00 Coffee Break
11:00 - 12:45 Session 2
  • 11:00 - 11:20 SKETCH: Structured Knowledge Enhanced Text Comprehension for Holistic Retrieval Aakash Mahalingam, Vinesh Kumar Gande, Aman Chadha, Vinija Jain , Divya Chaudhary (paper 18)
  • 11:20 - 11:40 On Reducing Factual Hallucinations in Graph-to-Text Generation using Large Language Models - Dmitrii Iarosh, Alexander Panchenko, Mikhail Salnikov (paper 30)
  • 11:40 - 12:00 GraphRAG: Leveraging Graph-Based Efficiency to Minimize Hallucinations in LLM-Driven RAG for Finance Data - Mariam Barry, Gaëtan Caillaut, Pierre Halftermeyer, Raheel Qader, Mehdi Mouayad, Dimitri Cariolaro, Fabrice Le Deit, Joseph Gesnouin (paper 32)
  • 12:00 - 12:15Structured Knowledge meets GenAI: A Framework for Logic-Driven Language Models Farida Eldessouky, Nourhan Ehab, Carolin Schindler, Mervat Abuelkheir, Wolfgang Minker (paper 14)
  • 12:15 - 12:35 Performance and Limitations of Fine-Tuned LLMs in SPARQL Query Generation Thamer Mecharnia, Mathieu d’Aquin (paper 12)

12:40 - 13:40 Lunch Break
13:40 - 15:30 Session 3
  • 13:40 - 14:20 Keynote - Wenya Wang - Can LLMs function as, enhance and benefit from Knowledge Bases?
  • 14:20 - 14:40 Refining Noisy Knowledge Graph with Large Language Models - Dong Na, Natthawut Kertkeidkachorn, Xin Liu, Kiyoaki Shirai (paper 8)
  • 14:40 - 15:00 Can LLMs be Knowledge Graph Curators for Validating Triple Insertions? - André Gomes Regino, Julio Cesar dos Reis (paper 7)
  • 15:00 - 15:20 Text2Cypher: Bridging Natural Language and Graph Databases - Makbule Gulcin Ozsoy, Leila Messallem, Jon Besga, Gianandrea Minneci (paper 19)

15:30 - 16:00 Coffee Break
16:00 - 17:20 Session 4
  • 16:00 - 16:20 KGFakeNet: A Knowledge Graph-Enhanced Model for Fake News Detection - Anuj Kumar, Pardeep Kumar, Abhishek Yadav, Satyadev Ahlawat, Yamuna Prasad (paper 20)
  • 16:20 - 16:40 Style Knowledge Graph: Augmenting Text Style Transfer with Knowledge Graphs - Martina Toshevska, Slobodan Kalajdziski and Sonja Gievska (paper 29)
  • 16:40 - 17:00 Entity Quality Enhancement in Knowledge Graphs through LLM-based Question Answering Morteza Kamaladdini Ezzabady, Farah Benamara (paper 24)
  • 17:00 - 17:20 Multilingual Skill Extraction for Job Vacancy–Job Seeker Matching in Knowledge Graphs -Hamit Kavas, Marc Serra-Vidal, Leo Wanner (paper 15)

17:20 - 17:40 Closing

Submission Details

Submissions can fall in one of the following categories:
  • Full research papers (6-8 pages)
  • Short research papers (4-6 pages)
  • Position papers (2 pages)
These page limits only apply to the main body of the paper. At the end of the paper (after the conclusions but before the references) papers need to include a mandatory section discussing the limitations of the work and, optionally, a section discussing ethical considerations. Papers can include unlimited pages of references and an unlimited appendix.

Papers must follow the two-column format of CoLING 2025, using the official templates. The templates are available for download as style files and formatting guidelines. Submissions that do not adhere to the specified styles, including paper size, font size restrictions, and margin width, will be desk-rejected.


Authors should follow the general instructions for COLING 2025 proceedings . Submissions are open to all and must be anonymous, adhering to COLING 2025's double-blind submission and reproducibility guidelines. All accepted papers (after double-blind review of at least 3 experts) will appear in the workshop proceedings that will be published by ICCL (International Committee on Computational Linguistics).

At least one of the authors of the accepted papers must register for the workshop to be included into the workshop proceedings. The workshop will be a 100% in-person 1-day event at COLING 2025.

Submissions must be made using the START portal: https://softconf.com/coling2025/GenAIK25/

Important Dates

  • Submission deadline: 5 November 2024 15 November 2024
  • Notification of Acceptance: 5 December 2024
  • Camera-ready paper due: 13 December 2024
  • COLING2025 Workshop day: 19 January 2025
Read CFP

Organizing Committee

Genet Asefa Gesese

FIZ Karlsruhe, KIT, Germany

Harald Sack

FIZ Karlsruhe, KIT, Germany

Heiko Paulheim

University of Mannheim, Germany

Albert Meroño-Peñuela

King’s College London, UK

Lihu Chen

Imperial College London, UK

Program Committee

  • Paul Groth, University of amsterdam, The Netherlands
  • Achim Rettinger, University of Trier, Germany
  • Cassia Trojahn, Institut de Recherche en Informatique de Toulouse, France
  • Gerard de Melo, Hasso Plattner Institute, Germany
  • Pushkar Singh, Google Search Ads AI, United States
  • Davide Buscaldi, Université Paris 13, France
  • Pierre-Henri Paris, Paris-Saclay University, France
  • Mayank Kejriwal, University of Southern California, United States
  • Finn Arup Nielsen, Technical University of Denmark, Denmark
  • Marco Bombieri, University of Verona, Italy
  • Fabien Gandon, Institut national de recherche en informatique et en automatique, France
  • Giuseppe Rizzo, LINKS Foundation, Italy
  • Edlira Vakaj Kalemi, Birmingham City University, UK
  • Vojtech Svatek, Prague University of Economics and Business, Czech Republic
  • Ondřej Zamazal, Prague University of Economics and Business, Czech Republic
  • Daniel Garijo, Universidad Politécnica de Madrid, Spain
  • Sanju Tiwari, TIB Hannover, Germany
  • Shufan JIANG, FIZ-Karlsruhe, Germany
  • Jesualdo Fernández-Breis, University of Murcia, Spain
  • Mauro Dragoni, Fondazione Bruno Kessler, Italy
  • Kristiina Jokinen, University of Helsinki, Finland
  • Graham Wilcock, University of Helsinki, Finland
  • Mathieu d'Aquin, Université de Lorraine, France
  • Giancarlo Guizzardi, University of Twente, EEMCS, The Netherlands
  • Oscar Rodríguez Rocha, Institut national de recherche en informatique et en automatique, France
  • Suvodeep Mazumdar, University of Sheffield, UK

**If you have published in *ACL conferences previously, and are interested to be part of the program committee of GenAIK2024, please fill in this form.