Abstract Summary
In the online digital ecosystem, we are surrounded by distinct forms of information pollution, posing significant threats to both individuals and society as a whole. Just think of the recent riots in England and Wales which were triggered by deliberate misinformation. Other instances of false news, for instance, wield power to sway public opinion on matters of politics and finance. Deceptive reviews can either bolster or tarnish the reputation of businesses, while unverified medical advice may steer people toward harmful health practices. In light of this challenging landscape, it has become imperative to ensure that users have access to both topically relevant and factually accurate information that does not warp their perception of reality, and there has been a surge of interest in various strategies to combat misinformation through different contexts and multiple tasks. The purpose of the ROMCIR Workshop, for some years now, is precisely that of engaging the Information Retrieval community to explore potential solutions that extend beyond conventional misinformation detection approaches. Key objectives include identifying subjective and objective factors associated with information credibility and truthfulness respectively, and integrating such factors as fundamental dimensions of relevance within Information Retrieval Systems (IRSs), achieving early detection of misinformation, and ensuring that the search results retrieved are not only truthful but also explainable to the users of IRSs. Moreover, it is essential to evaluate the role of generative models such as Language Models (LLMs) in inadvertently amplifying misinformation problems, and how they can be used to support IRSs, together with the contribution that the human-in-the-loop paradigm can have in this context.