{"id":37260,"date":"2025-08-04T15:46:34","date_gmt":"2025-08-04T13:46:34","guid":{"rendered":"https:\/\/kinit.sk\/fact-check-retrieval-using-causal-llms\/"},"modified":"2025-08-04T15:51:21","modified_gmt":"2025-08-04T13:51:21","slug":"fact-check-retrieval-using-causal-llms","status":"publish","type":"post","link":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/","title":{"rendered":"Fact-Check Retrieval Using Causal LLMs"},"content":{"rendered":"<div id=\"\" class=\"element core-paragraph\">\n<p><strong>In the <a href=\"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-text-embedding-models\/\">first instalment<\/a> of our 3-part series, we saw how a text embedding model can be used to retrieve existing fact-checks across different languages and help fact-checkers to cut down on duplicate effort. In this second part, we are going to look at how one can plug causal, generative LLMs (large language models) into the pipeline to refine the results further.<\/strong><\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>Generative LLMs are now widely understood to have the ability to capture richer semantic representations and to better handle contextual nuances in comparison to text embedding models. They owe this to several factors, such as the scale and generality of their pretraining data, their instruction tuning, their (limited) ability to perform reasoning, and even just the sheer number of parameters.<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-heading\">\n<h3 class=\"wp-block-heading\">LLMs Will Sift through Your Retrieved Fact-Checks<\/h3>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>So how can we incorporate generative LLMs into the retrieval process? Let us recall what we did last time using TEMs (text embedding models). As shown in Fig. 1, the idea was to simply feed both our documents (fact-checks) and our query (the claim we are interested in) into the TEM to get their embedding vectors. We would then compare the embedding vector of our query against those of all the documents and rank the documents by the resulting similarity scores. The most relevant documents should then be at the top of the list. Note that the individual documents and the query are all embedded independently of each other, so the embeddings of all the documents can be precomputed ahead of time and then reused across all searches.<\/p>\n<\/div>\n\n<div class=\"wp-block-image\"><div id=\"\" class=\"element core-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" data-src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXfKNZSFEFtYWEG7TOhEuF7WRZTgsS6MUYHxgVPWBf4QZLSBrObnkAlnV2efXLieTumWvhM-0bFOwiVOVN5lJP-KQVrQmwWfxXvZ6CzyLRy_m7-cdUrqah3aZhzGGXg_1ShfAmoS7g?key=h7FRyqI0LG_7YM0vsTj5dwP8\" alt=\"\" style=\"width:371px;height:auto\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" class=\"lazyload\" \/><\/figure>\n<\/div><\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p class=\"has-text-align-center\"><em>Fig. 1: Document retrieval using text embedding models (TEMs).<\/em><\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>When incorporating generative LLMs into the process, we are not going to touch the TEM part, as it is still a very efficient way to get the first rough ranking. You can think of this as performing a Google search \u2013 you get a list of results which are more or less relevant already, but then you need to sift through them and figure out which are really useful to you and which, although superficially similar, are not what you were looking for.<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>In our new setup, it is this sifting task that is going to be performed by the generative LLM. The LLM is going to sort through a list of top-ranked documents retrieved by the TEM. It will push the most relevant items towards the top of the list, and it will filter some items out altogether if it finds that they are not actually relevant. The pipeline as a whole is illustrated in Fig. 2.<\/p>\n<\/div>\n\n<div class=\"wp-block-image\"><div id=\"\" class=\"element core-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" data-src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXfeT1kmTiwiPh1JaMLK-wc6zyoEHSEhH4oeunG8fOsa-YovycFOO4dO1KENjF6S6bXhm1pihbtqSAgxaF2nunbdf_WqAcugxyu_G4rYLWrS_ZvP9Kk8T_QpzgBQlvJIpAlGqWHtqA?key=h7FRyqI0LG_7YM0vsTj5dwP8\" alt=\"\" style=\"width:568px;height:auto\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" class=\"lazyload\" \/><\/figure>\n<\/div><\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p class=\"has-text-align-center\"><em>Fig. 2: A retrieval pipeline combining a text embedding model with an LLM.<\/em><\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-heading\">\n<h3 class=\"wp-block-heading\">Why LLMs Have an Easier Time Figuring Out Details<\/h3>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>Note that the TEM and the LLM tackle the task in two very different ways. The TEM is asked to embed every item independently into a fixed-size vector. Consequently, what it really needs to do is take each document and squash all the information that could possibly be relevant into a very compact form \u2013 and all that without having any idea of what the query is going to be about! It is no wonder that the TEM-based ranking can sometimes miss particular details or ignore various nuances.<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>The LLM, in contrast, would naturally have an easier time figuring out details. After all, we are not asking it to blindly squash all the content into one small vector and then make do with the information that survived. On the contrary, the LLM attends to the entire query and to all the documents as it works. It can literally weigh the meaning of each particular word and, having seen the query, consider all this information in the context of what the user is looking for just now.<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-heading\">\n<h3 class=\"wp-block-heading\">Prompting Is Crucial!<\/h3>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>If you work with large language models often, you may not be surprised that it matters <em>how <\/em>you ask the LLM to perform the task \u2013 LLMs can be quite sensitive to what prompt you provide, and a prompt can often even make or break the entire pipeline. We have experimented with several different setups in our accompanying paper (which is now out on <a href=\"https:\/\/arxiv.org\/abs\/2503.02737\">arXiv<\/a>):<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-list\">\n<ul class=\"wp-block-list\"><div id=\"\" class=\"element core-list-item\">\n<li><strong>Zero-shot:<\/strong> We just ask the LLM: &#8222;Is the claim relevant to the social media post?&#8220; \u2013 without providing any additional context.<\/li>\n<\/div>\n\n<div id=\"\" class=\"element core-list-item\">\n<li><strong>Zero-shot with task description:<\/strong> We provide the LLM with more detailed instructions on how to approach the task, derived from guidelines provided to human annotators.<\/li>\n<\/div>\n\n<div id=\"\" class=\"element core-list-item\">\n<li><strong>Few-shot with task description:<\/strong> We provide the LLM with the task description and also include 10 examples (5 of relevant and 5 of irrelevant pairs). These 10 examples are actually not fixed ahead of time, but retrieved from a larger set by a TEM, based on similarity to the user&#8217;s query, to make them as helpful to the LLM as possible.<\/li>\n<\/div>\n\n<div id=\"\" class=\"element core-list-item\">\n<li><strong>Chain-of-thought:<\/strong> We additionally instruct the LLM to \u201cthink step by step\u201d \u2013 this encourages it to spend some time reasoning about the post and fact-checks before it provides an answer.<\/li>\n<\/div>\n\n<div id=\"\" class=\"element core-list-item\">\n<li><strong>Cross-lingual thought<\/strong> prompting (XLT): We additionally instruct the model to translate the content into English before it proceeds with the rest of the task \u2013 it has been observed that some models\u2019 skills just work better in English than in other languages, so this can sometimes help.<\/li>\n<\/div><\/ul>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>If you ask which setup is the best for fact-check retrieval, the answer is that it depends. The optimal strategy differs a lot from model to model. Out of the open-weight models that we experimented with, Mistral-Large showed the most robust performance, improving over TEM baselines in all setups, but achieving the best scores when using few-shot prompting with the task description \u2013 the results are summarised in Fig. 3. Note that in the plain zero-shot setup (without the more comprehensive task description), the improvement is very marginal \u2013 the results are more or less the same as in the TEM-only setup.<\/p>\n<\/div>\n\n<div class=\"wp-block-image\"><div id=\"\" class=\"element core-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" data-src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXdqJtKLEX15Cuy2KuVTW2cVPJcQMWHO1vOK8xxYXEFPJVYpi6xNafiEv4yuloax7G9wCPG4sZSKi9r6SsVq1foAOFAbJuWvW4olKdvGtoP42woCKKsLW7kMCv63dOGZsFqG01LFag?key=h7FRyqI0LG_7YM0vsTj5dwP8\" alt=\"\" style=\"width:589px;height:auto\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" class=\"lazyload\" \/><\/figure>\n<\/div><\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p class=\"has-text-align-center\"><em>Fig. 3: Performance comparison of LLMs across five prompting strategies in the original language, measured by Macro F1 score with confidence intervals. The dashed horizontal lines represent the best-performing baselines.<\/em><\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-heading\">\n<h3 class=\"wp-block-heading\">Multilingual Performance of Generative LLMs<\/h3>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>One particularly challenging aspect of searching for existing fact-checks is that they may easily have been written in a different language than your post, so to find them efficiently, you really need a tool that can conduct cross-lingual searches.<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>Thankfully, there are now some models \u2013 both TEMs and generative LLMs \u2013 with reasonably good multilingual capabilities. As we already know, Mistral-Large takes the lead overall \u2013 this can also be observed in Fig. 4, which shows the results achieved by different generative LLMs and by TEM baselines (shown in bold) across 20 individual languages and 20 cross-lingual pairs (i.e. the searched-for claim is in a different language than the target fact-checks).<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>Interestingly, though, it seems that even small models can be competitive \u2013 Qwen 2.5 7B\u2019s results arguably do not lag very far behind Mistral Large in some cases and even manage to outpace it in a few. Furthermore, although both versions of Qwen actually underperform the highest-achieving TEM (DeBERTa v3 Large) in many monolingual cases, they do seem to be more robust in terms of cross-lingual performance.<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>For additional information, have a look at our paper on <a href=\"https:\/\/arxiv.org\/abs\/2503.02737\">arXiv<\/a>.<\/p>\n<\/div>\n\n<div class=\"wp-block-image\"><div id=\"\" class=\"element core-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" data-src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXciPJHyv-QvYw8eth_SS9IBNIh13Cz1qtS2C57nDR49kimTjDFzMG3aSY8X6Dq6NpUG2hLuD_hT1-y5x3U__qf2K-mUV2WGJkSWcPPiZ3V0UZaBIpCSdE2sHj4BtXjEx2C3fTo7-g?key=h7FRyqI0LG_7YM0vsTj5dwP8\" alt=\"\" style=\"width:718px;height:auto\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" class=\"lazyload\" \/><\/figure>\n<\/div><\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p class=\"has-text-align-center\"><em>Fig. 4: Performance of TEMs (bold) and generative LLMs (normal) across 20 individual languages (left) and 20 cross-lingual pairs (right).<\/em><\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-heading\">\n<h3 class=\"wp-block-heading\">The Way Ahead<\/h3>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>It is clear that as models keep getting better, they are able to tackle increasingly complex tasks \u2013 and across a growing number of languages. At present, one can often push performance by carefully designing prompts and combining several kinds of models in a pipeline. Hopefully, going forward, the amount of engineering required to get the best results will decrease, and as models gain more skills, improved multilingual support and more advanced reasoning capabilities, they will also get even easier to apply \u2013 we will see.<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-paragraph\">\n<p>As a quick recap, our blog series has 3 parts. The first covered TEMs. In this, the second part, we had a look at generative LLMs and the third and final part \u2013 on multimodal fact-check retrieval \u2013 is yet to come:<\/p>\n<\/div>\n\n<div id=\"\" class=\"element core-list\">\n<ol class=\"wp-block-list\"><div id=\"\" class=\"element core-list-item\">\n<li><a href=\"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-text-embedding-models\/\"><strong>Fact-Check Retrieval Using Text Embedding Models<\/strong><\/a>, where we explain the fact-check retrieval task and how it can be addressed using text embedding models (TEMs), while supporting search across different languages.<\/li>\n<\/div>\n\n<div id=\"\" class=\"element core-list-item\">\n<li><strong>[This is where you are!] Fact-Check Retrieval Using LLMs<\/strong>, where we explain how generative large language models (LLMs) enter into the equation and how they can help to refine retrieval to improve results further.<\/li>\n<\/div>\n\n<div id=\"\" class=\"element core-list-item\">\n<li><strong>Multimodal Fact-Check Retrieval<\/strong>, where we explain how visual content can be leveraged to improve retrieval performance, starts with a very straightforward setup based on a combination of several smaller models and then goes on to how generative LLMs can be applied to the same task.<\/li>\n<\/div><\/ol>\n<\/div>","protected":false},"excerpt":{"rendered":"<p>In the first instalment of our 3-part series, we saw how a text embedding model can be used to retrieve existing fact-checks across different languages and help fact-checkers to cut&#8230;<\/p>\n","protected":false},"author":34,"featured_media":37258,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":true,"footnotes":""},"categories":[83,520],"tags":[395,530],"class_list":["post-37260","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news-sk","category-2025-sk","tag-fact-checking-sk","tag-disai-amplified"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Fact-Check Retrieval Using Causal LLMs - KInIT<\/title>\n<meta name=\"description\" content=\"In the first instalment of our 3-part series, we saw how a text embedding model can be used to retrieve existing fact-checks across...\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/\" \/>\n<meta property=\"og:locale\" content=\"sk_SK\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Fact-Check Retrieval Using Causal LLMs - KInIT\" \/>\n<meta property=\"og:description\" content=\"In the first instalment of our 3-part series, we saw how a text embedding model can be used to retrieve existing fact-checks across...\" \/>\n<meta property=\"og:url\" content=\"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/\" \/>\n<meta property=\"og:site_name\" content=\"KInIT\" \/>\n<meta property=\"article:published_time\" content=\"2025-08-04T13:46:34+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2025-08-04T13:51:21+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/kinit.sk\/wp-content\/uploads\/2025\/08\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1200\" \/>\n\t<meta property=\"og:image:height\" content=\"628\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"Wanda Pribylincova\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@kinit\" \/>\n<meta name=\"twitter:site\" content=\"@kinit\" \/>\n<meta name=\"twitter:label1\" content=\"Autor\" \/>\n\t<meta name=\"twitter:data1\" content=\"Wanda Pribylincova\" \/>\n\t<meta name=\"twitter:label2\" content=\"Predpokladan\u00fd \u010das \u010d\u00edtania\" \/>\n\t<meta name=\"twitter:data2\" content=\"8 min\u00fat\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/\"},\"author\":{\"name\":\"Wanda Pribylincova\",\"@id\":\"https:\\\/\\\/kinit.sk\\\/#\\\/schema\\\/person\\\/64db52a830dcb6d4df386e78e7eb748b\"},\"headline\":\"Fact-Check Retrieval Using Causal LLMs\",\"datePublished\":\"2025-08-04T13:46:34+00:00\",\"dateModified\":\"2025-08-04T13:51:21+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/\"},\"wordCount\":1435,\"image\":{\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/kinit.sk\\\/wp-content\\\/uploads\\\/2025\\\/08\\\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png\",\"keywords\":[\"fact-checking\",\"disai-amplified\"],\"articleSection\":[\"News\",\"2025\"],\"inLanguage\":\"sk-SK\"},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/\",\"url\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/\",\"name\":\"Fact-Check Retrieval Using Causal LLMs - KInIT\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/kinit.sk\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/kinit.sk\\\/wp-content\\\/uploads\\\/2025\\\/08\\\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png\",\"datePublished\":\"2025-08-04T13:46:34+00:00\",\"dateModified\":\"2025-08-04T13:51:21+00:00\",\"author\":{\"@id\":\"https:\\\/\\\/kinit.sk\\\/#\\\/schema\\\/person\\\/64db52a830dcb6d4df386e78e7eb748b\"},\"description\":\"In the first instalment of our 3-part series, we saw how a text embedding model can be used to retrieve existing fact-checks across...\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/#breadcrumb\"},\"inLanguage\":\"sk-SK\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"sk-SK\",\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/#primaryimage\",\"url\":\"https:\\\/\\\/kinit.sk\\\/wp-content\\\/uploads\\\/2025\\\/08\\\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png\",\"contentUrl\":\"https:\\\/\\\/kinit.sk\\\/wp-content\\\/uploads\\\/2025\\\/08\\\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png\",\"width\":1200,\"height\":628},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/fact-check-retrieval-using-causal-llms\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/kinit.sk\\\/sk\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"News\",\"item\":\"https:\\\/\\\/kinit.sk\\\/category\\\/news\\\/\"},{\"@type\":\"ListItem\",\"position\":3,\"name\":\"Fact-Check Retrieval Using Causal LLMs\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/kinit.sk\\\/#website\",\"url\":\"https:\\\/\\\/kinit.sk\\\/\",\"name\":\"KInIT\",\"description\":\"Vyu\u017e\u00edvame v\u00fdskum pre \u013eud\u00ed a priemysel\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/kinit.sk\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"sk-SK\"},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/kinit.sk\\\/#\\\/schema\\\/person\\\/64db52a830dcb6d4df386e78e7eb748b\",\"name\":\"Wanda Pribylincova\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Fact-Check Retrieval Using Causal LLMs - KInIT","description":"In the first instalment of our 3-part series, we saw how a text embedding model can be used to retrieve existing fact-checks across...","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/","og_locale":"sk_SK","og_type":"article","og_title":"Fact-Check Retrieval Using Causal LLMs - KInIT","og_description":"In the first instalment of our 3-part series, we saw how a text embedding model can be used to retrieve existing fact-checks across...","og_url":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/","og_site_name":"KInIT","article_published_time":"2025-08-04T13:46:34+00:00","article_modified_time":"2025-08-04T13:51:21+00:00","og_image":[{"width":1200,"height":628,"url":"https:\/\/kinit.sk\/wp-content\/uploads\/2025\/08\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png","type":"image\/png"}],"author":"Wanda Pribylincova","twitter_card":"summary_large_image","twitter_creator":"@kinit","twitter_site":"@kinit","twitter_misc":{"Autor":"Wanda Pribylincova","Predpokladan\u00fd \u010das \u010d\u00edtania":"8 min\u00fat"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/#article","isPartOf":{"@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/"},"author":{"name":"Wanda Pribylincova","@id":"https:\/\/kinit.sk\/#\/schema\/person\/64db52a830dcb6d4df386e78e7eb748b"},"headline":"Fact-Check Retrieval Using Causal LLMs","datePublished":"2025-08-04T13:46:34+00:00","dateModified":"2025-08-04T13:51:21+00:00","mainEntityOfPage":{"@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/"},"wordCount":1435,"image":{"@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/#primaryimage"},"thumbnailUrl":"https:\/\/kinit.sk\/wp-content\/uploads\/2025\/08\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png","keywords":["fact-checking","disai-amplified"],"articleSection":["News","2025"],"inLanguage":"sk-SK"},{"@type":"WebPage","@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/","url":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/","name":"Fact-Check Retrieval Using Causal LLMs - KInIT","isPartOf":{"@id":"https:\/\/kinit.sk\/#website"},"primaryImageOfPage":{"@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/#primaryimage"},"image":{"@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/#primaryimage"},"thumbnailUrl":"https:\/\/kinit.sk\/wp-content\/uploads\/2025\/08\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png","datePublished":"2025-08-04T13:46:34+00:00","dateModified":"2025-08-04T13:51:21+00:00","author":{"@id":"https:\/\/kinit.sk\/#\/schema\/person\/64db52a830dcb6d4df386e78e7eb748b"},"description":"In the first instalment of our 3-part series, we saw how a text embedding model can be used to retrieve existing fact-checks across...","breadcrumb":{"@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/#breadcrumb"},"inLanguage":"sk-SK","potentialAction":[{"@type":"ReadAction","target":["https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/"]}]},{"@type":"ImageObject","inLanguage":"sk-SK","@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/#primaryimage","url":"https:\/\/kinit.sk\/wp-content\/uploads\/2025\/08\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png","contentUrl":"https:\/\/kinit.sk\/wp-content\/uploads\/2025\/08\/DisAI-AMPLIFIED_feature-image_Fact-Check-Retrieval-Using-Causal-LLMs_.png","width":1200,"height":628},{"@type":"BreadcrumbList","@id":"https:\/\/kinit.sk\/sk\/fact-check-retrieval-using-causal-llms\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/kinit.sk\/sk\/"},{"@type":"ListItem","position":2,"name":"News","item":"https:\/\/kinit.sk\/category\/news\/"},{"@type":"ListItem","position":3,"name":"Fact-Check Retrieval Using Causal LLMs"}]},{"@type":"WebSite","@id":"https:\/\/kinit.sk\/#website","url":"https:\/\/kinit.sk\/","name":"KInIT","description":"Vyu\u017e\u00edvame v\u00fdskum pre \u013eud\u00ed a priemysel","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/kinit.sk\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"sk-SK"},{"@type":"Person","@id":"https:\/\/kinit.sk\/#\/schema\/person\/64db52a830dcb6d4df386e78e7eb748b","name":"Wanda Pribylincova"}]}},"_links":{"self":[{"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/posts\/37260","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/users\/34"}],"replies":[{"embeddable":true,"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/comments?post=37260"}],"version-history":[{"count":4,"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/posts\/37260\/revisions"}],"predecessor-version":[{"id":37267,"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/posts\/37260\/revisions\/37267"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/media\/37258"}],"wp:attachment":[{"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/media?parent=37260"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/categories?post=37260"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/kinit.sk\/sk\/wp-json\/wp\/v2\/tags?post=37260"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}