Join our PhD program and become an expert in artificial intelligence. Dive into challenging research topics with world-class mentors and industry partners. We welcome curious minds to join our open culture.

Why PhD at KInIT

Rise to excellence

Take advantage of an inspiring environment and become a top expert in your selected AI topic. At Kempelen institute, we prioritise ethical responsibility and societal benefit.

Learn from the best

When working on your dissertation, you’ll have access to a distinguished supervising team, including industrial partners and/or renowned experts from world-class research institutions.

Research that impacts

Choose a research topic that deeply interests you and will enrich the field. At KInIT, we support excellent, original world-class research and the connection with industry and current research and innovation challenges.

Open culture

Our culture is based on trust, openness and respect. That is how great ideas are born. We share knowledge and innovations not only within the institute, but also with our industry partners and other curious people like you.

Build your network

You’ll be teamed up with the best researchers in the region and gain invaluable experience and contacts from abroad that will boost your career growth in industry or academia.

No distractions

You will become a full-time KInIT employee. We will give you all the time and space you need to focus on your research, skills development and growth.

KInIT PhD program is run in partnership with

What our PhD students said

Brano Pecher

„Even though I would have been nervous about the defense and presentation, I think I have grown in the last four years, and presented so much that I’m more calm right now.“

Branislav Pecher

#growth

goldschmidtweb

„The impact of my research is to catch bad guys on the internet by using artificial intelligence and machine learning techniques.“

Patrik Goldschmidt

#impact

Belanec_Robert

„I was able to meet brilliant and like-minded minds from various fields of artificial intelligence and machine learning.“

Róbert Belanec

#internship

benova web

„The opportunity to have an external supervisor was actually one of the reasons I decided to join KInIT. It wasn’t something offered anywhere else when I was looking for a PhD position.“

Ivana Beňová

#mentorship

mocko-web_51330189024_o

„Working with ESET has helped me to get a better understanding of the differences between academic research and industry needs.“

Martin Mocko

#innovative projects

Cegin_Jan

„It was a really good experience to show my work to the world, to the best people in the field, and to get recognition for it.“

Ján Čegiň

#conferences

Vykopal_Ivan_2022

„Writing a paper is a long and interactive process, it often challenged me to communicate my research clearly.“

Ivan Vykopal

#published paper

pavlik web

„It can be challenging because you have to balance the goals of the industry partner with actually publishing papers, so we’re doing something that is useful for them, and also that is novel enough to be publishable.“

Peter Pavlík

#innovative projects

santi square

„KInIT provides a network of researchers that I was excited to be a part of and that would give me a lot of opportunities.“

Santiago Jose de Leon Martinez

#team

Dissertation topics 2026/2027

Recommender and Adaptive Web-Based Systems

Supervisor: Michal Kompan (supervisor, KInIT)

Supervising team: Michal Kompan (supervisor, KInIT), Peter Brusilovsky (University of Pittsburgh), Branislav Kveton (Google Research), Peter Dolog (Aalborg University),

Keywords: personalised recommendation, machine learning, user model, fairness, multi
objective and multi stakeholder recommendations, session-based

Viac info Menej info

The recommender systems are an integral part of almost every modern Web application. Personalized, or at least adaptive, services have become a standard that is expected by the users in almost every domain (e.g., news, chatbots, social media, or search).

Obviously, personalization has a great impact on the everyday life of hundreds of million users across many domains and applications. This results in a major challenge – to propose methods that are not only accurate but also trustworthy and fair. Such a goal offers plenty of research opportunities in many directions:

  • Trustworthy recommendation methods for sequence and session-based applications
  • Multi-objective and multi-stakeholder recommender systems
  • Explaining recommendations
  • Fairness and justice in recommendations
  • Biases in the recommendations

There are several application domains where these research problems can be addressed, e.g., search, e-commerce, social networks, news, and many others.

Relevant publications:

The research will be performed at the Kempelen Institute of Intelligent Technologies (KInIT, https://kinit.sk) in Bratislava in cooperation with industrial partners or researchers from highly respected research units from abroad. A combined (external) form of study and full employment at KInIT is expected.

Supervising team
Michal Kompan
Lead researcher, KInIT
Peter Brusilovsky
Professor, University of Pittsburgh, USA
Branislav Kveton
Principal Scientist, Adobe Research, USA
Peter Dolog
Associate Professor, Aalborg University, Denmark

Evaluating Vulnerabilities and Capabilities of Large Language Models

Supervisor: Mária Bieliková (supervisor, KInIT)

Supervising team: Mária Bieliková (supervisor, KInIT), Branislav Pecher (KInIT)

Keywords: large language models, in-context learning, benchmarking, mechanistic interpretability, metrics and evaluation, robustness, reasoning, dynamic benchmarks

Viac info Menej info

Despite the impressive performance of large language models (LLMs), these models are still not well understood. A lot of effort is dedicated to evaluating the capabilities of LLMs and creating guardrails to prevent potentially harmful behaviours. However, small prompt variations still significantly alter the behaviour of LLMs, they often fail in tasks that are similar to the ones seen in training data but contain a small change, or they can be easily used to generate confident misinformation. As the large language models increasingly mediate access to information and are increasingly used by expert and non-expert users, it is essential to develop rigorous methods for evaluating their true capabilities and identifying the sources of their vulnerabilities.

The goal of this research is to build a deeper and more reliable understanding of the LLM behaviour. The first possible research direction involves designing a benchmark that can accurately measure both advanced capabilities (such as reasoning, planning, understanding, math or multilingual capabilities) and common failure modes and vulnerabilities (including hallucination, brittleness, tendency or ease of generating problematic content such as misinformation). An important focus should be on making the benchmark robust to the current issues of data contamination and minimising the resources needed to run it. The second possible research direction leverages mechanistic interpretability to study the internal structures that drive the LLM behaviours. These tools will be used to investigate why specific vulnerabilities arise, how the capabilities arise, and whether targeted interventions can reduce failures while maintaining core capabilities.

By integrating robust behavioural benchmarking with mechanistic insights, the project aims to produce a more principled understanding of how LLMs work and how they can be made safer and more reliable. The application domain includes (but are not limited to) low-resource tasks,  multilingual understanding, misinformation and other problematic behaviour detection (narrative detection) or even LLM user simulation.

Relevant publications:

  • Vykopal, I., Pikuliak, M., Srba, I., Moro, R., Macko, D., and Bielikova, M., 2023. 2024. Disinformation Capabilities of Large Language Models. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 14830–14847, Bangkok, Thailand. Association for Computational Linguistics. https://aclanthology.org/2024.acl-long.793/ 
  • Zugecova, A., Macko, D., Srba, I., Moro, R., Kopal, J., Marcinčinová, K., and Mesarčík, M., 2025. Evaluation of LLM Vulnerabilities to Being Misused for Personalized Disinformation Generation. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 780–797, Vienna, Austria. Association for Computational Linguistics. https://aclanthology.org/2025.acl-long.38/

The research will be performed at the Kempelen Institute of Intelligent Technologies (KInIT, https://kinit.sk) in Bratislava in cooperation with researchers from highly respected research units. A combined (external) form of study and full employment at KInIT is expected.

Supervising team
Mária Bieliková
Lead researcher, KInIT
Branislav Pecher
Addressing small labeled data in machine learning training

Low-Resource Text Processing with the Help of Data Synthesis

Supervisor: Jakub Šimko (supervisor, KInIT),

Supervising team: Jakub Šimko (supervisor, KInIT), Ján Čegiň (KInIT)

Keywords: low-resource AI, LLM, data synthesis, data augmentation, text processing

Viac info Menej info

Despite the proliferation of large language models (LLMs) with their proclaimed universal applicability, many tasks of automatic text processing remain insufficiently solved. Tasks that need to be solved in low-resource languages are solved with lower success. Tasks that require sensitivity to details, context, or fresh information also remain problematic. Examples at the intersection of both these dimensions include processing data from social media for the purpose of detecting narratives, analyzing discourse, detecting toxic content, detecting organized influence campaigns, revealing manipulation, supporting fact-checking, or auditing social media recommenders.

Since the above problem can be understood as a lack of data for specific tasks, one of the solutions can be to use the capabilities of LLMs, but not directly as task solvers, but rather for generation or augmentation of data samples. Using these, specialized (smaller) models could be created for specific tasks. The advantage of such an approach is a higher degree of control over the final model (we can control the generated data), as well as a lower price and footprint of the final model (it is smaller). On the other hand, the disadvantage of such an approach is its lower straightforwardness and methodological ambiguity.That ambiguity, however, offers an opportunity for research.

The topic of the dissertation envisions research of methods and methodologies for creating automatic approaches for classification (or other automatic processing) of texts in heterogeneous and unstable domains. These are characterized by a lack of resources,  primarily labeled data, but secondarily also limited computing power.

Relevant publications:

  • Cegin, J., Simko, J. and Brusilovsky, P., 2023. ChatGPT to Replace Crowdsourcing of Paraphrases for Intent Classification: Higher Diversity and Comparable Model Robustness. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing https://aclanthology.org/2023.emnlp-main.117/ 
  • Cegin, J., Pecher, B., Simko, J., Srba, I., Bieliková, M. and Brusilovsky, P., 2024, August. Effects of diversity incentives on sample diversity and downstream model performance in LLM-based text augmentation. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics https://aclanthology.org/2024.acl-long.710/ 

The research will be performed at the Kempelen Institute of Intelligent Technologies (KInIT, https://kinit.sk) in Bratislava in cooperation with industrial partners or researchers from highly respected research units. A combined (external) form of study and full employment at KInIT is expected.

Supervising team
Jakub Šimko
Lead researcher, KInIT
Ján Čegiň
Machine learning with human in the loop

Meta-Evaluation of Multilingual Texts by Large Language Models

Supervisor: Dominik Macko (supervisor, KInIT)

Keywords: generative AI, large language models, metrics and evaluation, machine learning

Viac info Menej info

The multilingual capabilities of large language models (LLMs) enable us to generate texts in many languages, or even to analyze the existing multilingual texts. However, there are missing standard methods and metrics to measure the quality of multilingual texts. Analysis, evaluation, and annotation of such texts by humans is costly and time-consuming, while the replicability of the annotations is questionable (due to ensuring the annotators consistency). In some languages, it is quite difficult to obtain human annotations, while also balancing the level of expertise and demographic diversity of annotators across languages. For a while, the researchers are experimenting with using LLMs themselves to evaluate various aspects of text quality (e.g., coherence, grammar correctness, linguistic acceptability) in multiple languages, whether the texts are generated by language models or written by humans. Even the LLM annotations can be biased due to internal biases of the meta-evaluation models caused by training data.

It is important to continually increase the robustness of meta-evaluation (to limit biases and increase objectivity), whether by increasing the number of evaluation LLMs or by ensuring diversity via LLM (person-like) behavior modifications. The feasibility of the meta-evaluation method must be considered by assessing the computational efficiency and overall practical usability. Also, it is required to measure the correlation between meta-evaluation and human judgement and to evaluate for which aspects of text quality and in which languages is such meta-evaluation a suitable and usable alternative.

Relevant publications:

  • Macko, D., Kopal, J., Moro, R., and Srba, I., 2025. MultiSocial: Multilingual Benchmark of Machine-Generated Text Detection of Social-Media Texts. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 727–752, Vienna, Austria. Association for Computational Linguistics. https://aclanthology.org/2025.acl-long.36/
  • Zugecova, A., Macko, D., Srba, I., Moro, R., Kopal, J., Marcinčinová, K., and Mesarčík, M., 2025. Evaluation of LLM Vulnerabilities to Being Misused for Personalized Disinformation Generation. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 780–797, Vienna, Austria. Association for Computational Linguistics. https://aclanthology.org/2025.acl-long.38/

The research will be performed at the Kempelen Institute of Intelligent Technologies (KInIT, https://kinit.sk) in Bratislava in cooperation with industrial partners or researchers from highly respected research units involved in international projects. A combined (external) form of study and full employment at KInIT is expected.

Supervising team
Dominik Macko
Senior researcher, KInIT

Improving Human–LLM Interaction

Supervisor: Róbert Móro (supervisor, KInIT)

Keywords: generative AI, large language models, human-LLM interaction, LLM alignment

Viac info Menej info

Large language models (LLMs) are increasingly being used for a wide range of tasks and in various contexts in an interactive (chat-like) manner. The challenge is to make this interaction aligned with human expectations and values while avoiding reinforcing biases or negative social behavior of a model such as sycophancy, manipulativeness, etc. To do this, we need to be able to measure the extent of a positive or a negative behavior of the model and (ideally) have large datasets of human-LLM interactions together with human preferences that can be used to tune the models. For the latter, synthetic data generated by models trained on actual users’ data can sometimes be used instead.

The main goal is to improve the human-LLM interaction by using human inputs (such as their preferences) to fine-tune LLMs’ conversational capabilities while controlling for unwanted behavior (such as biases, sycophancy, manipulativeness, etc.). Alternatively, the topic can focus on measurement of the negative or positive social behavior of LLMs in various situations and contexts, or on improving safety of the models while taking the context and user preferences into account. 

The application domains are diverse, including (but not limited to) auditing of LLMs’ safety with respect to various user groups (including children and young adults), misuse of LLMs for disinformation generation, or credibility signals detection and assessment.

Relevant publications:

  • Aneta Zugecova, Dominik Macko, Ivan Srba, Robert Moro, Jakub Kopál, Katarína Marcinčinová, and Matúš Mesarčík. 2025. Evaluation of LLM Vulnerabilities to Being Misused for Personalized Disinformation Generation. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 780–797, Vienna, Austria. Association for Computational Linguistics. https://aclanthology.org/2025.acl-long.38/ 
  • Ivan Srba, Olesya Razuvayevskaya, João A. Leite, Robert Moro, Ipek Baris Schlicht, Sara Tonelli, Francisco Moreno García, Santiago Barrio Lottmann, Denis Teyssou, Valentin Porcellini, Carolina Scarton, Kalina Bontcheva, and Maria Bielikova. 2025. A Survey on Automatic Credibility Assessment Using Textual Credibility Signals in the Era of Large Language Models. ACM Trans. Intell. Syst. Technol. Just Accepted (September 2025). https://doi.org/10.1145/3770077 

The research will be performed at the Kempelen Institute of Intelligent Technologies (KInIT, https://kinit.sk) in Bratislava in collaboration with industrial partners or researchers from highly respected research units involved in international projects. A combined (external) form of study and full employment at KInIT is expected.

Supervising team
Róbert Móro
Senior researcher, KInIT

User Activity Prediction in Social Media Environments

Supervisor: Ivan Srba (supervisor, KInIT)

Keywords: user interaction prediction, generative AI, large language models

Viac info Menej info

The proliferation of AI-driven recommendation systems on social media has intensified the need for methods that can anticipate how users engage with presented content. While current “digital twin” approaches simulate the behaviour of specific individuals, this PhD project is motivated by the need to predict the next interaction of user archetypes, i.e., personas defined by demographic and interest attributes. With the rapid evolution of generative AI models, including LLMs and multimodal systems, their capabilities to forecast user actions such as clicking, liking, skipping, or dwelling on social-media content remain unexplored.

The core objective of the PhD research is to investigate and compare AI-based approaches for predicting the next user interaction in social media environments, focusing on user archetypes rather than individuals. To do this, the work will explore generative modelling strategies (e.g., LLM-based predictors), multimodal encoders, and hybrid architectures capable of processing and reacting to streamed online content. A significant part of the thesis will address real-time content annotation, identifying which text/image/video attributes best predict behavioural responses, and how Parameter-Efficient Finetuning Techniques (PEFTs) can be used to adapt models rapidly and cost-effectively. The project will also analyse evaluation methodologies for sequential user-interaction forecasts, including simulation-based benchmarks that reflect realistic platform dynamics.

The domain of application will be algorithmic auditing, a research practice that evaluates how algorithmic systems operate and whether they comply with legal, ethical, or performance expectations. More accurate next-interaction prediction models can make these auditing approaches substantially more authentic, enabling simulated agents to behave in a more organic, human-like manner when interacting with recommender systems. Such improvements can strengthen audit validity, allow testing at scale, and produce a deeper understanding of the impact that recommender design has on different user archetypes.

Relevant publications:

  • Ivan Srba, Branislav Pecher, Jakub Simko, Robert Moro, Maria Bielikova. 2025. Model-based Algorithmic Auditing of Social Media AI Algorithms. In Proceedings of Fourth European Workshop on Algorithmic Fairness, Vol. 294. 439-445. https://proceedings.mlr.press/v294/srba25a.html
  • Matej Mosnar, Adam Skurla, Branislav Pecher, Matus Tibensky, Jan Jakubcik, Adrian Bindas, Peter Sakalik, and Ivan Srba. 2025. Revisiting Algorithmic Audits of TikTok: Poor Reproducibility and Short-term Validity of Findings. In Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’25). Association for Computing Machinery, New York, NY, USA, 3357–3366. https://doi.org/10.1145/3726302.3730293

The research will be performed at the Kempelen Institute of Intelligent Technologies (KInIT, https://kinit.sk) in Bratislava in collaboration with industrial partners or researchers from highly respected research units involved in international projects. A combined (external) form of study and full employment at KInIT is expected.

Supervising team
Ivan Srba
Senior researcher, KInIT

Using Machine Learning to Solve Energy and Environmental Problems

Supervisor: Viera Rozinajová (supervisor, KInIT)

Supervising team: Viera Rozinajová (supervisor, KInIT),  Anna Bou Ezzeddine (KInIT), Gabriela Grmanová (KInIT)

Keywords: machine learning models, scientific machine learning, knowledge representation, transfer learning, energy domain, environment

Viac info Menej info

Current challenges that the energy sector faces today include maintaining stability in the energy grid, aggregation of flexibility, operation of energy communities, smart charging of electric vehicles, and efficient use of battery storage. Renewable resources are strongly dependent on weather, so weather prediction is also a focus of interest.

The ENVIRO team at KInIT is dedicated to the aforementioned tasks, whereas the proposed solutions are based on machine learning and artificial intelligence. We investigate and experiment with various methods such as taking into account physical laws in machine learning models (physics-informed machine learning), transfer or federative learning, as well as reinforcement learning, a currently popular approach used to optimize decision-making in a dynamic environment. Foundational models and their comparison with traditional approaches also offer an interesting direction of research. We solve tasks that process time series as well as image data using computer vision methods.

The aim of the PhD thesis will be to propose a solution to a selected energy or environmental problem (for example, flexibility aggregation or climate modeling/weather forecasting) by applying advanced artificial intelligence approaches. The selection of appropriate machine learning methods will depend on the specifics of the problem being solved and the choice of the doctoral candidate.

Relevant publications:

The research will be performed at the Kempelen Institute of Intelligent Technologies (KInIT, https://kinit.sk) in Bratislava in cooperation with industrial partners or researchers from highly respected research units from abroad. A combined (external) form of study and full employment at KInIT is expected.

Supervising team
Viera Rozinajová
Lead researcher, KInIT
Anna Bou Ezzeddine
Senior researcher, KInIT
Gabriela Grmanová
Senior researcher, KInIT

Improving Natural language Processing

Supervisor:  Marián Šimko (supervisor, KInIT)

Supervising team:  Marián Šimko (supervisor, KInIT), Jana Kosecka (George Mason University)

Keywords:  large language models, natural language processing, trustworthy NLP, multilingual learning, information extraction

Viac info Menej info

The rapid progress of large language models (LLMs) and other foundation models has transformed natural language processing (NLP), enabling powerful applications such as conversational agents, code assistants, and advanced information access systems. These technologies already affect everyday life and are reshaping how organisations work with text and knowledge.

Despite their success, LLMs still face important open challenges. Many models behave as black boxes, can hallucinate content, reproduce social biases, or fail when moved to new domains, tasks, or low-resource languages. At the same time, there is growing demand for models that are trustworthy, efficient, and easy to adapt, while respecting data, safety, and regulatory constraints.

This PhD project will investigate methods for understanding, adapting, and controlling large language models, with a particular focus on trustworthiness and low-resource / domain-specific NLP. The work can combine theory, modelling, and practical experimentation on real-world datasets.

Illustrative research challenges include:

  • Properties of LLMs – analysing and mitigating hallucinations, robustness issues, and failure modes.
  • Trustworthy NLP – bias detection and mitigation, safety and reliability, explainability and interpretability of model decisions.
  • Adapting LLMs to specific contexts and tasks – e.g. parameter-efficient fine-tuning (PEFT), retrieval-augmented generation (RAG), instruction tuning, or continual learning.
  • Advanced learning techniques – transfer and multilingual learning, active learning, low-shot and low-resource methods for under-represented languages.
  • Domain-specific NLP applications – information extraction and text classification for specialised domains (e.g. legal, financial, technical or customer-support texts), improving dialog quality in conversational agents, or building task-oriented assistants.

The topic offers substantial flexibility and can be tailored to the candidate’s interests, ranging from more theoretical work on model behaviour to application-driven research in collaboration with academic or industrial partners.

Relevant publications:

  • Vykopal, I., Ostermann, S., Simko, M. Soft Language Prompts for Language Transfer. In Proceedings of the 2025 Conference of the NAACL: Human Language Technologies (Volume 1: Long Papers), pages 10294-10313, ACL, 2025. https://doi.org/10.18653/v1/2025.naacl-long.517 
  • Pikuliak, M., Hrčková, A., Oreško, Š., Šimko, M. Women Are Beautiful, Men Are Leaders: Gender Stereotypes in Machine Translation and Language Modeling. In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 3060-3083, ACL, 2024. https://doi.org/10.18653/v1/2024.findings-emnlp.173   

The research will be performed at the Kempelen Institute of Intelligent Technologies (KInIT, https://kinit.sk) in Bratislava in cooperation with industrial partners or researchers from highly respected research units from abroad. A combined (external) form of study and full employment at KInIT is expected.

Supervising team
Marián Šimko
Lead researcher, KInIT
Jana Kosecka
Professor, George Mason University, USA

Competent Reasoning and Reliable Execution – Towards Truly Capable LLMs

Supervisor:  Michal Gregor (supervisor, KInIT)

Supervising team:  Michal Gregor (supervisor, KInIT), Marián Šimko (KInIT)

Keywords: large language models, AI, reasoning, machine learning, deep learning, LLM limitations, reinforcement learning, mechanistic interpretability

Viac info Menej info

Large language models (LLMs) are currently driving much of the progress in the area of artificial intelligence, especially due to their ability to perform a wide range of tasks zero-shot – based on a natural language description of the task – and due to the large amount of world knowledge they possess.

However, the current generation of LLMs still struggles in several important ways that bar them from becoming truly capable at a wide range of challenging tasks; these struggles include e.g. their:

  • Limited ability to correctly and robustly execute algorithms: Even frontier models currently struggle with correct execution of algorithms, especially when such execution is prolonged. Even models that theoretically support input contexts with millions of tokens, struggle to maintain accuracy as the context length grows.
  • Limited ability to bootstrap complex reasoning: While models are now being explicitly trained to perform deliberative reasoning, multiple works demonstrate that the current generation of training methods used in this context (such as GRPO and GSPO) do not successfully bootstrap to new, improved reasoning skills – rather, they just reinforce reasoning paths that the base model already knew from pretraining.

Apart from these, there is also the limited ability to efficiently learn from experience (which is a component in how humans achieve true competence – we possess powerful priors, but we can also adapt quickly) and many other struggles. While this is – to a very considerable extent – true of all current models, the issues are typically much more pronounced for small models. This, of course, has serious implications for a lot of applications that require compact models, which can be run locally.

The aim of this research is to:

  • Diagnose: Examine where the current models and training approaches struggle and why. In this phase, the student can draw upon existing insights (such as works on over-squashing and representational collapse in transformers), form their own hypotheses and validate them using the tools of mechanistic interpretability, use analogies with cases where similar approaches work (e.g. reinforcement learning methods like AlphaZero and MuZero – or even analogies from cognitive science regarding different types of human memory and such), etc.
  • Fix: Use insights formed in the diagnosing phase to mitigate some of the issues – by adjusting the identified weaknesses, such as certain constraints imposed by the architecture, the training methodology, the training data and how it is preprocessed, etc.

Relevant publications:

  • Gurgurov, Daniil, Michal Gregor, Josef van Genabith, Simon Ostermann. „On Multilingual Encoder Language Model Compression for Low-Resource Languages“. International Joint Conference on Natural Language Processing & Asia-Pacific Chapter of the Association for Computational Linguistics, 2025
  • Vykopal, I., Pikuliak, M., Ostermann, S., Anikina, T., Gregor, M. and Šimko, M., 2025. Large Language Models for Multilingual Previously Fact-Checked Claim Detection. EMNLP 2025 (pp. 15741–15765).

The research will be performed at the Kempelen Institute of Intelligent Technologies (KInIT, https://kinit.sk) in Bratislava in cooperation with industrial partners or researchers from highly respected research units from abroad. A combined (external) form of study and full employment at KInIT is expected.

Supervising team
Michal Gregor
Lead researcher, KInIT
Marián Šimko
Lead researcher, KInIT

Deadlines

    1. Express your interest — anytime (for academic year 2026/2027 no later than 31.3.2026)
      If you’re interested in the PhD study program at KInIT, let us know as soon as possible.
    2. Participate in testing — April 2026
      Test your knowledge and match with KInIT values.
      You’ll take part in systematic selection steps such as team activity, knowledge review test, etc.
    3. Find a topic and a supervisor match — May 2026
      We’ll get back to you and organize a matchmaking for a research topic and/or supervisor.
    4. Wait for the decision — July 2026
      Wait for the admission processes at KInIT and FIT VUT Brno (don’t worry, we’ll guide you through the application submission process) to finish. We’ll let you know our final decision.
    5. Start at KInIT
      If we choose you, we’ll provide you with further information on your work contract at KInIT.
      PhD Study at FIT VUT Brno officially starts on 1 September 2026.

How the PhD study at KInIT works

1

Work at KInIT & study at FIT VUT

Viac info Menej info

We provide PhD study in collaboration with the Faculty of Information Technology at VUT, Brno (Czechia). You’ll be enrolled at the faculty as an external doctoral student. After you’re accepted for study, you’ll work as a full-time employee at our institute. Your doctoral research will be your primary job role.

2

Carry out research in a selected topic

Viac info Menej info

You can choose from multiple interesting topics of your dissertation. They are based on different mentoring schemes:

  1. Research topics supported by industrial research partners
  2. Research topics mentored by top scientists from abroad
  3. Research topics connected to international projects

You’ll have access to an expert supervising team. Depending on the selected scheme, it can include industrial partners and/or a scientist from a renowned foreign university.

3

Join our research teams

Viac info Menej info

While working at our institute you’ll be a member of a KInIT research team. The teams collaborate on various projects or on other activities:

4

Change the world

Viac info Menej info

The rest is up to you! Turn your curiosity and passion for intelligent technologies into new discoveries or inventions making the world a better place.

KInIT supervisors

Mária Bieliková is an expert researcher at KInIT. She focuses on human-computer interaction analysis, user modeling and personalization. Recently, she has been working in data analysis and modeling of antisocial behavior on the Web. She is active in discussions on trustworthy AI at the national and European levels. Maria has supervised 19 successful doctoral graduates to date. She co-authored 70+ journal publications, 200+ conference papers, received 4,400+ citations (Google Scholar h-index 30), and serves on the editorial board of two CC journals. She has been the principal investigator in 40+ research projects.

Michal Gregor is an expert researcher at KInIT. He focuses especially on artificial neural networks and deep learning, on reinforcement learning, and more recently on multi-modal learning and learning that involves language supervision. Michal also has experience in other areas of AI such as metaheuristic optimization methods, representation of uncertain knowledge, probabilistic models and more.

Michal Kompan is an expert researcher at KInIT. He focuses on recommender systems, machine learning, user modeling, and information retrieval. His research is focused on predictive modeling and customer behavior (e.g., churn prediction, next-item recommendation), as well as content-based adaptive models. Michal serves as a reviewer or/and program committee member at several international conferences, such as RecSys, SIGIR, WWW, ADBIS, Hypertext, UMAP and SMAP.

Viera Rozinajová is an expert researcher at KInIT. She is focusing on intelligent data analysis, particularly predictive modeling, cluster analysis, anomaly detection and optimization. Before her employment at KInIT, she worked as an associate professor at the Faculty of Informatics and Information Technologies at the Slovak University of Technology in Bratislava, where she headed up the Big Data Analysis group. She has authored/co-authored more than 70 publications in scientific journals and conferences and has participated in more than 25 national and international research projects and has led several of them. 

Jakub Šimko is an expert researcher at KInIT, where he also leads the Web and User Data Processing team. Jakub focuses on the intersection of human computation, machine learning and user modeling. He has recently been working on social media algorithm auditing, misinformation modeling and promotes interdisciplinary approaches to computer science research. He graduated from Slovak University of Technology in Bratislava, where, after receiving his PhD, he worked for 7 years as a researcher and teacher. He co-authored more than 30 internationally recognized publications, together receiving more than 350 citations.

Marián Šimko is an expert researcher at KInIT. Marián focuses on natural language processing, information extraction, low-resource language processing and trustworthiness of neural models. He is a former vice-dean for Master’s study and alumni co-operation at the Slovak University of Technology.

External mentors

Peter Brusilovsky is a Professor at the School of Computing and Information, University of Pittsburgh, where he directs the Personalized Adaptive Web Systems (PAWS) lab. His research is focused on user-centered intelligent systems in the areas of adaptive learning, recommender systems, and personalized health. He is a recipient of Alexander von Humboldt Fellowship, NSF CAREER Award, and Fulbright-Nokia Distinguished Chair. Peter served as the Editor-in-Chief of IEEE  Trans. on Learning Technologies, and a program chair for several conferences including RecSys.

Peter Dolog is an Associate Professor at the Department of Computer Science, Aalborg University, Denmark. His current research interests include machine learning and data mining in the areas of user behavior analysis and prediction, recommender systems, preference learning, and personalization. Peter is a senior member of ACM, served as a senior program commitee member of AI related conferences as well as a general chair of UMAP, HT and Web Engineering conferences.

Jana Kosecka is a Professor at the George Mason University. She is interested in computational models of vision systems, acquisition of static and dynamic models of environments by means of visual sensing, high-level semantic scene understanding and human-computer interaction. She held visiting positions at UC Berkeley, Stanford University, Google and Nokia Research, and served as Program chair, Area chair or senior member of editorial board for  leading conferences in the field CVPR, ICCV, ICRA.

Jana is currently mentor of our PhD student: Ivana Beňová

Branislav Kveton is a Principal Scientist at Adobe research. He proposes, analyzes, and applies algorithms that learn incrementally, run in real time, and converge to near-optimal solutions as they learn. He made several fundamental contributions to the field of multi-armed bandits. His earlier work focused on structured bandit problems with graphs, submodularity, and low-rank matrices, and ranked lists. His recent work focuses on making bandit algorithms practical

Peter Richtárik is a Professor of Computer Science & Mathematics at KAUST. He is one of the founders and a Fellow of the Alan Turing Institute. Through his work on randomized and distributed optimization algorithms, he has contributed to the foundations of machine learning and federated learning. He serves as an Area Chair of leading machine learning conferences, including NeurIPS, ICML and ICLR.

Martin Takáč is an Associate Professor and Deputy Department Chair of Machine Learning Department at MBZUAI, where he is a core faculty at the Optimization & Machine Learning Lab. His current research interests include the design and analysis of algorithms for machine learning, applications of ML, optimization, HPC. He serves as an Area Chair of ML conferences such as AISTATS, ICML, and NeurIPS.

Peter Tino is a Professor at the School of Computer Science, University of Birmingham, UK. He is interested in the interplay between mathematical modelling and machine learning (dynamical systems, probabilistic modelling, statistical pattern recognition, natural computation). Peter is interested in both foundational aspects and applications in interdisciplinary contexts (e.g. astrophysics, biomedical sciences, cognitive neuroscience).

He is a Fellow of the Alan Turing Institute and has served on editorial boards of leading journals such as IEEE TNNLS, IEEE TCYB, Neural Networks and Scientific Reports.

Your future
starts at KInIT

Transform your curiosity into excellence. If you are interested in doing a PhD degree at KInIT, let us know as soon as possible. Fill in the PhD @ KInIT expression of interest form no later than March 31st.

I have a question

Find our more about our external mentors and their research