Article written by Dr. Philippe Jacquet, Data Scientist, Division Calcul et Soutien à la Recherche (DCSR).

Have you ever wanted to ask the same questions to thousands of documents, compare several AI models, or turn a large collection of texts into structured research data?
These are precisely the kind of research tasks for which DCSR-LLM was developed at UNIL.
What is the DCSR?
The DCSR stands for Division Calcul et Soutien à la Recherche, or Scientific Computing and Research Support Unit in English. We are part of the Centre Informatique at UNIL. We are a team of about 20 people. Our role is to help the UNIL research community with computing, data storage, and technical support for research projects.
Some members of the team manage the computing infrastructure used at UNIL. Others work directly with researchers and provide scientific and technical consulting.
What kind of support does the DCSR provide?
The DCSR provides two main kinds of support.
The first is infrastructure: computers, storage space, and technical systems that researchers can use for demanding research tasks. The second is expertise: DCSR staff help researchers with topics such as scientific programming, machine learning, databases, and web development.
So the DCSR is not only a place with machines. It is also a support unit with people who help researchers use these tools in a useful way.
What are HPC clusters?
The term HPC means High-Performance Computing. It refers to powerful computing systems used for tasks that are too large, too slow, or too demanding for an ordinary laptop. An HPC cluster is a group of computers working together. Instead of doing everything on one machine, a cluster can distribute the work across many machines.
At UNIL, there are two HPC clusters: Curnagl and Urblauna. Their names are derived from Romansh bird names. They are used for demanding research tasks such as simulations, data analysis, machine learning, and AI workflows.
A simple way to think about it is this: if your laptop is enough, you use your laptop. If your work becomes too heavy, too slow, or needs more memory, the cluster becomes useful.
What are Large Language Models?
Large Language Models, or LLMs, are AI systems trained on very large amounts of text. They can answer questions, summarize documents, rewrite text, classify information, or extract facts from written material. Most people know them through tools such as ChatGPT, Claude, or Gemini.
What is Hugging Face?
Hugging Face is a platform where people share AI models and datasets. A simple comparison is that it is a bit like GitHub, but for machine learning models and data. It also provides software tools that make it easier to download and run open-source language models on a computer or a server.
What is LM Studio?
LM Studio is an application that lets people run some language models locally through a chat interface. In practice, it feels a bit like using ChatGPT, except that the model can run on your own computer. This is useful for simple local experiments.
Why did the DCSR create DCSR-LLM?
Because many researchers want to use AI, but their needs often go beyond a simple chatbot.
A chatbot is useful for asking a question or drafting a paragraph. But research often requires something more structured. Researchers may want to:
– test several models on the same task;
– work with a large collection of documents;
– keep data on UNIL infrastructure;
– save the exact settings used in an analysis;
– repeat the same workflow later.
This is why the DCSR developed DCSR-LLM.
What is DCSR-LLM?
DCSR-LLM is a toolkit developed at UNIL for working with large language models in a more structured way.
It allows researchers to download open-source models from Hugging Face, run them locally or on UNIL servers, compare them, evaluate them on specific tasks, extract structured information from text, adapt some models to more specialized uses, and export models to GGUF format for use with LM Studio.
It is not just a chatbot. It is a tool designed for research workflows.
Why not just use ChatGPT or Claude?
For many everyday uses, ChatGPT or Claude are very useful. But in research, three extra questions often matter.
The first is: where does the data go?Some projects involve sensitive, unpublished, or internal material. In such cases, researchers may want a more controlled environment.
The second is: which model am I using?
Different models behave differently. In research, it is often useful to compare them rather than rely on only one assistant.
The third is: can I repeat the same workflow clearly?
If an AI result matters for a project, researchers usually need to document how it was produced.
DCSR-LLM is useful because it helps with these three points.
What does “reproducible” mean here?
It means that the work is done in a way that can be repeated and documented. If you use the same model, the same data, and the same settings, you should be able to rerun the workflow and understand what happened. This matters in research because methods need to be described clearly.
What can DCSR-LLM do in practice?
DCSR-LLM can do several practical things. It can help a researcher inspect and download models from Hugging Face, run some of these models locally, compare several models on the same benchmark, turn unstructured text into structured data, and adapt a model for a more specific task.
These ideas become easier to understand with examples.
Can you give a first concrete example?
Imagine a researcher in the social sciences has 2,000 interview transcripts and wants to organize them.
Each interview is written in free text. The researcher wants, for each transcript, to extract a small set of clearly defined fields such as:
– the title or identifier of the interview;
– the date of the interview;
– the location;
– the name or profile of the person interviewed;
– the institution or organization mentioned;
– the main topic discussed;
– a short quotation supporting the extraction.
A normal chatbot can help with a few interview transcripts, one by one. But this becomes difficult if there are thousands of texts and the researcher wants the same structure every time.
With DCSR-LLM, the team can define the fields they want, run the same extraction process on the whole collection, save the results in a structured format, and then review them. The AI does not replace the researcher. But it can help turn a large collection of text into something easier to inspect and analyze.
Can you give a second concrete example?
Imagine a biology or medical research team wants to use AI to answer a set of domain-specific questions.
Before choosing a model, the team wants to know which one performs best on its task. If they only use chatbots manually, comparison is difficult. One person may ask slightly different questions. Another may use different wording. Results are harder to compare fairly.
With DCSR-LLM, the team can prepare one fixed list of questions and run the same evaluation on several models. They can then compare the outputs more systematically. Instead of saying, “this model feels better,” they can say, “we tested these models on the same task under the same conditions.” That is much closer to a research approach.
Can you give a third concrete example?
Imagine a researcher has a very large collection of texts, for example 10,000 journal articles or 10,000 YouTube transcripts.
The researcher may want to ask questions such as:
– Does this article discuss climate change, migration, or public policy?
– Is this speaker talking positively or negatively about artificial intelligence?
– Does this text mention a specific concept, author, or method?
– Is the main purpose of this article to explain, criticize, or compare?
– Does this transcript contain personal testimony, expert opinion, or political argument?
– Which passages talk about ethics, cost, or risk?
A normal chatbot can help with one article or one transcript at a time. But this becomes impractical when the collection contains thousands of texts.
With DCSR-LLM, the researcher can run the same workflow across the whole corpus in a more systematic way. Instead of manually copying text into a chatbot, they can process the collection in a structured manner and save the results for later analysis.
The value here is not just that the AI can answer questions. The value is that it becomes possible to ask the same research questions at scale, on a large body of text, in a way that is more organized and reproducible.
Can you give a fourth concrete example?
Imagine a team has found a model that works reasonably well, but not quite well enough for its own field. For example, the model may struggle with the vocabulary of a specific discipline, or with the exact format of answers needed for a project.
In that case, the team may want to adapt the model to a specific task or field and then measure whether the adaptation actually helped.
DCSR-LLM can support this kind of before-and-after workflow. A team can test the original model, adapt it for the task, and then test it again on the same benchmark. This gives a clearer answer than a simple impression such as “the new version seems better.”
So DCSR-LLM is not mainly for chatting?
Exactly. It can be used for chat, but that is not its main purpose. Its main purpose is to support structured work with language models: evaluation, extraction, comparison, and adaptation. That is why it is more useful to think of it as a research tool than as a chatbot.
Do I need to know how to code?
Not necessarily in depth, but some technical support is helpful.
DCSR-LLM is a command-line tool. This means it is used through a terminal rather than through a point-and-click web interface. Users do not need to be expert programmers, but they should be comfortable with a basic technical environment, or work with support from the DCSR or from technically experienced colleagues.
The important point is that DCSR-LLM is not reserved to AI specialists. But it is also not intended as a consumer product for completely casual use.
Where can DCSR-LLM run?
It can run on a personal computer for small experiments. It is also designed to run on UNIL infrastructure, including Curnagl and Urblauna.
This matters because a project can start small and then grow. A researcher may first test an idea on a laptop, then later run a larger workflow on the UNIL clusters.
Why is that useful?
Because many projects do not start with a huge investment. Researchers often begin with a simple question: “Can this model help with my material?”
If the first results are promising, they may then want to scale up: more documents, more models, more comparisons, more demanding computations. DCSR-LLM supports this gradual scaling.
What is the main idea behind DCSR-LLM?
The main idea is simple. DCSR-LLM helps researchers move from informal chatbot use to a more controlled way of working with AI. That means more clarity about the model used, where the data is processed, how the results are generated, and how the workflow can be repeated.
Does AI replace the researcher?
No. AI can help with some tasks, sometimes very effectively. But researchers still need to define the question, choose the method, review the outputs, and interpret the results. DCSR-LLM is useful because it helps organize AI-based work. It does not remove the need for scientific judgment.
So what is DCSR-LLM, in one sentence?
DCSR-LLM is a practical UNIL tool that helps researchers use large language models in a more concrete, controlled, and reproducible way than a simple chatbot interface.
Useful links
– Repository: https://git.dcsr.unil.ch/Scientific-Computing/dcsr-llm
– More information: https://wiki.unil.ch/ci/link/2266
– Contact: helpdesk@unil.ch (subject: DCSR-LLM)