Random presentations (both in Spanish and English)
This project is maintained by jgbarah
Madrid, Spain, May 11, 2026
Jesus M. Gonzalez-Barahona
https://jgbarah.github.io/presentations/
But also:


https://arxiv.org/abs/2403.13784
Open model: architecture, model parameters, weights & metadata, tech report, model card, data card
Open tooling: open model plus training, inference, and evaluation code, libraries, evaluation data
Open science: open tooling plus paper, datasets, log files, intermediate models parameters, weights & metadata
The “four freedoms” from “What is Free Software?”
Access: can you run inferences the way you want?
Control on the model: can you modify the way the model works? (eg, finetune it)
Control on your data: can you control the prompt, the results?
Autonomy: how much you depend on the model provider?
Trust: can you ensure the model works as intended? (eg, backdoors, etc.)
Model access: Use cases, innovation, integration
Model control: Use cases, innovation, integration
Data control: Privacy, ndependence, reliance
Autonomy: Market competition, independence, reliance
Trust: Security, transparency.
An application uses one or more models to provide some service
It can be a local or cloud application
The application may be generalist or very specific
The model may change over time




Access: use the model only in the intended way
Model control: none
Data control: none
Autonomy: none
Trust: none
Access usually via HTTP API
The API defines to which extent the model can be controlled
Libraries and SDKs may be available
Designed for building apps, depending on the API
The model may change over time


Access: use as you want, but API restricts parameters
Model control: limited, depending on the API
Data control: none
Autonomy: none
Trust: none
Weights are available
Usually, software for inferences is available / FOSS
Can be run on trusted infrastructure
Finetuning, etc. is usually possible
Redistribution, modification, use may be conditioned or forbidden
In some cases, referred as “open weight models”
Access: use as you want if conditions are met
Model control: deep control, if conditions are met
Data control: complete
Autonomy: depends on the conditions
Trust: none
LLaMa3 models (2025-04)
Allows use, redistribution, derived works
No conditions for use
Derived works: finetuning, integration…
Does not require information about the model, its training, etc. (no freedom of study)
Open-Weight AI Models: What They Are, and Why OpenAI’s Next Move Matters
Access: use as you want
Model control: deep control
Data control: complete
Autonomy: only study is restricted
Trust: none
Allows use, redistribution, derived works
No conditions for use
Derived works: finetuning, integration…
Open source software for training, inferencing
Detailed description of training, doesn’t require availability of the training dataset
Proposal – Interpretation of DFSG on Artificial Intelligence (AI) Models
Access: use as you want
Model control: deep control
Data control: complete
Autonomy: detailed study is restricted
Trust: partial
Allows use, redistribution, derived works
No conditions for use
Derived works: finetuning, integration…
All information about the model
Requires availability of the training dataset
Access: use as you want
Model control: deep control
Data control: complete
Autonomy: complete
Trust: complete

| Access | Model Control | Data Control | Autonomy | Trust | |
|---|---|---|---|---|---|
| Behind-app | App-defined | None | None | None | None |
| Directly accessible | API restrictions | API restrictions | None | None | None |
| Available weights | With conditions | With conditions | Complete | With conditions | None |
| Open weight | Use as you want | Deep control | Complete | Study restricted | None |
| Open source | Use as you want | Deep control | Complete | Detailed study restricted | Partial |
| Reproducible | Use as you want | Deep control | Complete | Complete | Complete |
Reproducible AI: Why it Matters & How to Improve it
Guidelines for Empirical Studies in Software Engineering involving Large Language Models
Conditions on use: “ethical use”
Conditions on training: “ethical datasets”
Depends on what is considered as “ethical”

At least, “available weights” if inference code is available
Self-hostable models


Technical skills required!
You can also deploy in a cloud-based host
You have to do the math
safetensors filesHuggingFace Guide on Quantization
Adapting a model to a specific task with a (usually smaller) specialized dataset
Starting point: weights of given model.
Adjust weights to fit the specialized dataset
Adapter: new layers of weights added to a model to finetune it

https://civitai.com/
LangChain: Chain together large LLM operations into sophisticated workflows, usually to build agent tools
LiteLLM: Agile toolset designed for efficiency and simplicity.
Both can use local models, of models via HTTP API
Ollama, based on llama.cpp
Open WebUI: “almost” FOSS
Jan: Local assistant
Most of them also provide an HTTP API
curl -fsSL https://ollama.com/install.sh | sh
ollama serve
ollama run gemma3:1b
curl http://localhost:11434/api/generate -d '{
"model": "gemma3:1b",
"prompt":"Why is the sky blue?"
}'
Using Ollama to host an LLM on CPU-only equipment to enable a local chatbot and LLM API
uv venv --python 3.11
uv pip install open-webui
uv run open-webui serve
Now, open http://localhost:8080

sudo pkg -i Jan_0.6.9_amd64.deb
Jan

Qwen-Image, Apache 2.0 (2025-08)
HiDream-I1, MIT License (2025-07)
FLUX.1Kontext[dev], models, Flux Non-Commercial License (2025-08)
Stable Diffusion, models, StabilityAI Community License (2025-01)
A Guide to Open-Source Image Generation Models
ComfyUI, repo: front-end and UI for several self-hostable text-to-image and text-to-video models
Wan2GP: front-ed and UI for several self-hostable text-to-video models
CivitAI: models and finetunes
Whisper, MIT License
uv venv
uv pip install openai-whisper
uv run whisper speech.wav --language Spanish
#!/usr/bin/python3
import whisper
model = whisper.load_model('tiny')
transcription = model.transcribe('recording.wav')
print(transcription['text'])
$ tts --text "Texto" \
--model_name tts_models/es/mai/tacotron2-DDC \
--out_path speech.wav
Kokoro, Apache 2.0 (2025-01)
Higgs Audio V2, Boson Higgs Audio 2 Community License (2025-07)
Chatterbox, MIT License (2025-04)
MeloTTS & OpenVoice v2 MIT License (2024-02, 2024-04)
FishSpeech, CC Attribution-NonCommercial-ShareAlike (2025-08) (2024-11)
Exploring the World of Open-Source Text-to-Speech Models
Understanding and reasoning about time series: ChatTS Apache 2.0, includes training dataset (2025-08)
360 immersive and explorable 3D worlds: HunyuanWorld
Text to 3D: LlamaMesh, Llama Community License (2024-11)
Open model and tools for building videos with AI: OpenSora
Hive: orchestrator for coding agents
AnythingLLM: assistant, agentic
FastSDCPU: image generation optimized for CPU
OpenClaw: Agentic system
Hermes agent: Agentic system
Autoresearch: Self-improvign agent
Good night, have fun: Autoresearch-style agent, for coding
SurfSense: comprehensive assistant
DeerFlow: comprehensive assistant
Hyprnote: note taking tool for meetings
TransformerLab: Train, Tune, Chat with LLMs
MobiRAG: chat with PDFs in your mobile
Common Corpus: The Largest Collection of Ethical Data for LLM Pre-Training
The Common Pile v0.1: An 8TB Dataset of Public Domain and Openly Licensed Text
Reproducible research: we want experiments that can be repeated by others, with the same results
Reproducible results: in some cases, it is important to be sure that a given input produces a given output. Always.
Reliable debugging: for debugging a problem, exact reproduction is often needed
Deterministic software: in some cases, we need software that is deterministic. Always.
Once weights are settled… the network itself doesn’t change
Determinism depends on:
It is “regular”, deterministic software…
except when it tries to be random
seed can be fixed (initializes the pseudo-random number generator)temperature: 0 means “greedy sampling” (most probable next token)top_k (shortlist selector pool): 1 means the pool for selectable words is 1 (the most likely)top_p (nucleus sampling): chains of tokens to be considered. Difficult to control, interferes with top_kOther parameters (eg, frequency or presence penalty) should be equal
Beware: the software may use random number generators in some other places
Randomness in inference is there for a reason: it can be useful for creativity, for getting better outputs
Two strategies:
temperature, top_k, top_pBoth can be combined
Mixtures of experts: prompt tokens routed differently depending on composition of batches from different users
Framework (PyTorch, TensorFlow): non-deterministic convolution algorithms (can be configured to be deterministic)
Differences in compilers, GPU drivers…
Achieving Consistency and Reproducibility in Large Language Models (LLMs)
Solving Reproducibility Challenges in Deep Learning and LLMs: Our Journey
temperature = 0top_k = 1