flowchart LR
W[Weights<br/>Learned numbers] --> M((Model))
C[Code<br/>Instructions] --> M
M --> T[Does tasks]
From Using to Contributing (AI4LAM Fantastic Futures 2025 Workshop)
Hugging Face
2025-12-09
“Let’s build a chatbot for our collections!”
But is this the best starting point?
“A frontier model like GPT-4 is like a Ferrari. It’s an obvious triumph of engineering, designed to win races. But it takes a special pit crew just to change the tires.
In contrast, a smaller specialized model is like a Honda Civic. It’s engineered to be affordable, reliable, and extremely useful. And that’s why they’re absolutely everywhere.”
— Adapted from “Finally, a Replacement for BERT”
flowchart LR
W[Weights<br/>Learned numbers] --> M((Model))
C[Code<br/>Instructions] --> M
M --> T[Does tasks]
The weights are the “brain” - patterns learned from training data
flowchart TD
D[(Training Data<br/>Books, websites, images)] --> L[Learning Process]
L --> W[Weights File<br/>.safetensors]
W --> E["Billions of numbers:<br/>[0.023, -0.891, 0.442, ...]"]
These numbers encode everything the model “knows”
Open Weights
flowchart TD
O[Download weights] --> R[Run anywhere]
R --> I[Inspect & modify]
Closed Weights
flowchart TD
A[API only] --> B[Black box]
B --> V[Vendor controlled]
True open source AI includes:
Examples of truly open models:
| Choice | 1M+ models vs ~100 closed APIs |
| Control | Run where you want, pin versions |
| Flexibility | Fine-tune for your domain |
| Cost | Often cheaper at scale |
| Privacy | Data never leaves your infra |
| Transparency | Inspect model & training data |
AI/ML is more than LLMs!
Many tasks don’t need a large language model:
Finding models: Hugging Face Hub - 1M+ open models, filterable by task, language, size
Open models != local models
A few genres of inference for open models:
| Approach | Setup | Best For |
|---|---|---|
| Pay per token | Minimal | Prototyping, low volume |
| Local hardware | Medium | Privacy, offline use |
| Rent hardware | Higher | Production, scale |
Setup: Minimal - use via OpenAI-compatible client
Pros
Cons
Setup: Medium - install runtime + download model
Pros
Cons
Setup: Higher - configure cloud instance
Pros
Cons
mindmap
root((Local model ecosystem))
Python Libraries
Pytorch / Tensorflow / JAX
Transformers
Diffusers
Sentence Transformers
ONNX Runtime
Optimum
JavaScript
Transformers.js
Inference Frameworks
GPU Poor focused
llama.cpp
LM Studio
Ollama
MLX Apple Silicon
mlx-lm
mlx-vlm
ExLlamaV2
Jan
Enterprise Focused
vLLM
SgLang
Triton Inference Server
Quantization Formats
GGUF
GPTQ
AWQ
UIs
Gradio
Open WebUI
flowchart LR
Start([Start: No Training Data]) --> TestSOTA[Test with SOTA Closed Model<br/>Validate 'any model can do it']
TestSOTA --> CheckWorking{Working?}
CheckWorking -->|No| IteratePrompts[Iterate on Prompts]
IteratePrompts --> CheckIteration{Working Now?}
CheckIteration -->|Yes| TryOpenWeights[Try Open Weights Model]
CheckIteration -->|No| Rethink[Rethink Task]
Rethink --> TestSOTA
CheckWorking -->|Yes| TryOpenWeights
TryOpenWeights --> OpenWorking{Working?}
OpenWorking -->|Yes| TrySmaller[Try Smaller<br/>Open Weights Model]
TrySmaller --> CheckSize{Small Enough<br/>& Working Well?}
CheckSize -->|Yes<br/>Small & Working| Deploy1[Deploy & Capture Traces]
Deploy1 --> End1([Done])
CheckSize -->|No<br/>Too Big but Working| CreateData[Use Model to<br/>Create Training Data]
CheckSize -->|No<br/>Not Working Well| TestSOTA
CreateData --> FineTune[Fine-tune Smaller Model]
FineTune --> TestSmaller[Test Smaller Model]
TestSmaller --> Deploy2[Deploy & Share<br/>with Community]
Deploy2 --> End2([Done])
style Start fill:#e1f5ff
style TestSOTA fill:#fff4e1
style Deploy1 fill:#d4edda
style Deploy2 fill:#d4edda
style End1 fill:#d4edda
style End2 fill:#d4edda
style CheckWorking fill:#fff3cd
style CheckIteration fill:#fff3cd
style CheckSize fill:#fff3cd
style OpenWorking fill:#fff3cd
Switch to notebook
How can GLAMs contribute to Open Source AI?
GLAMs hold unique assets for AI training:
Move beyond “vibe checks” with domain expertise:
Share trained models back to the community:
danielvanstrien.xyz | Open Source AI for GLAMs