AI & ML interests
None defined yet.
Recent Activity
View all activity
chs20
authored a
paper about 1 month ago
chs20
submitted a
paper to Daily Papers about 1 month ago
nv-nguyen
authored a
paper about 2 months ago
julien-c
submitted a
paper to Daily Papers about 2 months ago
Post
3961
👉 What happened in AI in 2025? 👈
We prepared the 2025 version of the HF AI Timeline Grid, highlighting open vs API-based model releases, and allowing you to browse and filter by access, modality, and release type!
Play with it here:
2025-ai-timeline/2025-ai-timeline
Here's my personal quarterly TL;DR:
1️⃣ Q1 — Learning to Reason
Deepseek not only releases a top-notch reasoning model, but shows how to train them and compete with closed frontier models. OpenAI debuts Deep Research.
Significant milestones: DeepSeek R1 & R1-Zero, Qwen 2.5 VL, OpenAI Deep Research, Gemini 2.5 Pro (experimental)
2️⃣ Q2 — Multimodality and Coding
More LLMs embrace multimodality by default, and there's a surge in coding agents. Strong vision, audio, and generative models emerge.
Significant milestones: Llama 4, Qwen 3, Imagen 4, OpenAI Codex, Google Jules, Claude 4
3️⃣ Q3 — "Gold" rush, OpenAI opens up, the community goes bananas
Flagship models get gold in Math olympiads and hard benchmarks. OpenAI releases strong open source models and Google releases the much anticipated nano-banana for image generation and editing. Agentic workflows become commonplace.
Significant milestones: Gemini and OpenAI IMO Gold, gpt-oss, Gemini 2.5 Flash Image, Grok 4, Claude Sonnet 4.5
4️⃣ Q4 — Mistral returns, leaderboard hill-climbing
Mistral is back with updated model families. All labs release impressive models to wrap up the year!
Significant milestones: Claude Opus 4.5, DeepSeek Math V2, FLUX 2, GPT 5.1, Kimi K2 Thinking, Nano Banana Pro, GLM 4.7, Gemini 3, Mistral 3, MiniMax M2.1 🤯
Credits
🙏 NHLOCAL for the source data https://github.com/NHLOCAL/AiTimeline
🫡 @reach-vb for the original idea, design and recipe
🙌 @ariG23498 and yours truly for compiling and verifying the 2025 edition
🥳 Here's to 2026, wishing it becomes the best year ever for open releases and on-device-first use-cases! 🥂
We prepared the 2025 version of the HF AI Timeline Grid, highlighting open vs API-based model releases, and allowing you to browse and filter by access, modality, and release type!
Play with it here:
2025-ai-timeline/2025-ai-timeline
Here's my personal quarterly TL;DR:
1️⃣ Q1 — Learning to Reason
Deepseek not only releases a top-notch reasoning model, but shows how to train them and compete with closed frontier models. OpenAI debuts Deep Research.
Significant milestones: DeepSeek R1 & R1-Zero, Qwen 2.5 VL, OpenAI Deep Research, Gemini 2.5 Pro (experimental)
2️⃣ Q2 — Multimodality and Coding
More LLMs embrace multimodality by default, and there's a surge in coding agents. Strong vision, audio, and generative models emerge.
Significant milestones: Llama 4, Qwen 3, Imagen 4, OpenAI Codex, Google Jules, Claude 4
3️⃣ Q3 — "Gold" rush, OpenAI opens up, the community goes bananas
Flagship models get gold in Math olympiads and hard benchmarks. OpenAI releases strong open source models and Google releases the much anticipated nano-banana for image generation and editing. Agentic workflows become commonplace.
Significant milestones: Gemini and OpenAI IMO Gold, gpt-oss, Gemini 2.5 Flash Image, Grok 4, Claude Sonnet 4.5
4️⃣ Q4 — Mistral returns, leaderboard hill-climbing
Mistral is back with updated model families. All labs release impressive models to wrap up the year!
Significant milestones: Claude Opus 4.5, DeepSeek Math V2, FLUX 2, GPT 5.1, Kimi K2 Thinking, Nano Banana Pro, GLM 4.7, Gemini 3, Mistral 3, MiniMax M2.1 🤯
Credits
🙏 NHLOCAL for the source data https://github.com/NHLOCAL/AiTimeline
🫡 @reach-vb for the original idea, design and recipe
🙌 @ariG23498 and yours truly for compiling and verifying the 2025 edition
🥳 Here's to 2026, wishing it becomes the best year ever for open releases and on-device-first use-cases! 🥂
kenobi
authored 3
papers 3 months ago
On Invariance Penalties for Risk Minimization
Paper • 2106.09777 • Published
Generating (Factual?) Narrative Summaries of RCTs: Experiments with Neural Multi-Document Summarization
Paper • 2008.11293 • Published
Bayesian Deep Learning for Exoplanet Atmospheric Retrieval
Paper • 1811.03390 • Published
Post
4049
The new Mistral 3 models are here !
Today, we announce Mistral 3, the next generation of Mistral models. Mistral 3 includes three state-of-the-art small, dense models (14B, 8B, and 3B) and Mistral Large 3 – our most capable model to date – a sparse mixture-of-experts trained with 41B active and 675B total parameters.
All models are released under the Apache 2.0 license.
Ministrals :
https://huggingface.co/collections/mistralai/ministral-3
Mistral Large 3:
https://huggingface.co/collections/mistralai/mistral-large-3
Today, we announce Mistral 3, the next generation of Mistral models. Mistral 3 includes three state-of-the-art small, dense models (14B, 8B, and 3B) and Mistral Large 3 – our most capable model to date – a sparse mixture-of-experts trained with 41B active and 675B total parameters.
All models are released under the Apache 2.0 license.
Ministrals :
https://huggingface.co/collections/mistralai/ministral-3
Mistral Large 3:
https://huggingface.co/collections/mistralai/mistral-large-3
kumuji
authored a
paper 4 months ago
glopardo
authored 6
papers 5 months ago
Attention Meets Post-hoc Interpretability: A Mathematical Perspective
Paper • 2402.03485 • Published • 2
Faithful and Robust Local Interpretability for Textual Predictions
Paper • 2311.01605 • Published • 2
Understanding Post-hoc Explainers: The Case of Anchors
Paper • 2303.08806 • Published • 2
A Sea of Words: An In-Depth Analysis of Anchors for Text Data
Paper • 2205.13789 • Published • 2
Comparing Feature Importance and Rule Extraction for Interpretability on Text Data
Paper • 2207.01420 • Published • 2
SMACE: A New Method for the Interpretability of Composite Decision Systems
Paper • 2111.08749 • Published • 2
Post
1562
New blog post analyzing the top 50 entities with the most downloaded models on @huggingface 🤗!
https://huggingface.co/blog/lbourdois/huggingface-models-stats
The purpose here is to get an idea of the profile of the models with the greatest impact in open source (we are not interested in closed models here!).
32 figures + data
Enjoy 🤗
https://huggingface.co/blog/lbourdois/huggingface-models-stats
The purpose here is to get an idea of the profile of the models with the greatest impact in open source (we are not interested in closed models here!).
32 figures + data
Enjoy 🤗
gregmialz
authored a
paper 6 months ago
HugoLaurencon
authored a
paper 6 months ago