Frozen AI News archive

Halfmoon is Reve Image: a new SOTA Image Model from ex-Adobe/Stability trio

**Reve**, a new composite AI model from former Adobe and Stability alums **Christian Cantrell**, **Taesung Park**, and **Michaël Gharbi**, has emerged as the top-rated image generation model, surpassing previous state-of-the-art models like Recraft and Ideogram in text rendering and typography. The team emphasizes *"enhancing visual generative models with logic"* and *"understanding user intent with advanced language capabilities"* to iteratively amend visuals based on natural language input. Additionally, **DeepSeek-V3-0324** and **Alibaba's Qwen2.5-VL-32B-Instruct** models were released with notable performance improvements, including better vision task benchmarks and mathematical reasoning.

Canonical issue URL

AI News for 3/21/2025-3/24/2025. We checked 7 subreddits, 433 Twitters and 29 Discords (227 channels, and 10464 messages) for you. Estimated reading time saved (at 200wpm): 1129 minutes. You can now tag @smol_ai for AINews discussions!

A couple of nice updates from Qwen and Deepseek today, but we give title spot to a lesser known but ambitious new entrant.

Reve, pronounced [ʀɛv], from “rêve”, has emerged from Artificial Analysis' leaderboard as the top rated imagegen model, displacing former SOTA Recraft. "The model stands out for its impressive text rendering, prompt adherence, and aesthetics." We found it remarkably easy to play with.

image.png

image.png

And it beats Ideogram for typography:

image.png

It's interesting that it comes from Christian Cantrell, former VP Product at Stability, Taesung Park, and Michaël Gharbi. All are Adobe alums, and Michael's announcement gives the most insight into how they do it:

Reve’s mission is to invent the future of intent-driven visual creation. Capturing creative intent requires advanced machine understanding of natural language and other interactions. Turning this intent into compelling visual calls for interactive systems that have a deep understanding of the visual world they generate, so they can iteratively amend it.

Taesung agrees:

Today's text-to-image models are essentially that—random slice-of-the-world generator. There's no intelligence. This is both a data and representation problem. We need to leverage the equivalent of full documents for images, but we don't have a good representation for it. Our mission at Reve is to enhance visual generative models with logic. As the first step, we focus on understanding user intent with advanced language capabilities, resulting in superior complex prompt understanding and text writing.

There's no suggestion that it's a single model, but rather some composite of models. Probably this is what Christian wanted to build at Stability, but couldn't.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

Here's a summary of the AI-related discussions from the provided tweets, categorized for a technical audience:

Model Releases and Updates, Including Performance

Frameworks and Tools

Agent Engineering and Applications

Economic and Strategic Implications

ARC-AGI Benchmark

Humor and Memes


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. DeepSeek V3-0324: Performance and Expectations vs R1

Theme 2. Meta's ParetoQ Explored: Promise of 2-bit Models

Theme 3. Expanding LLM Functionalities: From Text to Multimodal

Theme 4. TeapotLLM's Impact: Lightweight Q&A Models

Other AI Subreddit Recap

/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding

Theme 1. New Improved Memory Alpha in ChatGPT Enhances Interaction

Theme 2. Anthropic's Revenue Surge Matches OpenAI's 2023 Numbers

Theme 3. AI-Driven Bug Fixing Automation: A 27-Day Experiment

Theme 4. Advanced Claude Workflow Integration: MCP External Tools

Theme 5. Wan 2.1 Video Frame Feature Innovations in AI


AI Discord Recap

A summary of Summaries of Summaries by o1-preview-2024-09-12

Theme 1. DeepSeek V3's Surprise Launch Shakes AI Community

Theme 2. Qwen Models and Upcoming AI Innovations

Theme 3. Debates and Advances in LLM Reasoning Training

Theme 4. Agent Engineering and MCP Developments

Theme 5. NVIDIA's Nemotron-H Models and Hardware Advances


PART 1: High level Discord summaries

Perplexity AI Discord


Unsloth AI (Daniel Han) Discord


LMArena Discord


Cursor Community Discord


aider (Paul Gauthier) Discord


Nous Research AI Discord


OpenAI Discord


OpenRouter (Alex Atallah) Discord


LM Studio Discord


Yannick Kilcher Discord


GPU MODE Discord


Interconnects (Nathan Lambert) Discord


Latent Space Discord


Notebook LM Discord


Eleuther Discord


HuggingFace Discord


MCP (Glama) Discord


Nomic.ai (GPT4All) Discord


Modular (Mojo 🔥) Discord


LlamaIndex Discord


Cohere Discord


Torchtune Discord


DSPy Discord


tinygrad (George Hotz) Discord


LLM Agents (Berkeley MOOC) Discord


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Codeium (Windsurf) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Gorilla LLM (Berkeley Function Calling) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Perplexity AI ▷ #general (998 messages🔥🔥🔥):

o3 mini, Grok 3, Chinese AI, Gemini deep research, Complexity plugin

Links mentioned:


Perplexity AI ▷ #sharing (18 messages🔥):

Trump, SSA shutdown, Boeing fighter, sunbathe, bluesky debates


Perplexity AI ▷ #pplx-api (21 messages🔥):

Perplexity API in Windsurf, API Credit vs Pro Subscription, Deep Research Limit, Sonar Model Truncated Responses, RAG Project with Sonar and Llama Index


Unsloth AI (Daniel Han) ▷ #general (602 messages🔥🔥🔥):

Bonsai bitnet, Mistral Small 3.1, Orpheus TTS, Gemma 3 27B, Llama 3 performance

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (41 messages🔥):

Unsloth PR process, Fine-tuning Arabic LLMs, Consensus framework for LLMs, Rotary Position Embedding (RoPE), Unsloth fork vs original repo

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (257 messages🔥🔥):

Training specific parts of output, GRPO notebooks, Dependency issue Qwen model, CUDA Version, Mistral 3.1

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (7 messages):

Unsloth fine-tuning, Lc0 Chess LLM, Vibe coding


Unsloth AI (Daniel Han) ▷ #research (51 messages🔥):

Tree of Thoughts limitations, Graph of Thought improvements, GRPO multi-turn setup, LLMs vs human brain, Llama3 Thai language support

Links mentioned:


LMArena ▷ #general (844 messages🔥🔥🔥):

Mistral Naming Schemes, Phantom Chatbot, Nebula Chatbot, DeepMind's Nebula, OpenAI GPT-4o

Links mentioned:


LMArena ▷ #announcements (1 messages):

Alpha Testing Updates, Bug Fixes, O3-Mini Formatting, Leaderboard Improvements

Links mentioned:


Cursor Community ▷ #general (857 messages🔥🔥🔥):

Cursor's Cmd+Backspace issue, Claude 3.7 Thinking pricing and features, windsurf better, MCP Combinations, AI's Limited Understanding of 3D Designs

Links mentioned:


aider (Paul Gauthier) ▷ #general (585 messages🔥🔥🔥):

Firecrawl, o1 vs o3 mini debugging, Claude Think Tool, Aider Homepage, Qwen 2.5 release

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (148 messages🔥🔥):

Anthropic API, Aider development workflow, Claude 3.7, Svelte 5 + SvelteKit, MCP servers in Claude App

Links mentioned:


aider (Paul Gauthier) ▷ #links (2 messages):

Aider Conventions, Prompts, LLM Documentation Snippets, Maybe Codebase Cursor Rules, Project Management Guidelines

Link mentioned: ctxs.ai context registry: An open-source, community-curated registry of contexts for use with LLMs


Nous Research AI ▷ #general (436 messages🔥🔥🔥):

LCPP Context Length, Quantization and Performance, Chinese Thinking Models, Agentic Workflows, Deepseek V3

Links mentioned:


Nous Research AI ▷ #ask-about-llms (46 messages🔥):

Steering Thinking Models, Deepseek V3 vs Sonnet 3.7, Fine-tuning LLMs on Codebases, Transformers without Normalization, Raytracing with LLMs

Link mentioned: llmbenchmark/raytracer at master · cpldcpu/llmbenchmark: Various LLM Benchmarks. Contribute to cpldcpu/llmbenchmark development by creating an account on GitHub.


Nous Research AI ▷ #research-papers (19 messages🔥):

Hunyuan-T1 Model, R1-Zero-Like Training, MathFusion for LLMs, GRPO on Coding Benchmarks, Satya Nadella on AGI

Links mentioned:


Nous Research AI ▷ #interesting-links (3 messages):

Qwen3, CPU inference

Link mentioned: Adding Qwen3 and Qwen3MoE by bozheng-hit · Pull Request #36878 · huggingface/transformers: Adding Qwen3This PR adds the support of codes for the coming Qwen3 models. For information about Qwen, please visit https://github.com/QwenLM/Qwen2.5. @ArthurZucker


Nous Research AI ▷ #research-papers (19 messages🔥):

Hunyuan-T1 Model, R1-Zero-Like Training, MathFusion Framework, GRPO on Coding Benchmarks, Open-R1 Project by Hugging Face

Links mentioned:


OpenAI ▷ #ai-discussions (226 messages🔥🔥):

GPT-4 Transcriber, Voicebot Tools, Turnitin AI Similarity, GPT-5 Release, Free Chatbots for Story Generation


OpenAI ▷ #gpt-4-discussions (2 messages):

GPT-4o mini TTS, Custom instructions


OpenAI ▷ #prompt-engineering (122 messages🔥🔥):

GPT-4o is a perfect model, NPCs in a customer service voice, AI Identity, UPSUM Chain Prompt, coherent multi-context conversation with an emergent persona


OpenAI ▷ #api-discussions (122 messages🔥🔥):

GPT-4o, AI NPCs, AI Identity Preservation Template, UPSUM Chain Prompt, Many-shot Prompting


OpenAI ▷ #api-projects (1 messages):

FormulaGPT, AI Racing Simulator, Open Source AI Racing

Link mentioned: GitHub - dawid-maj/FormulaGPT: FormulaGPT – AI-powered Formula 1 race simulator with real-time team management and strategy decisions.: FormulaGPT – AI-powered Formula 1 race simulator with real-time team management and strategy decisions. - dawid-maj/FormulaGPT


OpenRouter (Alex Atallah) ▷ #announcements (4 messages):

OpenAI o1-pro, Markdown Export, DeepSeek V3, Anthropic Outage

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (440 messages🔥🔥🔥):

OpenAI o1-pro API Pricing, Gemini's Image Generation, Lambda Endpoint Issues, DeepSeek R1 Model

Links mentioned:


LM Studio ▷ #general (199 messages🔥🔥):

NPU support, KV cache 8-bit quants, LM Studio runtimes, GPUs, Gemma 3 1B

Links mentioned:


LM Studio ▷ #hardware-discussion (159 messages🔥🔥):

VRAM Usage, Google Coral dual TPU, RX 6800 ROCm support, RTX 4060-Ti vs RX 7800 XT, AI APUs

Links mentioned:


Yannick Kilcher ▷ #general (326 messages🔥🔥):

VPN Injection, Amodal3R, NVIDIA cuOpt, CUDA Python, Mixture of Experts (MoEs)

Links mentioned:


Yannick Kilcher ▷ #paper-discussion (3 messages):

DeepSeek-V3, DeepSeek-R1, Multi-Head Latent Attention (MLA)

Link mentioned: 🥇Top AI Papers of the Week: The Top AI Papers of the Week (Mar 17 - 23)


Yannick Kilcher ▷ #ml-news (17 messages🔥):

ChatGPT & Loneliness, AITER Tensor Engine for ROCm, DeepSeek-V3-0324, Pokemon Red DRL

Links mentioned:


GPU MODE ▷ #general (22 messages🔥):

Cloud Providers with Profilers, In-depth Dive into NCCL, Quantization Benchmarking, Understanding Flash Attention, ILGPU 2.0 Availability

Links mentioned:


GPU MODE ▷ #triton (15 messages🔥):

Triton and Pip Confusion, cuTIl Performance, BF16 Atomic Operations, Triton IR Generation, Flash Attention 1 Kernel Issues

Links mentioned:


GPU MODE ▷ #cuda (42 messages🔥):

WMMA instructions, PyTorch RTX 5080 CUDA 12.8 Support, Flash Attention Optimization, Hopper Architecture Swizzle, CUDA Performance Counters Permission Error

Links mentioned:


GPU MODE ▷ #torch (5 messages):

torch.compile() graph breaks, VRAM reduction techniques, FA3 attention FP8

Links mentioned:


GPU MODE ▷ #announcements (1 messages):

Tanishq Kumar, Scaling Laws for Low Precision, Precision-aware scaling laws, post-training quantization, compute optimal

Link mentioned: Scaling Laws for Precision: Low precision training and inference affect both the quality and cost of language models, but current scaling laws do not account for this. In this work, we devise "precision-aware" scaling la...


GPU MODE ▷ #cool-links (1 messages):

srivarshan4271: https://lights0123.com/blog/2025/01/07/hip-script/


GPU MODE ▷ #jobs (1 messages):

AI & Neuroscience Fellowship at the University of Oxford, AI / RL in games and neuroimaging, non-invasive diagnosis and treatment of neurological disorders

Links mentioned:


GPU MODE ▷ #beginner (56 messages🔥🔥):

GPU/CUDA learning resources, Warp scheduler significance, Context switching, SIMD vs SIMT execution, Flash attention setup on Windows

Link mentioned: GPU Glossary: A glossary of terms related to GPUs.


GPU MODE ▷ #pmpp-book (1 messages):

Amazon Book Release Date, 5th Edition of Book


GPU MODE ▷ #jax (1 messages):

bigfoot1144: Any progress so far?


GPU MODE ▷ #rocm (2 messages):

ROCm, tilelang HIP backend, row-row bank conflict-free swizzle, AMD sponsoring cards


GPU MODE ▷ #lecture-qa (2 messages):

Hopper Flops, H100 Clock Rate, H100 SMs, Nvidia Boost Clocks

Link mentioned: NVIDIA H100 Tensor Core GPU Architecture Overview: A high-level overview of NVIDIA H100, new H100-based DGX, DGX SuperPOD, and HGX systems, and a H100-based Converged Accelerator. This is followed by a deep dive into the H100 hardware architecture, ef...


GPU MODE ▷ #tilelang (10 messages🔥):

Tilelang 2:4 sparsity support, Tilelang v0.1.3 Release, SPGEMM issue

Link mentioned: Release v0.1.3 · tile-ai/tilelang: What's Changed[Docker] Add libstdcxx-ng-12 to Dockerfiles for CUDA versions by @LeiWang1999 in #160Add cpu jit with backend ctypes by @xs-keju in #154[Carver] Multi-Threads Compilation for Fast...


GPU MODE ▷ #metal (3 messages):

Parallelized Cholesky, Python + MLX + Metal


GPU MODE ▷ #self-promotion (10 messages🔥):

WheelNext Initiative, CUDA Indexing Blogpost, Container-First Triton Development, GemLite bfloat16 Support

Links mentioned:


GPU MODE ▷ #🍿 (1 messages):

LLM Kernel Understanding, RL for Operation Understanding, Reducing Hallucinations in Kernel Creation


GPU MODE ▷ #reasoning-gym (5 messages):

veRL rollouts with sglang, low precision data types, quantization strategies for RL, ARC-AGI2 announcement

Links mentioned:


GPU MODE ▷ #gpu模式 (5 messages):

CUDA core, CUDA_fp6.hpp, CUDA_fp4.hpp


GPU MODE ▷ #general (9 messages🔥):

Submission Guide, Kernel profiling, Conv2D error

Link mentioned: Getting Started | GPU MODE Kernel Leaderboard: Welcome! If you are excited about building GPU kernels, this leaderboard is the place for you! We


GPU MODE ▷ #submissions (119 messages🔥🔥):

matmul benchmarks on H100, grayscale benchmarks on A100, grayscale benchmarks on T4, L4, A100, H100, histogram benchmarks on T4, vectorsum tests on A100


GPU MODE ▷ #status (2 messages):

CUDA, load_inline(), PyTorch headers, KernelBot

Link mentioned: load_inline no_implicit_headers mode by msaroufim · Pull Request #149480 · pytorch/pytorch: In the kernelBot leaderboard we support people competing with custom cuda extensions via load_inline(), however even on toy kernels this can result in cold starts of up to 90s - this problem is pri...


GPU MODE ▷ #hardware (17 messages🔥):

GPU prices, VRAM requirements for LLMs, RTX Pro 6000, CUDA Capability


GPU MODE ▷ #tpu (1 messages):

rocka2424: This is awesome, looking forward to it!


Interconnects (Nathan Lambert) ▷ #news (86 messages🔥🔥):

Nvidia Mamba-Transformer Hybrid, Qwen 2.5 Omni Model, DeepSeek V3 Model Update, Reve Image Halfmoon Model, Qwen2.5-VL-32B-Instruct

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (25 messages🔥):

Impact of noisy data in multi-turn SFT, Transformer usage in RL, Community model preferences, Trusting eval benchmarks, Gemini's image generation

Link mentioned: Building on evaluation quicksand: On the state of evaluation for language models.


Interconnects (Nathan Lambert) ▷ #random (36 messages🔥):

LLM input/output tokens, o1-pro performance, Mistral 24B is impressive, Claude Compass starter prompts, DAPO and Dr. GRPO

Links mentioned:


Interconnects (Nathan Lambert) ▷ #memes (4 messages):

O1-pro vs BoN, O1-pro reasoning paths marginalization, Tech CEOs in Open Source RL

Links mentioned:


Interconnects (Nathan Lambert) ▷ #rl (127 messages🔥🔥):

R1-Zero Training, GRPO Bias, LOOP & RLOO, PPO Objective, Creative Writing LLMs

Links mentioned:


Interconnects (Nathan Lambert) ▷ #cv (6 messages):

Operator agent limitations, Infinibranch Browsers as a solution, Intelligent Browser Automation

Links mentioned:


Interconnects (Nathan Lambert) ▷ #reads (16 messages🔥):

R1-Zero-Like Training, DeepSeek-V3-Base, GRPO Bias in RL-tuning, CoT Philosophy, Math errors in AI papers

Link mentioned: Tweet from Zichen Liu (@zzlccc): 🪂Understanding R1-Zero-Like Training: A Critical Perspective* DeepSeek-V3-Base already exhibits "Aha moment" before RL-tuning??* The ever-increasing output length in RL-tuning might be due to...


Interconnects (Nathan Lambert) ▷ #lectures-and-projects (2 messages):

Claude PR, Header Copy Links

Link mentioned: (experimental) Add heading anchor links for easy section linking by natolambert · Pull Request #82 · natolambert/rlhf-book: Add copyable links to all headings that appear on hoverLinks copy the current URL with fragment identifier to clipboardAdd CSS for styling the anchor linksUpdate Makefile to copy new JS file to ...


Interconnects (Nathan Lambert) ▷ #policy (9 messages🔥):

China's Open Source AI Blitz, DeepSeek's Impact, US vs China AI Competition, Chinese consumer market for software, China commoditizing hardware

Link mentioned: Tweet from Balaji (@balajis): AI OVERPRODUCTIONChina seeks to commoditize their complements. So, over the following months, I expect a complete blitz of Chinese open-source AI models for everything from computer vision to robotics...


Interconnects (Nathan Lambert) ▷ #expensive-queries (17 messages🔥):

Grok DeeperSearch, OpenAI Deep Research, Twitter Premium, HF model comparisons

Link mentioned: Tweet from Tibor Blaho (@btibor91): @TheXeophon-bench


Latent Space ▷ #ai-general-chat (89 messages🔥🔥):

Gemini Updates, Claude Code New Features, Model Context Protocol (MCP), AI Agents and Email, RF-DETR Object Detection Model

Links mentioned:


Latent Space ▷ #ai-announcements (2 messages):

Rishi Agarwal on Distillation, Swyx's Agent Engineering Talk, Agent Engineering Elements, Agents as ChatGPT's Growth Path

Link mentioned: Tweet from swyx 🌉 (@swyx): 🆕 talk + essay: Agent Engineeringhttps://latent.space/p/agentWhy we went all in on Agents @aiDotEngineerDefining Agents (thanks to @simonw)The Six Elements of Agent EngineeringWhy Agents are ChatGPT&...


Latent Space ▷ #ai-in-action-club (226 messages🔥🔥):

DORA report, Gemini API, AI code generation, Agile adoption, Ruby on Rails

Links mentioned:


Notebook LM ▷ #announcements (1 messages):

Mobile Study Participants, AI Model Updates


Notebook LM ▷ #use-cases (52 messages🔥):

Mindmaps in NotebookLM, Research with NotebookLM, HR policies Hub in NotebookLM, NotebookLM for literature search, External Users Share NotebookLM

Links mentioned:


Notebook LM ▷ #general (202 messages🔥🔥):

Mind Map Pixelation Fix, Mind Map Feature Feedback, NotebookLM vs ChatGPT, Access to New NotebookLM, Feedback Methods for NotebookLM

Links mentioned:


Eleuther ▷ #general (106 messages🔥🔥):

RWKV architecture development, AI model viability prediction, EleutherAI evaluation methods, Low precision data types for RL, MkDocs site for lm-evaluation-harness

Links mentioned:


Eleuther ▷ #research (121 messages🔥🔥):

AI simulation environments, Continual learning in production LLMs, Architecture-aware optimizers, Sharpness Disparity across Transformer blocks, VectorAdam optimizer

Links mentioned:


Eleuther ▷ #interpretability-general (20 messages🔥):

mechinterp backlash, token level activations, SAE visualizations, single token activations, untied embeddings


Eleuther ▷ #multimodal-general (1 messages):

Recursive Design, GAN vs. CNN vs. RL Architectures


Eleuther ▷ #gpt-neox-dev (1 messages):

lm_eval update, CI test failures

Links mentioned:


HuggingFace ▷ #announcements (1 messages):

StarVector, SpatialLM, Hugging Face Agents Course, Xet on the Hub, HF Welcome Page Makeover

Links mentioned:


HuggingFace ▷ #general (136 messages🔥🔥):

ComfyUI Samplers, Open Schizo Leaderboard, Short Story Generator with Pytorch, Photorealism Settings for SD1.5/SDXL, Flux.1 Model Performance

Links mentioned:

"This authentication method does…": no description foundWhisper — vLLM: no description foundWindchill – Kattis, AI-olympiadens Kval 2025: no description found@julien-c on Hugging Face: "Important notice 🚨

For Inference Providers who have built support for our…": no description foundPricing and Rate limits: no description foundWan-AI/Wan2.1-I2V-14B-480P-Diffusers · Hugging Face: no description foundOpen R1: Update #3: no description foundcannot import name 'AutoencoderKLWan' from 'diffusers' · Issue #10963 · huggingface/diffusers: Describe the bug ImportError: cannot import name 'AutoencoderKLWan' from 'diffusers' (/usr/local/lib/python3.10/dist-packages/diffusers/init.py) Reproduction from diffusers import Auto...Inference Endpoints: no description foundInference Endpoints (dedicated) - Hugging Face Open-Source AI Cookbook: no description foundGitHub - huggingface/text-generation-inference: Large Language Model Text Generation Inference: Large Language Model Text Generation Inference. Contribute to huggingface/text-generation-inference development by creating an account on GitHub.Expert Support – Hugging Face: no description foundDiffusers Image Outpaint - a Hugging Face Space by fffiloni: no description foundImage-to-image: no description foundGitHub - justinpinkney/stable-diffusion: Contribute to justinpinkney/stable-diffusion development by creating an account on GitHub.GitHub - TheDenk/images_mixing: Сombine images using usual diffusion models.: Сombine images using usual diffusion models. Contribute to TheDenk/images_mixing development by creating an account on GitHub.Spaces - Hugging Face: no description foundSpaces - Hugging Face: no description foundFine-Tuning Open-Source Language Models: A Step-by-Step Guide | by Vi…: no description foundQwen/Qwen2.5-VL-7B-Instruct · Hugging Face: no description foundmicrosoft/Phi-4-multimodal-instruct · Hugging Face: no description foundAdvanced RAG on Hugging Face documentation using LangChain - Hugging Face Open-Source AI Cookbook: no description foundAutomatic speech recognition: no description found


HuggingFace ▷ #today-im-learning (5 messages):

audio processing, AI agents, Tokenisers, BPE, Unigram language modelling


HuggingFace ▷ #i-made-this (8 messages🔥):

Logfire Callback for HF Transformers Trainer, TrashLens for image organization, pdf2notes: AI-powered PDF to Notes conversion, Kids feedback on UI/UX, Local API Usage

Links mentioned:


HuggingFace ▷ #computer-vision (6 messages):

Qwen for video annotation, Opus clip opensource, LLMs and VLMs in autonomous driving

Link mentioned: Autonomous driving with LLMs, VLMs, and MLLMs: Discussing the application of Large Language/Vision Models in autonomous driving and the most significant developments and approaches.


HuggingFace ▷ #gradio-announcements (1 messages):

Gradio Deep Links

Link mentioned: black-forest-labs/FLUX.1-schnell: no description found


HuggingFace ▷ #smol-course (1 messages):

Hackathon Timing, Hackathon Details


HuggingFace ▷ #agents-course (33 messages🔥):

LangGraph rigidity, Local LLMs for smolagents, Gemini in LangGraph, API costs for notebooks, Agent storing retrieved info

Links mentioned:


HuggingFace ▷ #open-r1 (9 messages🔥):

r1, vllm, cuda kernel


MCP (Glama) ▷ #general (155 messages🔥🔥):

MCP and K8s, Anthropic's MCP, MCP server directories, C# MCP SDK, Vercel's AI SDK with MCP Clients

Links mentioned:


MCP (Glama) ▷ #showcase (29 messages🔥):

mcpwizard, vscode-mcp, DICOM servers MCP, google sheet MCP server, Narrative Spittoon Inversion project

Links mentioned:


Nomic.ai (GPT4All) ▷ #general (102 messages🔥🔥):

Speech to Text Solutions, GPT4All and NSFW content, LocalDocs Disappearing, LLMs for Office Tasks, Running Models on Multiple Devices

Links mentioned:


Modular (Mojo 🔥) ▷ #general (7 messages):

High performance software, Vendor lock-ins, OpenCL, OpenMP, OpenACC, Vulkan’s Compute API, and SYCL, Democratizing AI Compute, Hardware Lottery

Links mentioned:


Modular (Mojo 🔥) ▷ #mojo (82 messages🔥🔥):

Mojo Logging Library, Mojo Formatter Tool, Mojo Dict Default Values, GPU Support for Windows, Mojo Inline Assembly Documentation

Link mentioned: Question: vpermi2b inline assembly output incorrect in loop context due to register allocation: Maybe you could try this from sys import llvm_intrinsic alias T = SIMD[DType.int8, 64] @always_inline("nodebug") fn vpermi2b(a: T, b: T, idx: T) -> T: return llvm_intrinsic["llv...


Modular (Mojo 🔥) ▷ #max (3 messages):

MAX Platform, pixi.toml, max-pipeline, Python model graphs, magic CLI

Links mentioned:


LlamaIndex ▷ #blog (4 messages):

AGNTCY, Large-Scale Structured Extraction, Deepseek R1 + LlamaIndex RAG app, WeAreDevs WebDev & AI Day


LlamaIndex ▷ #general (71 messages🔥🔥):

Haystack Uninstall LlamaIndex Install, Ollama Integration Error, RTX 3060 Token Issues, Custom AI Interview Prep, Agent Workflow Timeout Error

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):

Multi-Agent Systems, Program-Wide Backoff Mechanism, Function Calling


Cohere ▷ #「💬」general (25 messages🔥):

RAG source return, data retention policy, security information about chat with cohere, sampler settings for Command A, AI assistant powered by Cohere's command-r-plus

Links mentioned:


Cohere ▷ #「🔌」api-discussions (35 messages🔥):

Command models, SSL Errors, API Rate Limits, MongoDB


Cohere ▷ #「💡」projects (2 messages):

Discord Bot, RAG Pipeline, vnc-lm, Context Augmentation, Docker

Links mentioned:


Torchtune ▷ #general (33 messages🔥):

Synthetic Data Generation with vllm and deepseek r1, Llama4 Release, Qwen3 MoE, Good Data Problem, PDF Extraction

Links mentioned:


Torchtune ▷ #dev (23 messages🔥):

datasets library issue, GRPO LoRA 3B Single Device, vLLM support for data generation, CUDA graphs

Link mentioned: GRPO LoRA Single Device by ianbarber · Pull Request #2467 · pytorch/torchtune: ContextWhat is the purpose of this PR? Is it to[x ] add a new feature fix a bug update tests and/or documentation other (please add here)#2421 - exploring a LoRA recipe.ChangelogWhat are ...


DSPy ▷ #show-and-tell (1 messages):

DLCoT Optimizer, Chain-of-Thought Distillation, Token Usage Reduction, DSPy Optimizers

Link mentioned: Add DLCoT Optimizer for efficient Chain-of-Thought distillation by jmanhype · Pull Request #8000 · stanfordnlp/dspy: Add DLCoT (Deconstructing Long Chain-of-Thought) OptimizerOverviewThis PR adds a new optimizer to the DSPy teleprompt module: the DLCoT (Deconstructing Long Chain-of-Thought) optimizer. This feat...


DSPy ▷ #general (20 messages🔥):

DSPy for creative content generation, PAPILLON example, Agentic-Reward-Modeling link, DLCoT Optimizer, MIPROv2

Links mentioned:


DSPy ▷ #examples (9 messages🔥):

DSPy Modules, Creative Writing Prompts, PAPILLON, Privacy Preservation

Link mentioned: PAPILLON/papillon_tutorial.ipynb at main · Columbia-NLP-Lab/PAPILLON: Code for our paper PAPILLON: PrivAcy Preservation from Internet-based and Local Language MOdel ENsembles - Columbia-NLP-Lab/PAPILLON


tinygrad (George Hotz) ▷ #general (19 messages🔥):

sops.gz dataset, Tinygrad CUDA port, Meeting #63 Agenda, AMD LLVM progress, ONNX Frontend for Tinygrad

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (4 messages):

Disable colored terminal output, tinygrad facades, GPU code generation, OpenCLEmpty guarantees

Link mentioned: Introduction to the internals: Tutorials on tinygrad


LLM Agents (Berkeley MOOC) ▷ #mooc-questions (9 messages🔥):

Quiz Typos, AgentX Research Track, Remote Research Mentorship, Unpaid Research





{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}