Frozen AI News archive

not much happened today

**Olmo 2** released a detailed tech report showcasing full pre, mid, and post-training details for a frontier fully open model. **PRIME**, an open-source reasoning solution, achieved **26.7% pass@1**, surpassing **GPT-4o** in benchmarks. Performance improvements include **Qwen 32B (4-bit)** generating at **>40 tokens/sec** on an **M4 Max** and **libvips** being **25x faster** than **Pillow** for image resizing. New tools like **Swaggo/swag** for Swagger 2.0 documentation, **Jujutsu (jj)** Git-compatible VCS, and **Portspoof** security tool were introduced. Robotics advances include a weapon detection system with a meters-wide field of view and faster frame rates. Hardware benchmarks compared **H100** and **MI300x** accelerators. Applications span medical error detection using PRIME and a financial AI agent integrating **LangChainAI** and **Vercel AI SDK**. Architectural insights suggest the need for breakthroughs similar to **SSMs** or **RNNs**.

Canonical issue URL

AI News for 1/2/2025-1/3/2025. We checked 7 subreddits, 433 Twitters and 32 Discords (217 channels, and 2120 messages) for you. Estimated reading time saved (at 200wpm): 236 minutes. You can now tag @smol_ai for AINews discussions!

Lots of "open o1" imitators causing noise, but mostly not leaving much confidence and meanwhile o1 is continuing to impress. Olmo 2 released their tech report (our first coverage here), with characteristic full {pre|mid|post}-training detail for one of the few remaining frontier fully open models.

image.png


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

AI Models and Performance

AI Tools and Frameworks

Robotics and Hardware

AI Applications and Use Cases

Industry Updates and News

Community and Personal Reflections

Memes and Humor


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. LLM Performance Leap Creates Demand for New Benchmarks

Theme 2. Deepseek V3 Hosted on Fireworks, Privacy and Pricing

Theme 3. Tsinghua's Eurus-2: Novel RL Methods Beat Qwen2.5

Theme 4. OLMo 2.0: Competitive Open Source Model Released

Other AI Subreddit Recap

/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT

Theme 1. Video Generation Tool Comparison: Sora vs Veo2 vs Minimax

Theme 2. GPT-4o: Advanced Reasoning Over GPT-3.5


AI Discord Recap

A summary of Summaries of Summaries by o1-2024-12-17

Theme 1. Performance Sagas and Slowdowns

Theme 2. Credit Crunch and Cost Confusion

Theme 3. Model Debuts and Fine-Tuning Frenzy

Theme 4. Tooling Triumphs and Tensions

Theme 5. Hardware, VRAM, and HPC Adventures


PART 1: High level Discord summaries

Codeium (Windsurf) Discord


aider (Paul Gauthier) Discord


Unsloth AI (Daniel Han) Discord


Cursor IDE Discord


Nous Research AI Discord


LM Studio Discord


Notebook LM Discord Discord


Stackblitz (Bolt.new) Discord


Perplexity AI Discord


OpenRouter (Alex Atallah) Discord


Stability.ai (Stable Diffusion) Discord


Eleuther Discord


Latent Space Discord


GPU MODE Discord


Interconnects (Nathan Lambert) Discord


OpenAI Discord


Cohere Discord


Torchtune Discord


LlamaIndex Discord


OpenInterpreter Discord


Modular (Mojo 🔥) Discord


LLM Agents (Berkeley MOOC) Discord


DSPy Discord


tinygrad (George Hotz) Discord


MLOps @Chipro Discord


Nomic.ai (GPT4All) Discord


LAION Discord


The Axolotl AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The HuggingFace Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Gorilla LLM (Berkeley Function Calling) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Codeium (Windsurf) ▷ #discussion (188 messages🔥🔥):

Windsurf Performance Issues, Cascade Credits Consumption, Codeium Plugin Suggestions, User Support Experiences, Learning and Coding Tools

Links mentioned:


Codeium (Windsurf) ▷ #windsurf (194 messages🔥🔥):

Performance Issues with Windsurf, DeepSeek v3 vs. Sonnet 3.6, Code Editing Errors, Prompt and Configuration Management, Credit System Feedback

Links mentioned:


aider (Paul Gauthier) ▷ #general (198 messages🔥🔥):

Linting Confusion, Deepseek Performance, Architect Mode in Aider, New AI Model Announcements, Using OpenRouter with Aider

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (47 messages🔥):

OpenRouter API issues, Aider configuration and context management, Tailwind CSS documentation integration, Graphrag tool for RAG, Feature requests for Aider

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (188 messages🔥🔥):

OpenWebUI dataset export, Inference methods in Unsloth, Model quantization issues, VLLM for LLM inference, Fine-tuning choices for text classification

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (16 messages🔥):

Unsloth library installation issues, Granite training errors, Fine-tuning models for specific tasks, Using Colab with custom datasets, Understanding embedding vs fine-tuning

Links mentioned:


Unsloth AI (Daniel Han) ▷ #research (8 messages🔥):

Process Reinforcement through Implicit Rewards, O1 Reproduction Efforts, Fudan Report, Reinforcement Learning Code


Cursor IDE ▷ #general (170 messages🔥🔥):

DeepSeek V3 Model Discussion, Email Spoofing Incident, Using Cursor for Project Development, Website Design Inspiration, Language Selection for Marketplace Development

Links mentioned:


Nous Research AI ▷ #general (139 messages🔥🔥):

Swarm Library Release, Research Proposal Assistance, NLP Workshops at ICML, Fine-Tuning Models, Training Costs for AI Models

Links mentioned:


Nous Research AI ▷ #ask-about-llms (24 messages🔥):

Hermes Training Data, VLM Use Cases, Fine-Tuning Models, Model Weight Distribution, Mergoo Introduction


LM Studio ▷ #general (119 messages🔥🔥):

Extractor.io Website, LM Studio Image Generation, Qwen2-VL Model Limitations, AI Training with Internet Data, Model Performance Issues

Links mentioned:


LM Studio ▷ #hardware-discussion (38 messages🔥):

Local LLM Usage, API Concerns, GPU vs CPU Utilization, Quest Generation, Hardware Recommendations

Link mentioned: Download and run NousResearch/Hermes-3-Llama-3.1-8B-GGUF in LM Studio: Use NousResearch/Hermes-3-Llama-3.1-8B-GGUF locally in your LM Studio


Notebook LM Discord ▷ #use-cases (15 messages🔥):

Various media references, Conflict of interest in studies, Translation checks, Long-term mycelium storage techniques, Turing Test discussions

Links mentioned:


Notebook LM Discord ▷ #general (101 messages🔥🔥):

Sharing Notebooks Issues, Beta Testing Experience, Multilingual Features, AI Summarization Product, Customization Function Use

Links mentioned:


Stackblitz (Bolt.new) ▷ #prompting (3 messages):

Bolt code handling issues, UI implementation challenges, Web app development tips


Stackblitz (Bolt.new) ▷ #discussions (110 messages🔥🔥):

Billing issues with Bolt, Debugging in Bolt, Integrating APIs with Bolt, Building frontend with mock data, Using Supabase with Bolt

Links mentioned:


Perplexity AI ▷ #general (103 messages🔥🔥):

Perplexity O1 Feature, ChatGPT vs. Perplexity, Grok Model Opinions, UI Changes in Perplexity, Using AI Subscriptions

Links mentioned:


Perplexity AI ▷ #sharing (7 messages):

Musk Lawsuit Support, AI Interview Preparation, Oceans and CO2 Absorption, 2025 Predictions, Perplexity AI Origins

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (1 messages):

API server location, Chatbot integration, Token utilization


OpenRouter (Alex Atallah) ▷ #general (86 messages🔥🔥):

OpenRouter authentication issues, DeepSeek performance, Model recommendations for structured output, Janitor AI integration with OpenRouter, Payment processing issues

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (67 messages🔥🔥):

SwarmUI vs Forge, SANA and Omnigen Models, Video Generation Models, Text Output in Images, GPU Recommendations for AI

Link mentioned: GitHub - butaixianran/Stable-Diffusion-Webui-Civitai-Helper: Stable Diffusion Webui Extension for Civitai, to manage your model much more easily.: Stable Diffusion Webui Extension for Civitai, to manage your model much more easily. - butaixianran/Stable-Diffusion-Webui-Civitai-Helper


Eleuther ▷ #general (21 messages🔥):

RTX 50xx series VRAM limitations, Integration of SSDs with GPUs, Memory performance between VRAM and RAM, Cache hierarchy in GPUs


Eleuther ▷ #research (35 messages🔥):

Attention on Attention Weights, Quartic Transformer, Ring Attention, Higher-Order Attention, HYMBA and SWA

Links mentioned:


Latent Space ▷ #ai-general-chat (15 messages🔥):

LLM developments in 2024, Text extraction tools evaluation, Image generation trends, SmallThinker-3B model performance, OLMo2 tech report release

Links mentioned:


Latent Space ▷ #ai-announcements (4 messages):

AI Engineer Summit, AI Engineer World's Fair, Understanding Transformers

Links mentioned:


Latent Space ▷ #ai-in-action-club (31 messages🔥):

Discord Bot Building, Obsidian Tool Discussions, Screen Sharing Issues, Webhook Limitations


GPU MODE ▷ #general (2 messages):

GEMM performance on GPU, Identifying inefficiencies in GEMM, Optimizing GEMM computations, Suggestions for improvements in GEMM

Link mentioned: Tweet from YyWangCS (@YyWangCS17122): Matrix multiplication (GEMM) performance is crucial for deep learning. I’ve written an article on how to automatically detect computation inefficiencies in GEMM computations on GPU. https://yywangcs.n...


GPU MODE ▷ #triton (20 messages🔥):

Triton kernel performance, Matrix multiplication issues, Testing kernel equivalence, Pointer loading in Triton, Floating point operation ordering

Links mentioned:


GPU MODE ▷ #cuda (3 messages):

Dynamic br/bc values in Flash Attention, Fixed br/bc values performance, Compiling versions for selection


GPU MODE ▷ #torch (11 messages🔥):

Model Level Compile, Memory Profiling Issues, Inductor Cache Performance, Flex Attention Complications, Gradients and Activation Management

Link mentioned: TorchDynamo APIs for fine-grained tracing — PyTorch 2.5 documentation: no description found


GPU MODE ▷ #jobs (1 messages):

P-1 AI, Research Engineer role, Artificial General Engineering

Link mentioned: P-1 AI - Staff Research Engineer: no description found


GPU MODE ▷ #beginner (5 messages):

GPU Upgrade Considerations, Federated/Gossip Learning Resources, CUDA Learning, VRAM and Global Memory Importance, Upcoming Event on Turing Tensor Cores

Link mentioned: - YouTube: no description found


GPU MODE ▷ #off-topic (3 messages):

Learning Cuda, Mental Health Awareness, Felix Hill Tribute

Link mentioned: - YouTube: no description found


GPU MODE ▷ #self-promotion (2 messages):

TorchTitan, MFU vs HFU, Lecture 39 on YouTube

Link mentioned: - YouTube: no description found


GPU MODE ▷ #arc-agi-2 (1 messages):

Transduction Goals, Prompt Optimization, Training Procedure


Interconnects (Nathan Lambert) ▷ #ml-questions (14 messages🔥):

Fine-tuning LLMs with LoRA, HuggingFace TRL vs. alternatives, MoE routing techniques, Documentation availability

Link mentioned: LLM Fine-Tuning Library Comparison: Fine-Tuning LLMs with LoRA: A Comparative Analysis of Five Popular Libraries The rise of large language models (LLMs) has ushered in a new era in natural language processing, enabling remarkable advan...


Interconnects (Nathan Lambert) ▷ #random (18 messages🔥):

Post-training tutorial content, Quality of recorded talks, AI Safety Institute's activities, LinkedIn dynamics, Chatbotarena plot maintenance

Link mentioned: Tweet from Lauren Wagner (@typewriters): OH: I don't understand why people at the UK AI Safety Institute aren't considered intelligence operatives. They moved from London to SF, throw parties, get drunk researchers to talk, and send ...


Interconnects (Nathan Lambert) ▷ #reads (4 messages):

The Bittersweet Lesson, Felix's Contributions, Google Account Concerns

Link mentioned: The Bittersweet Lesson: The Bittersweet Lesson 😆 The strange case of inductive bias in Transformers Felix Hill 21 Oct 2024 Do you remember a few years back when the notion of inductive bias was central to machine learnin...


Interconnects (Nathan Lambert) ▷ #posts (6 messages):

SnailBot news, Entertainment value of SnailBot


OpenAI ▷ #ai-discussions (19 messages🔥):

FLUX.1 [dev], Image Filters on Minecraft, Community Dynamics, Using Discord Group, AGI Discussions

Link mentioned: black-forest-labs/FLUX.1-dev · Hugging Face: no description found


OpenAI ▷ #gpt-4-discussions (5 messages):

ChatGPT for search results, YouTube GPTs functionality, Cross posting etiquette


OpenAI ▷ #prompt-engineering (1 messages):

castilla99_87524: how do i keep consistent characters when making dif scenes in sora?


OpenAI ▷ #api-discussions (1 messages):

castilla99_87524: how do i keep consistent characters when making dif scenes in sora?


Cohere ▷ #discussions (7 messages):

New Year Wishes, Rerank on Azure


Cohere ▷ #rules (1 messages):

Server Guidelines, Promotion Rules, Spam Policy, Commercial Activities Restrictions


Cohere ▷ #questions (2 messages):

Command-R functionality, Issues with Command-R, Resuming processes


Cohere ▷ #cmd-r-bot (6 messages):

Increased Embedding Rate Limit, Cohere API Rate Limits


Torchtune ▷ #general (10 messages🔥):

Torchtune Benchmarking, Chunked Cross Entropy Implementation, Memory Gains during Compilation

Links mentioned:


Torchtune ▷ #dev (1 messages):

PyTorch Torchtune Bug, Flex Attention Compilation, Kernel Finding

Link mentioned: Issues · pytorch/torchtune: PyTorch native post-training library. Contribute to pytorch/torchtune development by creating an account on GitHub.


LlamaIndex ▷ #blog (1 messages):

Invoice processing agent, LlamaParse, Agentic workflow, Spend categories, Cost centers


LlamaIndex ▷ #general (9 messages🔥):

Dataset storage options, Query fusion for retrievers, JSON advantages, Compression techniques for data, Using SQL or NoSQL for datasets


OpenInterpreter ▷ #general (8 messages🔥):

Open Interpreter functionality, Open-source contributions, Installation Steps for Open Interpreter, Web WhatsApp messaging, Trading clicker execution mode


Modular (Mojo 🔥) ▷ #general (1 messages):

rd4com: 🥳 Happy new year !!


Modular (Mojo 🔥) ▷ #mojo (6 messages):

Linked List Implementation, Building CLI and TUI Tools, AST and Index-Style Trees, Mojo Debugging Issues

Link mentioned: [BUG] --debug-level full crashes when importing · Issue #3917 · modularml/mojo: Bug description Running a mojo script using the debugger seg faults, as opposed to when running regular mojo, which runs to completion (although I have noticed strange behavior in the regular scrip...


LLM Agents (Berkeley MOOC) ▷ #mooc-questions (5 messages):

Certificate Issuance, Fall 2024 Course Enrollment, Spring 2025 Course Sign-up

Link mentioned: Large Language Model Agents MOOC: MOOC, Spring 2025


DSPy ▷ #general (4 messages):

GraphRAG implementation, Donor's Game simulation, DSPy strategy updates

Link mentioned: cultural_evolution/donors_game/game/orchestrator.py at main · CakeCrusher/cultural_evolution: implements the methodology outlined in the paper Cultural Evolution of Cooperation among LLM Agents. The paper explores whether a society of large language model (LLM) agents can develop cooperat...


tinygrad (George Hotz) ▷ #general (2 messages):

Tinygrad Windows Support, Pull Requests for Windows Bugs


tinygrad (George Hotz) ▷ #learn-tinygrad (2 messages):

Shapetracker Documentation, Matrix Memory Layout, Memory Index Calculation, Stride in Matrix Access

Link mentioned: Shapetracker: Tutorials on tinygrad


MLOps @Chipro ▷ #general-ml (4 messages):

Weights and Biases vs MLflow, Recording Experimentation Results, State of Classical Machine Learning, 1-bit Large Language Models

Link mentioned: The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits: Recent research, such as BitNet, is paving the way for a new era of 1-bit Large Language Models (LLMs). In this work, we introduce a 1-bit LLM variant, namely BitNet b1.58, in which every single param...


Nomic.ai (GPT4All) ▷ #general (4 messages):

AI Reader Tool, Embedding Weights in Content, Indexing by Subject


LAION ▷ #general (2 messages):

AI and 3D modeling with Blender, Animals and EEG, Language/action mapping for animals


LAION ▷ #research (1 messages):

yoavhacohen: https://x.com/yoavhacohen/status/1875148348489113891






{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}