Frozen AI News archive

not much happened today

**Zyphra AI** launched **Zonos-v0.1**, a leading open-weight text-to-speech model supporting multiple languages and zero-shot voice cloning. **Meta FAIR** released the open-source **Audiobox Aesthetics** model trained on 562 hours of audio data. **Kyutai Labs** introduced **Moshi**, a real-time speech-to-speech system with low latency. **Perplexity AI** announced the **Sonar** model based on **Llama 3.3 70b**, outperforming top models like **GPT-4o** and **Claude 3.5 Sonnet** with 1200 tokens/second speed, powered by **Cerebras** infrastructure. **UC Berkeley** open-sourced a 1.5B model trained with reinforcement learning that beats **o1-preview** on math tasks. **ReasonFlux-32B** achieved 91.2% on the MATH benchmark, outperforming **OpenAI o1-preview**. **CrossPoster**, an AI agent for cross-platform posting, was released using **LlamaIndex** workflows. **Brilliant Labs** integrated the **Google DeepMind Gemini Live API** into smart glasses for real-time translation and object identification.

Canonical issue URL

AI News for 2/10/2025-2/11/2025. We checked 7 subreddits, 433 Twitters and 29 Discords (211 channels, and 5891 messages) for you. Estimated reading time saved (at 200wpm): 524 minutes. You can now tag @smol_ai for AINews discussions!

a quiet day. Dan Hendrycks released an interesting study on LLM bias which has come under some questions.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

New Models and Releases

Model Performance and Benchmarking

AI Applications and Tools

AI Safety, Ethics, and Bias

Other Topics

Humor/Memes


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Elon's Offer Complicates OpenAI's For-Profit Transition Plans

Theme 2. DeepScaleR-1.5B: Advancing Reinforcement Learning for Smaller Models

Theme 3. Open-Sourced R1 Reasoning Architecture for LLMs

Other AI Subreddit Recap

/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT

Theme 1. Elon Musk vs Sam Altman: Power Struggle at OpenAI

Theme 2. Grok 3's Underperformance in Competitive LLM Space


AI Discord Recap

A summary of Summaries of Summaries by Gemini 2.0 Flash Thinking

Theme 1. Model Performance and Benchmarking: The AI Model Arena Heats Up

Theme 2. Developer Tools and IDEs: Navigating the AI Code Jungle

Theme 3. Technical Deep Dives: Decoding LLM Challenges and Innovations

Theme 4. AI Applications: From Marketing to Music and Beyond

Theme 5. Infrastructure and Optimization: Powering the AI Revolution


PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord


Cursor IDE Discord


LM Studio Discord


Codeium (Windsurf) Discord


OpenAI Discord


MCP (Glama) Discord


Perplexity AI Discord


GPU MODE Discord


OpenRouter (Alex Atallah) Discord


Notebook LM Discord


aider (Paul Gauthier) Discord


Nous Research AI Discord


Eleuther Discord


Stability.ai (Stable Diffusion) Discord


Latent Space Discord


LlamaIndex Discord


LLM Agents (Berkeley MOOC) Discord


Yannick Kilcher Discord


Torchtune Discord


Nomic.ai (GPT4All) Discord


tinygrad (George Hotz) Discord


Gorilla LLM (Berkeley Function Calling) Discord


Modular (Mojo 🔥) Discord


Cohere Discord


DSPy Discord


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Unsloth AI (Daniel Han) ▷ #general (991 messages🔥🔥🔥):

GRPO vs SFT, Rewards in Fine-Tuning, Using LLMs for Code Assistance, Neural Network Legal Implications, Future of AMD vs NVIDIA

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (5 messages):

EveryOneCoder4x7b, Merging Models into MoE, GRPO Tutorial, Unsloth Tool, Reading Resources


Unsloth AI (Daniel Han) ▷ #help (64 messages🔥🔥):

Exllama performance, Llama 3.3 fine-tuning, DAPT techniques, GPT agent training issues, Hardware for large model training

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (1 messages):

Spark Engine v1, No-code AI sandbox, Integration with Unsloth

Link mentioned: Spark Engine - The AI Sandbox: Turn ideas into AI-powered products, no coding experience required


Unsloth AI (Daniel Han) ▷ #research (12 messages🔥):

Phi 4 limitations, DoRA improvements, Training Mistral, Fine-tuning models, LoRA and vLLM

Link mentioned: Tweet from Wing Lian (caseus) (@winglian): What's the trick? DoRA. I don't have a great hypothesis on why it works yet, but I've upstreamed the changes to TRL. The PR merges the LoRA weights into the base model and ships those to v...


Cursor IDE ▷ #general (568 messages🔥🔥🔥):

Cursor MCP Servers, Usage-Based Pricing, DeepSeek and Perplexity, New Features in Cursor, Implementing Cursor Rules

Links mentioned:


LM Studio ▷ #general (150 messages🔥🔥):

Model Configuration and Usage, Quantization Techniques, Performance of Different Models, LM Studio Capabilities, Music Generation Models

Links mentioned:


LM Studio ▷ #hardware-discussion (413 messages🔥🔥🔥):

GPU Performance and Usage, Intel Integrated Graphics Impact, Model Offloading Techniques, Multiple GPU Configurations, Deep Learning Model Benchmarking

Links mentioned:


Codeium (Windsurf) ▷ #discussion (14 messages🔥):

Codeium Extensions, Windsurf IDE, Jetbrains Connectivity Issues, Alternatives to Codeium, Extension Updates


Codeium (Windsurf) ▷ #windsurf (404 messages🔥🔥🔥):

Windsurf usage issues, Updates and features in Windsurf, Model comparisons in AI tools, Credit usage concerns, Error messages and troubleshooting

Links mentioned:


OpenAI ▷ #ai-discussions (199 messages🔥🔥):

AI Models Performance Comparison, Local LLM Setup, User Frustrations with AI, Spatial Reasoning in LLMs, Market Dynamics in AI

Link mentioned: Reddit - Dive into anything: no description found


OpenAI ▷ #gpt-4-discussions (24 messages🔥):

Using GPT for children's stories, Creating horror stories, Prompt refinement for effective storytelling, Psychological aspects in prompts, Marketing strategies and pitching


OpenAI ▷ #prompt-engineering (6 messages):

Preventing AI Laziness, Importance of Specificity, Iterative Prompting, Model Instruction Conflicts


OpenAI ▷ #api-discussions (6 messages):

Preventing AI Laziness, Iterative Prompting, Conflicting Instructions, Model Limitations


MCP (Glama) ▷ #general (167 messages🔥🔥):

Claude Desktop update issues, MCP server and Python SDK challenges, Sage for Android, Security concerns with MCP servers, OpenRouter authentication options

Links mentioned:


MCP (Glama) ▷ #showcase (2 messages):

Managing DO, OAuth Flows


Perplexity AI ▷ #general (149 messages🔥🔥):

Perplexity's model limitations, Sonar model performance, User interface concerns, R1 model usage, AI support issues

Links mentioned:


Perplexity AI ▷ #sharing (13 messages🔥):

Google's Gemini 2.0 release, Controversy over first iPhone porn app, DeepSeek's impact on energy industry, Various model outputs, Federal Executive Institute insights

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (1 messages):

mastercharter: Has anyone noticed fluctuaing quality in the reasoning models responses


GPU MODE ▷ #general (3 messages):

Nebius Meetup, GPU for Cuda Mode, Kubernetes operator for Slurm, Agentic systems

Link mentioned: Nebius AI Cloud Unveiled. San Francisco Meetup: Discover the most efficient way to build, tune and run your AI models and applications on top-notch NVIDIA® GPUs.


GPU MODE ▷ #triton (14 messages🔥):

Triton vs CUDA, New TMA feature in Triton, Inline ASM in Triton, Debugging in Triton

Links mentioned:


GPU MODE ▷ #cuda (6 messages):

Warp Group Specialized Persistent Kernels, CUDA audio processing, Ping-Pong kernels

Link mentioned: Deep Dive on CUTLASS Ping-Pong GEMM Kernel: In this post, we provide an overview, with relevant FP8 inference kernel benchmarking, of the CUTLASS Ping-Pong GEMM kernel.


GPU MODE ▷ #torch (8 messages🔥):

CPUOffload, DTensor full_tensor, Optimizer Steps


GPU MODE ▷ #algorithms (9 messages🔥):

CPU Attention Implementation, Efficient Scaled Dot-Product Attention, Flex Attention Developments, Memory-Bound Attention, Llama.cpp Attention Operations

Links mentioned:


GPU MODE ▷ #beginner (2 messages):

CUDA Learning Resources

Link mentioned: no title found: no description found


GPU MODE ▷ #pmpp-book (1 messages):

Cooking, Image Presentation


GPU MODE ▷ #off-topic (1 messages):

Quantization-Aware Training, QuEST Method, Model Compression Techniques, Hadamard Transform in LLMs, Comparative Analysis of FP16 and 8-bit Models

Link mentioned: QuEST: Stable Training of LLMs with 1-Bit Weights and Activations: One approach to reducing the massive costs of large language models (LLMs) is the use of quantized or sparse representations for training or deployment. While post-training compression methods are ver...


GPU MODE ▷ #rocm (9 messages🔥):

Optimized FP32 Matrix Multiplication, rocBLAS Optimization Concerns, GPU Kernel Optimization Challenges, CUDA to ROCm Conversion, hipBLAS and Tensile Insights

Link mentioned: Optimizing Matrix Multiplication on RDNA3: 50 TFlops and 60% Faster Than rocBLAS: Introduction


GPU MODE ▷ #intel (8 messages🔥):

Intel Extension for PyTorch, Gaudi Accelerator, Xeon Max and Data Center GPU Max confusion


GPU MODE ▷ #thunderkittens (3 messages):

LayerNorm on A100, Performance of Complex Matmul, Reinterpreting ST as CST


GPU MODE ▷ #edge (1 messages):

PyTorch Edge team updates, ExecuTorch library, Public Discord Channel


GPU MODE ▷ #reasoning-gym (54 messages🔥):

GRPO support for Axolotl, SymBench datasets, Evaluation metrics for reasoning models, DeepScaler model performance, Improving dataset prompts and outputs

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (114 messages🔥🔥):

Websearch functionality, Anthropic computer-use tools, Issues with Gemini model, Chathistory retrieval, Music chord detection AI

Links mentioned:


OpenRouter (Alex Atallah) ▷ #beta-feedback (1 messages):

mazvi: Cool


Notebook LM ▷ #announcements (1 messages):

NotebookLM Plus, Google One AI Premium, Student Discounts, NotebookLM features

Link mentioned: NotebookLM Plus is now available in the Google One AI Premium subscription.: NotebookLM is a research and thinking companion designed to help you make the most of your information. You can upload material, summarize it, ask questions and transfor…


Notebook LM ▷ #use-cases (9 messages🔥):

Customizing Deeper Insights, Technical Support Requests, Optimizing Neural Network Structures, Health Tracking Innovations, Podcast Workflow Instructions

Link mentioned: What makes a good feedforward computational graph?: Open Source Intelligence · Episode


Notebook LM ▷ #general (103 messages🔥🔥):

NotebookLM Access Issues, NotebookLM and Google One Subscription, User Limits in NotebookLM, Notebook Sharing Among Users, Education Use of NotebookLM

Links mentioned:


aider (Paul Gauthier) ▷ #general (89 messages🔥🔥):

DeepSeek Performance Issues, Aider Features and Usability, Architecture Models Usage, Visual Indicators for Processing, CMake Command Issues

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (18 messages🔥):

Aider Custom Model Aliases, Integrating Copilot with Aider, Extending Aider Functionality, Using Aider for AI Code Editing, Benchmarking Starcoder2

Links mentioned:


aider (Paul Gauthier) ▷ #links (3 messages):

SCM Files in LLMap, CodeSteer-v1 Paper

Link mentioned: Paper page - CodeSteer: Symbolic-Augmented Language Models via Code/Text Guidance: no description found


Nous Research AI ▷ #general (87 messages🔥🔥):

Pre-trained Models, Open Source AI Community, Meta's Business Strategy, Elon Musk's Influence on OpenAI, Gemini 2.0 Challenges

Links mentioned:


Nous Research AI ▷ #research-papers (2 messages):

Research topics for medical students, Psychology of medical students


Nous Research AI ▷ #interesting-links (1 messages):

Novel Language Model Architecture, Scaling Test-Time Computation, Recurrent Block Iteration, Reasoning Benchmarks, Parameter Efficiency

Link mentioned: Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach: We study a novel language model architecture that is capable of scaling test-time computation by implicitly reasoning in latent space. Our model works by iterating a recurrent block, thereby unrolling...


Nous Research AI ▷ #research-papers (2 messages):

Research topics for medical students, Psychology of medical students, Innovative research themes


Nous Research AI ▷ #reasoning-tasks (1 messages):

Anthropic's Economic Index, Reasoning Dataset Curriculum

Link mentioned: Anthropic/EconomicIndex · Datasets at Hugging Face: no description found


Eleuther ▷ #general (18 messages🔥):

AI in Trading, Deep Model Loss Issues, Deepfrying in Network Training, Sequence Length Tolerance, Model Depth and Plasticity


Eleuther ▷ #research (74 messages🔥🔥):

Curse of Depth in LLMs, Value Residuals vs. Value Embeddings, Compression Techniques in AI, Gated Skip Connections, Rotary Position Embeddings (RoPE)

Links mentioned:


Eleuther ▷ #interpretability-general (1 messages):

Superposition and Distributed Representations, Follow-up Work on Neural Network Structures, Further Discussions on Toy Testing

Link mentioned: Distributed Representations: Composition & Superposition: no description found


Stability.ai (Stable Diffusion) ▷ #general-chat (70 messages🔥🔥):

Flux Resolution Performance, Differences Between Flux Dev and Schnell, SDXL vs SD 1.5 Quality Comparison, Using Refiners with Models, Artistic Model Recommendations


Latent Space ▷ #ai-general-chat (31 messages🔥):

Data Breach of OpenAI Credentials, Ilya Sutskever's New Startup, AI Alignment and Value Systems, Matryoshka Quantization, Deep Research Insights

Links mentioned:


Latent Space ▷ #ai-announcements (5 messages):

Bret Taylor Podcast, AI Software Engineering, OpenAI Leadership, Customer Experience at SierraPlatform, Future of Autonomous AI

Links mentioned:


LlamaIndex ▷ #blog (3 messages):

GraphRAG pipelines, AI-driven marketing automation, DeepSeek AI deployment

Link mentioned: Dein persönlicher KI Assistent ​für Kampagnen in Life Sciences: KI gestütztes Life Sciences Marketing: Innovative, Künstliche Intelligenz basierte Marketinglösung für Pharma, Medtech, Biotech und Gesundheitswesen, die die Erstellung von Strategien und Marketingunt...


LlamaIndex ▷ #general (28 messages🔥):

AzureAI Search customization, Multi-agent workflows, Integrating MCP tools with LlamaIndex, OpenRouter usage, Blockchain development

Links mentioned:


LLM Agents (Berkeley MOOC) ▷ #mooc-announcements (1 messages):

Lecture 3 with Yu Su, Role of Language in AI, Core Competencies of Language Agents, LLM-based Language Agents

Link mentioned: CS 194/294-280 (Advanced LLM Agents) - Lecture 3, Yu Su: Ask questions here: https://www.bli.do/su-mem3


LLM Agents (Berkeley MOOC) ▷ #mooc-questions (23 messages🔥):

Certificate Completion Issues, Research Track Registration, Lecture Slides Availability, MOOC Curriculum Details, Quiz Links for Lectures


LLM Agents (Berkeley MOOC) ▷ #mooc-lecture-discussion (4 messages):

MOOC Curriculum Release, Reading Assignment Submission, Community Engagement


LLM Agents (Berkeley MOOC) ▷ #mooc-readings-discussion (1 messages):

DeepScaleR, Scaling RL


Yannick Kilcher ▷ #general (8 messages🔥):

Cursor/Copilot Diff Application, Provisional Patent for Vocal Agents, Thinking Models Behavior via SAE, Claude AI Enthusiasm


Yannick Kilcher ▷ #paper-discussion (4 messages):

ICLR 2022 Outstanding Paper S4, Discussion on legS and legT

Link mentioned: Efficiently Modeling Long Sequences with Structured State Spaces: A central goal of sequence modeling is designing a single principled model that can address sequence data across a range of modalities and tasks, particularly on long-range dependencies. Although conv...


Yannick Kilcher ▷ #agents (1 messages):

.sepoy: LLMs can't count at all 🤷


Yannick Kilcher ▷ #ml-news (14 messages🔥):

Anthropic's AI Performance, Microsoft Study on AI and Cognition, Elon Musk's OpenAI Bid, International AI Declaration Refusal, AI Self-Replication Research

Links mentioned:


Torchtune ▷ #general (2 messages):

Approval Process, GitHub Update


Torchtune ▷ #dev (16 messages🔥):

Support for UV package manager, Gradient accumulation in DPO/PPO recipes, Fixes related to checkpoint resuming, Standardization of dependency installations, Quality of tests in development

Links mentioned:


Torchtune ▷ #papers (3 messages):

Novel Language Model Architecture, Dynamic Model Depth, State Space Models, Test-Time Computation

Link mentioned: Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach: We study a novel language model architecture that is capable of scaling test-time computation by implicitly reasoning in latent space. Our model works by iterating a recurrent block, thereby unrolling...


Nomic.ai (GPT4All) ▷ #general (19 messages🔥):

Local AI Tools, Using GPT4All with Voice, Embedding PDFs, Mobile Alternatives to GPT4All, Community Interactions


tinygrad (George Hotz) ▷ #general (6 messages):

Research Before Asking, Closing Stale PRs, Pull Request #7456 Updates, Asking Technical Questions

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (6 messages):

CUDA installation issues, Tinygrad device support, Documentation updates

Link mentioned: docs: note if Device.DEFAULT shows GPU by LytixDev · Pull Request #9033 · tinygrad/tinygrad: Note for the noobs that forgot to install the correct CUDA drivers.


Gorilla LLM (Berkeley Function Calling) ▷ #discussion (4 messages):

HF dataset compatibility, Berkeley Function Calling Leaderboard, GitHub workflow for auto-committing, Dataset visualization needs

Link mentioned: gorilla-llm/Berkeley-Function-Calling-Leaderboard · Datasets at Hugging Face: no description found


Modular (Mojo 🔥) ▷ #mojo (2 messages):

Lazy Evaluation in Mojo, Benchmarking GB/s Parsing Speed


Cohere ▷ #discussions (2 messages):

Monkeys


DSPy ▷ #general (1 messages):

DSPy implementation, Python scripting, MUD server interaction, Llama-3 performance, Metric tracking



{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}