Frozen AI News archive

Bespoke-Stratos + Sky-T1: The Vicuna+Alpaca moment for reasoning

**Reasoning Distillation** has emerged as a key technique, with Berkeley/USC researchers releasing **Sky-T1-32B-Preview**, a finetuned model of **Qwen 2.5 32B** using 17k reasoning traces for just **$450**, matching benchmarks of **o1-preview**. **DeepSeek** introduced **R1**, a model surpassing **o1-preview** and enabling distillation to smaller models like a 1.5B Qwen to match **gpt-4o** and **claude-3-sonnet** levels. **Bespoke Labs** further distilled **R1** on Qwen, outperforming **o1-preview** with fewer samples. This progress suggests that *"SFT is all you need"* for reasoning without major architecture changes. Additionally, **DeepSeek-R1** uses pure reinforcement learning with supervised finetuning to accelerate convergence and shows strong reasoning and multimodal capabilities. **Google's Gemini 2.0 Flash Thinking** model boasts a **1 million token context window**, code execution, and excels in math, science, and multimodal reasoning. Critiques highlight challenges in model repeatability, behavioral self-awareness, and RLHF limitations in reasoning robustness.

Canonical issue URL

AI News for 1/21/2025-1/22/2025. We checked 7 subreddits, 433 Twitters and 34 Discords (225 channels, and 4297 messages) for you. Estimated reading time saved (at 200wpm): 496 minutes. You can now tag @smol_ai for AINews discussions!

In the ChatGPT heyday of 2022-23, Alpaca and Vicuna were born out of LMsys and Stanford as ultra cheap ($300) finetunes of LLaMA 1 that distilled from ChatGPT/Bard samples to achieve 90% of the quality of ChatGPT/GPT3.5.

In the last 48 hours, it seems the Berkeley/USC folks have done it again, this time with the reasoning models.

It's hard to believe this sequence of events happened just in the last 2 weeks:

  1. Berkeley's Sky Computing lab released Sky-T1-32B-Preview, a finetune of Qwen 2.5 32B (our coverage here) with 17k rows of training data from QwQ-32B (our coverage here) + rewriting traces with gpt-4o-mini + rejection sampling, all done for $450. Because QwQ outperforms o1-preview, distilling from QwQ brings Qwen up to match o1-preview's benchmarks: image.png image.png
  2. DeepSeek releases R1 (2 days ago) with benchmarks far above o1-preview. The R1 paper also revealed the surprise that you can distill from R1 to turn a 1.5B Qwen model to match 4o and 3.5 Sonnet (?!).
  3. Bespoke Labs (today) uses the Sky-T1 recipe to distill R1 on Qwen again to greatly outperform (not just match) o1-preview, again with 17k rows of reasoning traces. image.png

While Bespoke's distillation does not quite match DeepSeek's distillation in performance, they used 17k samples vs DeepSeek's 800k. It is pretty evident that they could keep going here if they wished.

The bigger shocking thing is that "SFT is all you need" - no major architecture changes are required for reasoning to happen, just feed in more (validated, rephrased) reasoning traces, backtracking and pivoting and all, and it seems like it will generalize well. In all likelihood, this explains the relative efficiency of o1-mini and o3-mini vs their full size counterparts.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

AI Model Developments and Evaluations

AI Applications and Tools

AI Research and Papers

AI Infrastructure and Compute

AI Community, Education, and Events

Memes/Humor

AI Policy and Ethics


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Mistral 10V: Exploring New Capabilities with 12K Tokens

Theme 2. O1-Pro: Revolutionary Use in Legislation Analysis

Theme 3. Gemini 1.5: Leading AI with Performance Edge

Other AI Subreddit Recap

/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT

Theme 1. Stargate AI Project: $500 Billion Investment's Impact

Theme 2. DeepSeek R1: Redefining AI Benchmarks

Theme 3. Model-Agnostic Reasoning: R1 Techniques

Theme 4. Deepseek R1 GRPO Code: Open-Sourcing Breakthrough

Theme 5. R1-Zero: AI Reinforcement Learning Breakthroughs


AI Discord Recap

A summary of Summaries of Summaries

o1-preview-2024-09-12

Theme 1. AI's Billion-Dollar Stargate Projects: Lofty Goals and Skepticism

Theme 2. AI Models Clash: DeepSeek R1 Outperforms the Giants

Theme 3. Censorship vs. Uncensored AI Models: Users Seek Freedom

Theme 4. New AI Tools and Innovations Empower Developers

Theme 5. AI Development Challenges: From Quantization to Privacy

o1-2024-12-17

Theme 1. AI Infrastructure & Funding Frenzy

Theme 2. LLM Showdowns & Math Marvels

Theme 3. Reinforcement Learning & GRPO Talk

Theme 4. HPC & GPU Codegen Adventures

Theme 5. RAG Systems & Tool Innovations

DeepSeek v3

Theme 1. DeepSeek R1 Model Performance and Integration

Theme 2. AI Model Quantization and Fine-Tuning

Theme 3. AI Infrastructure and Large-Scale Investments

Theme 4. AI for Creative and Technical Applications

Theme 5. AI Safety, Ethics, and Regulation

DeepSeek R1

Theme 1. Model Optimization Wars: Quantization, Fine-Tuning, and Scaling Battles

Theme 2. AI Infrastructure Arms Race: $500B Projects and Hardware Hurdles

Theme 3. Agentic AI: Hype vs. Reality in Autonomous Systems

Theme 4. Tooling Turbulence: IDE Wars, API Quirks, and RAG Realities

Theme 5. Ethics, Censorship, and Workforce Displacement Fears


PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord


Codeium (Windsurf) Discord


LM Studio Discord


Nous Research AI Discord


Interconnects (Nathan Lambert) Discord


aider (Paul Gauthier) Discord


Stackblitz (Bolt.new) Discord


Yannick Kilcher Discord


OpenRouter (Alex Atallah) Discord


Perplexity AI Discord


MCP (Glama) Discord


Latent Space Discord


GPU MODE Discord


Eleuther Discord


OpenAI Discord


Notebook LM Discord Discord


Stability.ai (Stable Diffusion) Discord


LlamaIndex Discord


Modular (Mojo 🔥) Discord


LAION Discord


Cohere Discord


LLM Agents (Berkeley MOOC) Discord


Nomic.ai (GPT4All) Discord


DSPy Discord


Torchtune Discord


OpenInterpreter Discord


Gorilla LLM (Berkeley Function Calling) Discord


tinygrad (George Hotz) Discord


Axolotl AI Discord


Mozilla AI Discord


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The HuggingFace Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Unsloth AI (Daniel Han) ▷ #general (266 messages🔥🔥):

Model Quantization, Fine-tuning Models, DeepSeek R1 Model Support, Chat Templates and Thinking Tags, Dynamic 4-bit Quantization

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (6 messages):

Unsloth training on Medium, Weights & Biases ETA tracking, Custom code for ETA, Fine-tuning challenges

Link mentioned: Fine-Tuning Llama-3.1-8B for Function Calling using LoRA: Leveraging Unsloth for fine-tuning with Weights & Biases integration for monitoring and vLLM for model serving


Unsloth AI (Daniel Han) ▷ #help (146 messages🔥🔥):

Phi-4 model issues, Running DPO training script, Unsloth notebooks updates, Using Triton with Unsloth, Fine-tuning suggestions for different models

Links mentioned:


Unsloth AI (Daniel Han) ▷ #research (191 messages🔥🔥):

Synthetic Data Training, Chinchilla Optimal Models, Emotion Tracking in AI, Agentic AI Claims, Dynamic vs. Static Learning Systems

Links mentioned:


Codeium (Windsurf) ▷ #discussion (48 messages🔥):

Codeium Extension Updates, Windsurf IDE Issues, Model API Integration, Diff Viewer Difficulties, Privacy Policy Queries

Links mentioned:


Codeium (Windsurf) ▷ #windsurf (426 messages🔥🔥🔥):

Windsurf's Auto-Generated Memories, Development Challenges with Cascade, Prompt Engineering and Context, AI Integration in Programming, User Experiences with Windsurf

Links mentioned:


LM Studio ▷ #announcements (1 messages):

LM Studio 0.3.8, Thinking UI, LaTeX rendering, Bug fixes

Link mentioned: Tweet from LM Studio (@lmstudio): LM Studio 0.3.8 🚢- Thinking UI for DeepSeek R1- LaTeX rendering improvements- Bug fixes


LM Studio ▷ #general (209 messages🔥🔥):

DeepSeek R1 performance, Model loading issues on Mac, Quantization settings, Using LMStudio effectively, Math problem-solving capabilities of models

Links mentioned:


LM Studio ▷ #hardware-discussion (180 messages🔥🔥):

DeepSeek R1 Pricing, MacBook Performance for LLMs, GPUQuantization, NVIDIA Digits Requirements, Model Size Efficiency

Link mentioned: bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF · Hugging Face: no description found


Nous Research AI ▷ #general (295 messages🔥🔥):

AI Model Development, Chip Manufacturing Competition, Blockchain and Crypto, DeepSeek Reasoning Extraction, Project Stargate Funding

Links mentioned:


Nous Research AI ▷ #ask-about-llms (6 messages):

Mechanical Interpretation of Model Activations, DeepSeek Capabilities, Synthetic Data Generation

Link mentioned: Tweet from LLM Fan (@llm_fan): I think LLM's are catching onto the 'r's in strawberry question. I asked deepseek"how many p's are there in the correct word for razzberry?" ( I tried 'proper word' als...


Nous Research AI ▷ #research-papers (5 messages):

FLAME model, Human-AI representation alignment

Links mentioned:


Nous Research AI ▷ #interesting-links (12 messages🔥):

Automated architecture search for LLMs, EvaByte tokenizer-free model, Tensor networks in ML

Links mentioned:


Nous Research AI ▷ #research-papers (5 messages):

FLAME model, Human-AI representation similarities, Small model advantages

Links mentioned:


Nous Research AI ▷ #reasoning-tasks (1 messages):

lowiqgenai: Hey i did some using MistralAI free Services fhai50032/medmcqa-solved-thinking-o1


Interconnects (Nathan Lambert) ▷ #news (87 messages🔥🔥):

Stargate Project Funding, Updates on AI Models, Google's Investment in Anthropic, Flash Thinking Model, Concerns over AI Data Usage

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (16 messages🔥):

Microsoft's Investment in Stargate, Billionaires on Twitter, AI Safety Discussions, Model Alignment, Influence of Tech Leaders

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (88 messages🔥🔥):

Robonato humor, OpenAI Media Manager developments, DeepSeek advancements, Creative writing benchmarks, AI podcast editing tools

Links mentioned:


Interconnects (Nathan Lambert) ▷ #memes (4 messages):

AI Safety Index, Whitehouse.com betting concerns, Undefined discussions, Contrast in social media posts

Links mentioned:


Interconnects (Nathan Lambert) ▷ #rl (13 messages🔥):

GRPO concerns, T1 RL paper, Deepseek response, HLF training plots, TRL GitHub discussion

Links mentioned:


Interconnects (Nathan Lambert) ▷ #reads (12 messages🔥):

Davos Interviews, AI Regulations, Transistor Radio Podcasts

Links mentioned:


Interconnects (Nathan Lambert) ▷ #posts (3 messages):

SnailBot News, Bot Performance


Interconnects (Nathan Lambert) ▷ #policy (4 messages):

Arms Race in Technology, Live Events, Stargate Discussions


aider (Paul Gauthier) ▷ #general (122 messages🔥🔥):

Gemini 2.0 Flash Thinking Model, Aider Workflow Enhancements, Model Comparison and Critique, Markdown Specifications for Development, RAG Approach with PDF References

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (91 messages🔥🔥):

Aider model configurations, Error handling in Aider, Using OpenAI keys with Aider, Integrating various models, Neovim plugins for Aider

Links mentioned:


aider (Paul Gauthier) ▷ #links (1 messages):

astor1: https://github.com/NixOS/nixpkgs/pull/375634 aider 0.72.1 in nixpkgs


Stackblitz (Bolt.new) ▷ #announcements (1 messages):

Bolt funding, Community appreciation

Link mentioned: Tweet from bolt.new (@boltdotnew): Today we're announcing $105.5m in funding to take Bolt to new heights! 🚀Our Series B was led by Emergence & GV, with participation from Madrona, The Chainsmokers (Mantis), Conviction, and some of...


Stackblitz (Bolt.new) ▷ #prompting (16 messages🔥):

Netlify routing issues, NextJS and Supabase integration, SSR challenges with large NextJS projects, Building a Tetris mini app for Telegram

Link mentioned: Telegram Apps Center: Community-driven catalog of applications developed by third-party developers. Not affiliated with Telegram Messenger.


Stackblitz (Bolt.new) ▷ #discussions (192 messages🔥🔥):

Bolt recursion issues, Token upgrade inquiries, User permissions and policies, CORS issues, Using Claude for troubleshooting

Links mentioned:


Yannick Kilcher ▷ #general (161 messages🔥🔥):

R1's Performance, DeepSeek's Advancements, AI Ethical Dilemmas, OpenAI vs. Competitors, AI Infrastructure Investments

Links mentioned:


Yannick Kilcher ▷ #paper-discussion (32 messages🔥):

DeepSeek R1 Model Performance, Challenges in Paper Reviewing, DeepSeekMath Paper Insights, Anthropomorphism in Models, DeepSeek Training Procedure

Links mentioned:


Yannick Kilcher ▷ #agents (1 messages):

IntellAgent, Conversational Agents Evaluation, Synthetic Interactions

Link mentioned: GitHub - plurai-ai/intellagent: A framework for comprehensive diagnosis and evaluation of conversational agents using simulated, realistic synthetic interactions: A framework for comprehensive diagnosis and evaluation of conversational agents using simulated, realistic synthetic interactions - plurai-ai/intellagent


Yannick Kilcher ▷ #ml-news (3 messages):

Stargate Project, UI-TARS Model, OpenAI Operator Feature

Links mentioned:


OpenRouter (Alex Atallah) ▷ #announcements (1 messages):

Web Search Pricing, API Access Launch


OpenRouter (Alex Atallah) ▷ #general (187 messages🔥🔥):

DeepSeek model performance, DeepSeek R1 issues, Censorship concerns, Uncensored models, Cerebras model availability

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):

Sonar API, Sonar Pro, AI Companion 2.0, SimpleQA benchmark, Data security

Link mentioned: Sonar by Perplexity: Build with the best AI answer engine API, created by Perplexity. Power your products with the fastest, cheapest offering out there with search grounding. Delivering unparalleled real-time, web-wide re...


Perplexity AI ▷ #general (128 messages🔥🔥):

Sonar API Performance Issues, Pro Model Usage Confusion, Model Comparison and Updates, Login and Server Errors, API Functionality and Documentation

Links mentioned:


Perplexity AI ▷ #sharing (8 messages🔥):

Perplexity API usage, Anduril's autonomous weapons, PhD-level Super Agents, College basketball, pyctc_decode for projects

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (16 messages🔥):

Sonar Pro API, Search Domain Filter, Error Messages, Deployment in Europe, Comparison Tool

Links mentioned:


MCP (Glama) ▷ #general (114 messages🔥🔥):

MCP Server Functionality, Brave Search for Documentation, Custom GPT Limitations, Code Editing with MCP, Prompt System in Claude Desktop

Links mentioned:


MCP (Glama) ▷ #showcase (7 messages):

MCP Server for Apify's Actors, Anthropic TS Client Issues, Connecting with SSE

Links mentioned:


Latent Space ▷ #ai-general-chat (96 messages🔥🔥):

Ai2 ScholarQA, Project Stargate, Decline of Stack Overflow, Bespoke-Stratos-32B, Investments in AI infrastructure

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

LLM Paper Club, Physics of Language Models, Retroinstruct, Event Notifications, Calendar Integration

Link mentioned: LLM Paper Club (Physics of Language Models, Retroinstruct) · Zoom · Luma: A 2 for 1 day!Shamima will cover https://arxiv.org/abs/2404.05405 and this guide on synthetic datasets…


GPU MODE ▷ #general (5 messages):

Proximal Policy Optimization, GRPO implementation, ChatGPT jailbreak possibilities, AI security concerns

Link mentioned: Policy Gradient Algorithms: [Updated on 2018-06-30: add two new policy gradient methods, SAC and D4PG.][Updated on 2018-09-30: add a new policy gradient method, TD3.][Updated on 2019-02-09: add SAC with automatically adjusted te...


GPU MODE ▷ #triton (18 messages🔥):

TMA Implementation Issues, Persistent Matmul Descriptors, TRITON_INTERPRET Behavior, Data Dependency in Triton Kernels, Group Implementation for GPU Papers

Link mentioned: GridQuant/scripts/gemm.py at main · niconunezz/GridQuant: An attempt to implement GridQuant. Contribute to niconunezz/GridQuant development by creating an account on GitHub.


GPU MODE ▷ #cuda (16 messages🔥):

NVIDIA Blackwell Codegen, Emulating GPUs on CPUs, Upcoming Blackwell Whitepaper, Accel-Sim Framework, STF Discussion

Links mentioned:


GPU MODE ▷ #torch (6 messages):

Torch Nightly with Triton 3.2, Torch Lightning + DeepSpeed Checkpointing, Learning Rate Schedulers, Torch Profiler Run Times


GPU MODE ▷ #beginner (13 messages🔥):

Speeding up Hugging Face generate(), GPU recommendations for programming, Challenges running large models, Cloud GPU rental options, Budget considerations for GPU setups

Link mentioned: Cloud GPUs: no description found


GPU MODE ▷ #pmpp-book (4 messages):

New Content in PMPP Book, Programming Exercises in PMPP Book, Cloud GPU Comparison for CUDA Programming

Links mentioned:


GPU MODE ▷ #jax (1 messages):

woct0rdho: Why can jax run fp8 operations on CUDA with sm < 89, but pytorch cannot?


GPU MODE ▷ #torchao (7 messages):

Triton 3.2 issues, torch.compile failures, AttrsDescriptor API breaks, Joint Triton project proposal

Links mentioned:


GPU MODE ▷ #off-topic (9 messages🔥):

Pexels API Usage, Pixabay as Alternative, Image Fetching Limits, Automated Queries Concerns


GPU MODE ▷ #self-promotion (4 messages):

Triton Livestreaming, Accelerating LLM Inference, LeetGPU Updates


GPU MODE ▷ #arc-agi-2 (10 messages🔥):

GRPO Algorithm Implementation, Kimi-k1.5 Paper Discussion, Curriculum Learning in RL, Tiny GRPO Repository, RL-hyped Experimentation

Links mentioned:


Eleuther ▷ #general (26 messages🔥):

Colab workflow challenges, AI podcasts recommendations, Google Titans model insights, Data transfer to rented servers issues, Grokking and numerical stability findings

Links mentioned:


Eleuther ▷ #research (16 messages🔥):

DeepSeek reward model architecture, Learning from Egomotion in Vision, Parametric loss functions with differentiable updates, Domino effect in skill learning, Efficient linear attention mechanisms

Links mentioned:


Eleuther ▷ #lm-thunderdome (29 messages🔥):

Minerva Math, Math-500 Dataset, DeepSeek AI Performance, Chat Template Requirements, Long Context Tasks

Links mentioned:


Eleuther ▷ #gpt-neox-dev (2 messages):

Exporting model to HF format, RuntimeError during conversion, Multi-node training configuration

Link mentioned: {: "pipe_parallel_size": 0, "model_parallel_size": 4, "make_vocab_size_divisible_by": 1, # model settings "num_layers": 32, &a...


OpenAI ▷ #ai-discussions (44 messages🔥):

DeepSeek functionality, AI in cybersecurity, User engagement with AI models, Investment expectations in AI, Accessibility of AI tools

Link mentioned: Sade - Smooth Operator - Official - 1984: Sade – Smooth OperatorDirector - Julien Temple - September 1984 The official YouTube channel for the British iconic band Sade www.sade.comSade (vocals) Stuar...


OpenAI ▷ #gpt-4-discussions (2 messages):

Custom GPT with image training, File upload functionality in API


OpenAI ▷ #prompt-engineering (6 messages):

OCR examples impact, Reading maps with OCR, OpenAI's O series model improvements


OpenAI ▷ #api-discussions (6 messages):

OCR and Hallucinations, Mapping Use Cases, OpenAI's O Series Models


Notebook LM Discord ▷ #use-cases (10 messages🔥):

NotebookLM in Church Services, NotebookLM for Study Workflow, Audio Content Generation Issues, Prompt Optimization with Gemini, CompTIA A+ Resource Creation

Link mentioned: We Need to Talk About NotebookLM: Is it the missing link in your AI game?


Notebook LM Discord ▷ #general (43 messages🔥):

APA Reference Generation, Notebook-LM Customization, Audio Overview Generation, Chrome Extension for Prompts, Creativity in Responses

Link mentioned: How does your business utilize artificial intelligence (AI)?: We are gathering data on how various businesses incorporate AI into their operations. Your insights will help us understand the diverse applications of AI across different sectors.


Stability.ai (Stable Diffusion) ▷ #general-chat (53 messages🔥):

De-distilled flux models performance, AI art public perception, Discord bot scams, CitivAI maintenance, Fixing faces in swarmUI


LlamaIndex ▷ #blog (3 messages):

AgentWorkflow release, DeepSeek-R1 model, Open-source RAG system guide


LlamaIndex ▷ #general (46 messages🔥):

LlamaIndex doc website bugs, Cached Augmented Generation with Gemini, Custom Reader for Python File Objects, Domain-specific vector stores in LlamaIndex, AgentWorkflow parallel calls

Links mentioned:


Modular (Mojo 🔥) ▷ #general (6 messages):

Community Showcase, Forum vs Discord Discussions, Project Sharing Clarity, Nightly Developments


Modular (Mojo 🔥) ▷ #mojo (29 messages🔥):

Mojo Domain Observations, MLIR Parallelization, Rust Work-Stealing Scheduler, Mojo Function Overriding, Async Programming Challenges


LAION ▷ #general (34 messages🔥):

SunoMusic Audio Input Feature, Audio Captioning Challenges, Audio Dataset Projects, Emotional Open Source TTS, High School Teacher Volunteering

Links mentioned:


Cohere ▷ #discussions (25 messages🔥):

OpenAI API for LLMs, Text Generation Models, Research Assistance in ML, Cohere's Model Development, Image to Video Generation

Link mentioned: <a href="https://api.deepseek.com")">no%22%3Eno) title found: no description found


Cohere ▷ #questions (1 messages):

Channel Sunset Announcement, Support Streamlining, New Model/API Questions


Cohere ▷ #api-discussions (1 messages):

competent: This channel is staying open!


Cohere ▷ #cmd-r-bot (6 messages):

Cohere Command R+ 08-2024 model, Duplicate content in chatbot responses, LlamaIndex integration, Troubleshooting suggestions, Cohere API version discussions


LLM Agents (Berkeley MOOC) ▷ #mooc-questions (20 messages🔥):

MOOC Syllabus Release, Guest Speaker Suggestions, LLM Hackathon Updates, Spring MOOC Content, Research Collaboration Interest


Nomic.ai (GPT4All) ▷ #general (17 messages🔥):

DeepSeek R1 models, API Key challenges, Language barriers in discussions, GPT4All updates, Chatbot integration in WordPress

Link mentioned: Model Catalog - LM Studio: The latest and greatest LLMs you can run on your computer.


DSPy ▷ #general (9 messages🔥):

DSPy-based RAG with dynamic data, Collaboration in DSPy research, Using DSPy with LM Studio REST API, Errors in DSPy with Ollama, Repo spam concerns


DSPy ▷ #examples (1 messages):

Model functionality, Using LM-Studio models


Torchtune ▷ #dev (8 messages🔥):

Custom Loss Functions in RLHF, Phi 4 PR Updates, Context for PR Discussions, Passing Custom Forward Functions, Deprecation of SimPO

Link mentioned: Custom losses redesign in alignment section · Issue #2206 · pytorch/torchtune: We have passed several iterations speaking about custom contrastive losses in torchtune. The last point in this direction was the deprecation of SimPO #2062 and the prohibition of new custom losses...


Torchtune ▷ #papers (2 messages):

Nature Communications Feature


OpenInterpreter ▷ #general (5 messages):

OpenInterpreter 1.0, Python code execution, Markdown and TXT formatting


Gorilla LLM (Berkeley Function Calling) ▷ #discussion (3 messages):

Gorilla Model from Ollama, LLaMA v2 Model Specifications

Link mentioned: adrienbrault/gorilla-openfunctions-v2:Q6_K/model: https://huggingface.co/gorilla-llm/gorilla-openfunctions-v2-gguf


tinygrad (George Hotz) ▷ #learn-tinygrad (2 messages):

Windows tests for Tinygrad, GPU support via OpenCL

Link mentioned: Windows tests ci by c143 · Pull Request #8715 · tinygrad/tinygrad: no description found


Axolotl AI ▷ #general (1 messages):

KTO Loss, Liger Kernel, Model Merging


Mozilla AI ▷ #announcements (1 messages):

Local-First X AI Hackathon, San Francisco events, Hackathon planning, February 22




{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}