Frozen AI News archive

OpenAI Sora Turbo and Sora.com

**OpenAI** launched **Sora Turbo**, enabling text-to-video generation for ChatGPT Plus and Pro users with monthly generation limits and regional restrictions in Europe and the UK. **Google** announced a quantum computing breakthrough with the development of the **Willow chip**, potentially enabling commercial quantum applications. Discussions on **O1** model performance highlighted its lag behind **Claude 3.5 Sonnet** and **Gemini** in coding tasks, with calls for algorithmic innovation beyond transformer scaling. The **Llama 3.3 Euryale v2.3** model was praised for storytelling and roleplay capabilities, with users suggesting parameter tuning to reduce creative liberties and repetition. Alternatives like **Mistral-Large**, **Behemoth**, and **Endurance v1.1** were also noted. Additionally, **Nvidia** faces an anti-monopoly investigation in China. Memes and humor around GPU issues and embargo mishaps were popular on social media.

Canonical issue URL

AI News for 12/6/2024-12/9/2024. We checked 7 subreddits, 433 Twitters and '31 Discords (206 channels, and 16978 messages) for you. Estimated reading time saved (at 200wpm): 1953 minutes. You can now tag @smol_ai for AINews discussions!

Sora launched today to all ChatGPT Plus and Pro users at no additional cost... but requiring a signup that was disabled because of the intense load.

https://www.youtube.com/live/2jKVx2vyZOY

While we wait for the GPUs to cool, you can watch the onboarding videos, watch MKBHD's botched embargo or listen to Latent Space's coverage of Generative Video World Simulators.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

Here are the key themes and discussions from the Twitter data, organized by major topics:

Sora Launch & Availability

Quantum Computing Breakthrough at Google

O1/Claude Model Performance Discussions

Memes & Humor


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Meta's LLaMA 3.3 Euryale v2.3 excites storytelling enthusiasts

Theme 2. Nvidia faces anti-monopoly investigation in China

Theme 3. Hugging Face's Apache 2.0 Image Dataset release

Theme 4. EXAONE 3.5 models get tested in GPU-Poor Arena

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

Theme 1. Sora Video Generation Launched to Mixed Reception

Theme 2. ChatGPT's Humorous Side: Users Share Insights

Theme 3. OpenAI's Pro Subscription Pricing Under Fire

Theme 4. Criticism of "AI Gotcha" Tests: A Reflective Discourse


AI Discord Recap

A summary of Summaries of Summaries by O1-preview

Theme 1. Llama 3.3 Models: Releases, Fine-Tuning, and Challenges


Theme 2. Gemini and Sora: The AI Showdown


Theme 3. AI Model Performance and Comparisons


Theme 4. Tools and Techniques for AI Efficiency


Theme 5. AI in Development: Challenges and Solutions



PART 1: High level Discord summaries

Codeium / Windsurf Discord


Cursor IDE Discord


Unsloth AI (Daniel Han) Discord


aider (Paul Gauthier) Discord


Modular (Mojo 🔥) Discord


Bolt.new / Stackblitz Discord


OpenRouter (Alex Atallah) Discord


LM Studio Discord


Latent Space Discord


Eleuther Discord


Notebook LM Discord Discord


Cohere Discord


Nous Research AI Discord


DSPy Discord


LlamaIndex Discord


Torchtune Discord


tinygrad (George Hotz) Discord


LLM Agents (Berkeley MOOC) Discord


OpenInterpreter Discord


LAION Discord


Axolotl AI Discord


Mozilla AI Discord


AI21 Labs (Jamba) Discord


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The HuggingFace Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Gorilla LLM (Berkeley Function Calling) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Codeium / Windsurf ▷ #announcements (6 messages):

Cascade Pricing Changes, Windsurf 1.0.7 Release, New Support Ticketing System, Pro Plan Pricing Honor, Cascade Features Update

Links mentioned:


Codeium / Windsurf ▷ #discussion (710 messages🔥🔥🔥):

Windsurf Pricing and Credits, AI Limitations and User Experiences, IDE Comparisons: Cursor vs. Windsurf, User Challenges with Sign-Up and Payment, Suggestions for Improvement in AI Interaction

Links mentioned:


Codeium / Windsurf ▷ #windsurf (508 messages🔥🔥🔥):

Windsurf Pricing Model, Model Switching Strategies, AI Context Windows, Codeium Features, User Experiences with Cascade

Links mentioned:


Cursor IDE ▷ #general (929 messages🔥🔥🔥):

Cursor performance issues, Current state of AI models, API usage and pricing comparison, Comparative experiences with Cursor and Windsurf, Feedback on open AI models

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (610 messages🔥🔥🔥):

Fine-tuning Llama 3.3, Collators and packing, Using models on limited hardware, Performance of AWQ and LoRA, Sora and its impact

Links mentioned:

  1. New methods for merging LoRA weights. Refer this…": no description foundunsloth/Llama-3.3-70B-Instruct-bnb-4bit · Hugging Face: no description foundFinetune Llama 3.1 with Unsloth: Fine-tune and run Meta's updated Llama 3.1 model with 6x longer context lengths via Unsloth!gist:9e912fe4390c3a6959340afb19804566: GitHub Gist: instantly share code, notes, and snippets.unsloth/Llama-3.3-70B-Instruct-GGUF at main: no description foundReddit - Dive into anything: no description foundGitHub - teknium1/ShareGPT-Builder: Contribute to teknium1/ShareGPT-Builder development by creating an account on GitHub.unsloth (Unsloth AI): no description foundUpdate Model Conversion Command in save.py to convert_hf_to_gguf.py by malibayram · Pull Request #730 · unslothai/unsloth: Update Model Conversion Command in save.py to convert_hf_to_gguf.pyDescription:This PR updates the model conversion command in save.py to use convert_hf_to_gguf.py, aligning with the latest tools...Reddit - Dive into anything: no description foundunsloth/Llama-3.3-70B-Instruct · Hugging Face: no description foundGitHub - ggerganov/llama.cpp: LLM inference in C/C++: LLM inference in C/C++. Contribute to ggerganov/llama.cpp development by creating an account on GitHub.Modular Model Spec: no description found

Unsloth AI (Daniel Han) ▷ #off-topic (30 messages🔥):

Open-source projects, Harmony project, Mental health data, LLM competition, Natural Language Processing

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (214 messages🔥🔥):

Multi-GPU training with Unsloth, Error resolution in unsloth installations, Optimization of models for specific tasks, Fine-tuning Llama models for various applications, Accessing and using model weights

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (2 messages):

Awesome RAG, Upcoming Articles

Link mentioned: GitHub - lucifertrj/Awesome-RAG: RAG-VectorDB-Embedings-LlamaIndex-Langchain: RAG-VectorDB-Embedings-LlamaIndex-Langchain. Contribute to lucifertrj/Awesome-RAG development by creating an account on GitHub.


Unsloth AI (Daniel Han) ▷ #research (11 messages🔥):

OpenAI Triton library, AWQ quantization approach, Hyperfitting phenomenon, Memory-efficient optimization techniques, Text-based model releases

Links mentioned:


aider (Paul Gauthier) ▷ #general (430 messages🔥🔥🔥):

Gemini model performance, O1 Pro vs Sonnet, AI assistant usage for coding, File handling in O1 Pro, Quantum computing advancements

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (78 messages🔥🔥):

Aider's functionality and modes, Troubleshooting API rate limits, Script automation in Aider, Aider integration with language servers, Aider's approach to handling new files

Links mentioned:


Modular (Mojo 🔥) ▷ #general (61 messages🔥🔥):

Mojo Compiler Features, Forum Bug Reports, Merchandise Requests, AI-Generated Content Policy

Links mentioned:


Modular (Mojo 🔥) ▷ #announcements (1 messages):

Modular Forum Launch, Discord's Role, Swag Challenge, Ask Ahmed about GPU Programming, Forum Feedback


Modular (Mojo 🔥) ▷ #mojo (283 messages🔥🔥):

Mojo Language Features, Linear Types Proposal, Game Development with Mojo, Comparison with Other Languages, Memory Management in Programming

Links mentioned:


Bolt.new / Stackblitz ▷ #prompting (16 messages🔥):

Bolt functionality issues, Prompting conventions in Bolt, Feature implementation challenges, Variable sensitivity in prompts, Tools for improving prompts

Link mentioned: I made a website that makes websites: 📚 𝗠𝗮𝘁𝗲𝗿𝗶𝗮𝗹𝘀/𝗥𝗲𝗳𝗲𝗿𝗲𝗻𝗰𝗲𝘀:GitHub Repository (give it a star ⭐) → https://github.com/hkirat/bolt.newer0:00 - Introduction and Architecture Di...


Bolt.new / Stackblitz ▷ #discussions (318 messages🔥🔥):

Token Management in Bolt, Supabase Integration, Technical Issues with Bolt, Open Source vs Production Version, Community Resources for Bolt

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (2 messages):

Countless.dev launch, Claude 3.5 Sonnet updates, Integration with Poe

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (318 messages🔥🔥):

Llama Models, API Errors, Sora Model Features, OpenRouter Rate Limits, Mistral Model Updates

Links mentioned:


OpenRouter (Alex Atallah) ▷ #beta-feedback (13 messages🔥):

Integration Beta Feature Requests, Custom Provider Keys, Amazon Bedrock Model Integrations, Google Flash Model Access


LM Studio ▷ #general (173 messages🔥🔥):

LM Studio GPU Usage, Aider Integration Issues, Model Compatibility with LM Studio, Frontend Clients for LM Studio, Hardware Recommendations for AI Models

Links mentioned:


LM Studio ▷ #hardware-discussion (98 messages🔥🔥):

LM Studio server capabilities, GPU setups and cooling solutions, Memory bandwidth and CPU performance, ROCm vs CUDA compatibility, Custom GPU riser designs

Links mentioned:


Latent Space ▷ #ai-general-chat (88 messages🔥🔥):

Gemini exp 1206, Aurora image model, Sora video generation, WaveForms AI, NeurIPS conference

Links mentioned:


Latent Space ▷ #ai-in-action-club (136 messages🔥🔥):

NeurIPS preparation, Networking at conferences, The role of tabular data in industry, Paper Club events, Communication tools for conferences

Links mentioned:


Eleuther ▷ #general (21 messages🔥):

Llama 3.3 Weights Release, Open-ended Information Storage Challenges, Text Adventure Continuity Issues, Eleuther Eval Harness Modification, JAX/Flax Model Integration

Links mentioned:


Eleuther ▷ #research (81 messages🔥🔥):

Variational encoders in different modalities, Memory-efficient optimizers, 3D generation frameworks, Catastrophic forgetting in training, Performance of Adam vs SGD

Links mentioned:


Eleuther ▷ #scaling-laws (2 messages):

GitHub Gist code sharing, Scaling laws overview

Link mentioned: get_scaling_laws: GitHub Gist: instantly share code, notes, and snippets.


Eleuther ▷ #interpretability-general (11 messages🔥):

Gradient Routing, Neural Network Specialization, Causal Inference and Gradient Routing, Credible Source Distinction, Interpretable Architecture

Link mentioned: Tweet from Alex Turner (@Turn_Trout): 1) AIs are trained as black boxes, making it hard to understand or control their behavior. This is bad for safety! But what is an alternative? Our idea: train structure into a neural network by config...


Eleuther ▷ #lm-thunderdome (57 messages🔥🔥):

MLX Examples PR, Eleuther AI Eval Harness, GSM8K Comparison Issues, ARC-Challenge Dataset Anomalies, Llama Model Evaluation Techniques

Links mentioned:


Eleuther ▷ #multimodal-general (8 messages🔥):

VLM Training Process, Causal Loss in VLMs, MSE on Visual Tokens, Apple AIM


Eleuther ▷ #gpt-neox-dev (1 messages):

karatsubabutslower: CC <@367104793292046338> Any hints for this?


Notebook LM Discord ▷ #use-cases (28 messages🔥):

Podcast Lengths, NotebookLM Use Cases, Interactive Storytelling, Data Handling in Sheets, NotebookLM Podcast Prompts

Links mentioned:


Notebook LM Discord ▷ #general (141 messages🔥🔥):

NotebookLM Limitations, Language Support in NotebookLM, Podcast Features Comparison, Audio Overview Issues, Using NotebookLM for Study

Links mentioned:


Cohere ▷ #discussions (65 messages🔥🔥):

Unsloth Finetuning Framework, Building Chat Models, AI in Commerce, Emotional Expression in Voice Generation, Traditional Chinese AI Training


Cohere ▷ #questions (59 messages🔥🔥):

Quantizing aya-expense model, LLM Deployment Options, Vector-based retrieval methods, Multi-step tool use in RAG, Community Engagement in AI Research

Links mentioned:


Cohere ▷ #api-discussions (17 messages🔥):

Dataset Upload Issues, File Format Errors, Absolute Path Recommendations, Cohere Dashboard Upload, Sample File Assistance

Link mentioned: Login | Cohere: Login for access to advanced Large Language Models and NLP tools through one easy-to-use API.


Cohere ▷ #cohere-toolkit (6 messages):

Introduction Messages, Cohere Toolkit Questions


Nous Research AI ▷ #announcements (1 messages):

Neurips Hangout


Nous Research AI ▷ #general (125 messages🔥🔥):

a16z Crypto Ideas, Nous Research Updates, AI x Crypto Discussions, Video on AI Development, DCT and Transpose in DeMo

Links mentioned:


Nous Research AI ▷ #ask-about-llms (4 messages):

Momentum in training, In-context learning efficiency, O1-type synthetic data generation


Nous Research AI ▷ #research-papers (6 messages):

Notable LLM Papers from Last Two Years, Mixture of Experts in LLMs, Resource-Efficient Training for LLMs, Training Small LLMs and Diffusion Models, Challenges in LLM Training

Links mentioned:


Nous Research AI ▷ #research-papers (6 messages):

Top AI/ML Research Papers, Medical AI Research, Mixture of Experts in LLMs, High-Efficiency LLM Training, Survey of Impactful LLM Papers

Links mentioned:


DSPy ▷ #general (66 messages🔥🔥):

Issues with Running Ollama Locally, Exploring Human Feedback in DSPy, Deployment Strategies for DSPy Programs, Using DSPy for Context-Aware Chunking, Anthropic Model Context Protocol with DSPy

Links mentioned:


LlamaIndex ▷ #blog (5 messages):

LlamaParse Multimodal Parsing, Claude Desktop PDF Integration, Agentless Software Issue Resolution, LlamaParse Auto Mode Benefits

Link mentioned: no title found: no description found


LlamaIndex ▷ #general (35 messages🔥):

Automating Ingestion Pipelines, LlamaIndex RAG Integration, LlamaParse Server Locations, Llama3 Cookbook for Intel Gaudi, OpenAI Seed Mechanism

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):

Chain of Thought Prompting, COT techniques, AI problem-solving

Link mentioned: What is Chain of Thought Prompting in AI?: Chain of Thought Prompting (CoT) - OverviewAn artificial intelligence method called Chain of Thought prompting encourages sequential thinking, which enables models to handle challenging tasks more eff...


Torchtune ▷ #general (34 messages🔥):

Adaptive Batching, Llama 3.3 Config Memory Issues, Flex Attention Kernel Bugs, New CPU Flex Kernel, Memory Optimization Techniques

Links mentioned:


Torchtune ▷ #dev (6 messages):

int8 mixed-precision training, AdamW optimizer usage, batch size adjustments, streamlining pre-commit, just command runner

Link mentioned: just/examples/pre-commit.just at master · casey/just: 🤖 Just a command runner. Contribute to casey/just development by creating an account on GitHub.


Torchtune ▷ #papers (1 messages):

Agents' method changes, Allegations of financial misappropriation


tinygrad (George Hotz) ▷ #general (7 messages):

Inf/Nan Handling in Code, Tinygrad Developer Engagement, TinyStats Improvement Suggestions, Upcoming Meeting Agenda, Smart Question Guidelines

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (28 messages🔥):

TinyJit behavior, Training with JIT, Data loading issues, Learning rate scheduling, Librosa installation problems

Links mentioned:


LLM Agents (Berkeley MOOC) ▷ #mooc-questions (32 messages🔥):

Assignment Deadlines, Lab Submission Results, Written Article Submission, Hackathon Participation, Certificate Distribution

Links mentioned:


LLM Agents (Berkeley MOOC) ▷ #mooc-lecture-discussion (3 messages):

Written Article Assignment, GPT-4 Function Calling Mechanism, Code Datasets for Training


OpenInterpreter ▷ #general (25 messages🔥):

Access to OpenInterpreter App, Model Compatibility and Tool Calls, Multi-Agent Systems Discussion, User Approval Workflow for Commands, User Experience with OI Pro


OpenInterpreter ▷ #O1 (2 messages):

O1 performance on weak laptops, O1 on Windows laptops, Windows 11 compatibility


OpenInterpreter ▷ #ai-content (3 messages):

New Product Launch, OpenAI Sora

Links mentioned:


LAION ▷ #general (9 messages🔥):

Spam Advertising Issues, German LLM Evaluation, Awareness of AI Capabilities


LAION ▷ #research (3 messages):

MagVit 2 for medical imaging, Memory-efficient optimizers for LLMs

Link mentioned: APOLLO: SGD-like Memory, AdamW-level Performance: Large language models (LLMs) are notoriously memory-intensive during training, particularly with the popular AdamW optimizer. This memory burden necessitates using more or higher-end GPUs or reducing ...


Axolotl AI ▷ #general (3 messages):

Shampoo Low Bit Implementation, Gradient Checkpointing Default Setting

Link mentioned: GitHub - axolotl-ai-cloud/axolotl at shampoo-low_bit: Go ahead and axolotl questions. Contribute to axolotl-ai-cloud/axolotl development by creating an account on GitHub.


Mozilla AI ▷ #announcements (1 messages):

Web Applets open standard, Graphical client-side apps, Live coding demos


AI21 Labs (Jamba) ▷ #general-chat (1 messages):

Dataoorts GPU Cloud




{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}