Frozen AI News archive

not much happened today

**Anthropic** released upgraded **Claude 3.5 Sonnet** and **Claude 3.5 Haiku** models featuring a new **computer use capability** that allows interaction with computer interfaces via screenshots and actions like mouse movement and typing. The **Claude 3.5 Sonnet** achieved state-of-the-art coding performance on SWE-bench Verified with a **49% score**, surpassing OpenAI's **o1-preview**. **Anthropic** focuses on teaching general computer skills rather than task-specific tools, with expected rapid improvements. Other releases include **Mochi 1**, an open-source video generation model, **Stable Diffusion 3.5** with Large and Medium variants, and **Embed 3** by **Cohere**, a multimodal embedding model for text and image search. **KerasHub** was launched by **François Chollet**, unifying KerasNLP and KerasCV with 37 pretrained models. Microsoft introduced the **Differential Transformer** to reduce attention noise via differential attention maps, and research on transformer attention layers was shared by **Rasbt**.

Canonical issue URL

AI News for 10/22/2024-10/23/2024. We checked 7 subreddits, 433 Twitters and 31 Discords (229 channels, and 3078 messages) for you. Estimated reading time saved (at 200wpm): 346 minutes. You can now tag @smol_ai for AINews discussions!

People are still very much exploring the implications of Anthropic's new Computer Use demo/usecases. Some are pointing out its failures and picking apart the terminology, others have hooked it up to phone simulators and real phones. Kyle Corbitt somehow wrote a full desktop app for Computer Use in 6 hours so you dont have to spin up the docker demo Anthropic shipped.

image.png

But not a single person in the room has any doubt that this will get a lot better very soon and very quickly.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

Anthropic's Claude 3.5 Release and Computer Use Capability

Other AI Model Releases and Updates

AI Research and Development

Industry Developments and Collaborations


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Major LLM Updates: Claude 3.5 and Stable Diffusion 3.5

Theme 2. Open Source AI Model Developments and Replication Efforts

Theme 3. AI Model Comparison Tools and Cost Optimization

Theme 4. GPU Hardware Discussions for AI Development

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

AI Model Releases and Improvements

AI Capabilities and Applications

AI Development and Research

AI Ethics and Security


AI Discord Recap

A summary of Summaries of Summaries by O1-preview

Theme 1: Claude 3.5 Takes the AI World by Storm

2. Innovative AI Applications in Creative and Practical Domains

Theme 3: Shiny New Tools Promise AI Advancement

Theme 4: AI's Dark Side Sparks Concern

Theme 5: ZK Proofs Give Users Control Over Their Data


PART 1: High level Discord summaries

HuggingFace Discord


OpenAI Discord


Unsloth AI (Daniel Han) Discord


Stability.ai (Stable Diffusion) Discord


aider (Paul Gauthier) Discord


OpenRouter (Alex Atallah) Discord


LM Studio Discord


Perplexity AI Discord


GPU MODE Discord


Nous Research AI Discord


Eleuther Discord


Latent Space Discord


LlamaIndex Discord


tinygrad (George Hotz) Discord


Interconnects (Nathan Lambert) Discord


OpenInterpreter Discord


Cohere Discord


Modular (Mojo 🔥) Discord


Torchtune Discord


LangChain AI Discord


DSPy Discord


LLM Agents (Berkeley MOOC) Discord


LLM Finetuning (Hamel + Dan) Discord


OpenAccess AI Collective (axolotl) Discord


Gorilla LLM (Berkeley Function Calling) Discord


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ▷ #general (547 messages🔥🔥🔥):

  • Stable Diffusion 3.5
  • Llama Models
  • Automating CAD Document Creation
  • Music Generation Models
  • Benchmarking AI Models

Links mentioned:


HuggingFace ▷ #today-im-learning (15 messages🔥):

  • MIT AI Course
  • Manim Animation Engine
  • Learning LLM Neural Networks
  • Statistics and Linear Algebra for AI

Links mentioned:


HuggingFace ▷ #cool-finds (18 messages🔥):

  • Quantum Computing Exploration
  • Vintern-3B-beta Model Development
  • Comparison Tool for LLM Services
  • Openblock ZK Proofs for ChatGPT
  • LLM License Clauses Criticism

Links mentioned:


HuggingFace ▷ #i-made-this (22 messages🔥):

  • ZK Proofs for ChatGPT
  • SNES Music Diffusion Model
  • Concerns about GitHub Organization
  • Fourier Dual Diffusion Repository

Links mentioned:


HuggingFace ▷ #NLP (3 messages):

  • Custom Model Tokenizers
  • Automating CAD File Creation

HuggingFace ▷ #diffusion-discussions (3 messages):

  • Gradio Questions
  • Channel Guidance

OpenAI ▷ #ai-discussions (300 messages🔥🔥):

  • Claude 3.5 Sonnet
  • Anthropic Computer Use Tool
  • Model Performance Improvements
  • AI Tokenization Issues
  • AI Systems and Optimization

Link mentioned: How to Keep Improving When You're Better Than Any Teacher - Iterated Distillation and Amplification: [2nd upload] AI systems can be trained using demonstrations from experts, but how do you train them to out-perform those experts? Can this still be done even...


OpenAI ▷ #gpt-4-discussions (9 messages🔥):

  • GPT-4 Upgrade Timeline
  • Caching Feature in GPT-4
  • ChatGPT Payment Issues

Link mentioned: Custom GPT's upgrade base model to omni?: Hi people! I’m realy wonderful with GPT 4o, really about fast request and reorder of topics. But i considere a problem to bring my custom GPT’s from the old model (i thinks thats are Gizmo model) to...


OpenAI ▷ #prompt-engineering (7 messages):

  • Spatial Sense in Models
  • Custom GPT Challenge
  • Function Call Context
  • Realtime API Performance

OpenAI ▷ #api-discussions (7 messages):

  • Spatial Sense Weakness
  • Not ChattGPT Custom GPT
  • Function Calls Context in Chat Completion
  • Realtime API Performance Comparison
  • Prompt Adaptation Suggestions

Unsloth AI (Daniel Han) ▷ #general (202 messages🔥🔥):

  • Claude 3.5 models
  • Kaggle & PyTorch issues
  • Fine-tuning challenges
  • Unsloth sloth symbolism
  • MMLU performance concerns

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (1 messages):

  • PhD Research Experience
  • Publication Rates in Academia
  • Comparative PhD Culture
  • AI/ML/CV Fields

Link mentioned: Reddit - Dive into anything: no description found


Unsloth AI (Daniel Han) ▷ #help (92 messages🔥🔥):

  • Multi-GPU Support
  • Unsloth Import Errors
  • Using Unsloth with Kaggle
  • Model Fine-Tuning Procedures
  • Gemma Templates

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (1 messages):

  • Fast Apply
  • Qwen2.5 Coder Model
  • Cursor's Blog Post
  • Performance Metrics

Link mentioned: Reddit - Dive into anything: no description found


Unsloth AI (Daniel Han) ▷ #community-collaboration (3 messages):

  • Studio Fix PR
  • Tokenizer Patch PR

Links mentioned:


Unsloth AI (Daniel Han) ▷ #research (1 messages):

edd0302: https://arxiv.org/pdf/2410.16663


Stability.ai (Stable Diffusion) ▷ #general-chat (294 messages🔥🔥):

  • Stable Diffusion 3.5
  • AI Generation Models
  • LoRA and VAE Handling
  • Image Prompting Techniques
  • Model Checkpoints and Sorting

Links mentioned:


aider (Paul Gauthier) ▷ #announcements (1 messages):

  • Aider v0.60.0
  • Sonnet 10/22 support
  • Improved code editing
  • Bugfixes
  • Model metadata management

aider (Paul Gauthier) ▷ #general (240 messages🔥🔥):

  • Claude 3.5 Sonnet
  • Aider Benchmarking
  • Fast Apply
  • Qwen 2.5 and Replete
  • OpenRouter and Rate Limits

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (45 messages🔥):

  • Claude 3.5 updates
  • Aider debugging tips
  • Mistral API authentication issues
  • Git commands and navigation
  • Repo map functionality

Links mentioned:


aider (Paul Gauthier) ▷ #links (1 messages):

  • DreamCut AI
  • Claude AI
  • Video Editing Software

Link mentioned: Tweet from Meng To (@MengTo): Introducing http://dreamcut.ai A video editor I built from scratch using Claude AI. This took 3 months and over 50k lines of code. I totally skipped design and went straight to code. Currently in e...


OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

  • Claude 3.5 Sonnet
  • Lumimaid v0.2
  • Magnum v4
  • Discounts on Models

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (272 messages🔥🔥):

  • OpenRouter Model Updates
  • API Key Usage
  • Prompt Caching
  • Tool Use in Models
  • Website Access Issues

Links mentioned:


OpenRouter (Alex Atallah) ▷ #beta-feedback (10 messages🔥):

  • Beta Access for Custom Provider Keys
  • Integrations Settings Access Requests

LM Studio ▷ #general (154 messages🔥🔥):

  • Downloading Models in LM Studio
  • Limitations of LLMs for Coding
  • Model Quantization Methods
  • Issues with System Prompts in Models
  • Using Vision Models for Image Captioning

Links mentioned:


LM Studio ▷ #hardware-discussion (13 messages🔥):

  • Ryzen AI Configuration
  • Intel Lunar Lake NPU Support
  • NPU Support in Llama.cpp
  • RX 7900 XTX vs RTX 3090 Comparison
  • Multi-GPU Support in ROCm

Links mentioned:


Perplexity AI ▷ #general (145 messages🔥🔥):

  • New Sonnet 3.5 Model
  • Perplexity Web Search Issues
  • User Experience with API Models
  • Feedback on Updates
  • Use of Custom Rag Systems

Links mentioned:


Perplexity AI ▷ #sharing (17 messages🔥):

  • Advanced AI-Driven Fact-Checking
  • Method of Loci Memory Trick
  • Component MRO in Aircraft Industry
  • Nvidia TSMC AI Alliance
  • Claude Computer Control in RPA

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (3 messages):

  • Account Credit Transfer Issues
  • Server Errors

GPU MODE ▷ #general (6 messages):

  • LLM Activations Quantization
  • bf16 vs fp32 Precision
  • Anyscale Inference Engine
  • CUDA Streams and Synchronization

Link mentioned: cuda-course/05_Writing_your_First_Kernels/05 Streams/01_stream_basics.cu at master · Infatoshi/cuda-course: Contribute to Infatoshi/cuda-course development by creating an account on GitHub.


GPU MODE ▷ #triton (47 messages🔥):

  • Triton kernels performance
  • Kernel caching strategies
  • Kernel launch overhead
  • Dynamic input shapes in Triton
  • Use of heuristics in Triton

Links mentioned:


GPU MODE ▷ #torch (7 messages):

  • torch.autocast usage with autoquant
  • Library linking with c10
  • Speeding up PyTorch models
  • Compiling PyTorch code
  • Autocast behavior with FP32 and BF16

Link mentioned: pytorch/aten/src/ATen/AccumulateType.h at b24c34426f47d6c311ad80ebba1d2575e6c7a6aa · pytorch/pytorch): Tensors and Dynamic neural networks in Python with strong GPU acceleration - pytorch/pytorch


GPU MODE ▷ #algorithms (1 messages):

.alphago: https://x.com/detectiveenters/status/1752067011113546234


GPU MODE ▷ #cool-links (3 messages):

  • Mapper Generation Automation
  • Weight Nowcaster Networks
  • Neuron Interaction and Nowcasting
  • Training Optimization
  • Graph Neural Networks

Links mentioned:


GPU MODE ▷ #beginner (3 messages):

  • CUDA Projects for Internships
  • Interactive Environments for CUDA Kernels

GPU MODE ▷ #pmpp-book (1 messages):

vim410: This whole chapter is getting rewritten. Once i have the new chapter will share it.


GPU MODE ▷ #torchao (7 messages):

  • int4_mm wrapping for torch.compile
  • torch.library.custom_op vs @register_meta
  • nightly wheels for non-Linux

Links mentioned:


GPU MODE ▷ #rocm (8 messages🔥):

  • ROCm vs ROCm/Triton/FA benchmarks
  • RCCL contributions
  • PyTorch symmetric memory
  • Matrix multiplication using Triton

GPU MODE ▷ #liger-kernel (8 messages🔥):

  • Monkey Patching CrossEntropy
  • Version Compatibility Challenges
  • Liger's Kernel Fusion
  • Inference Optimization Techniques
  • Gradient Accumulation Fix Discussion

Links mentioned:


GPU MODE ▷ #self-promotion (3 messages):

  • Triton kernels for quantization
  • NVIDIA Virtual Connect with Experts
  • cuDF and cuML
  • RAPIDS developers panel

Link mentioned: accelerated-computing-hub/connect-with-experts at main · NVIDIA/accelerated-computing-hub: NVIDIA curated collection of educational resources related to general purpose GPU programming. - NVIDIA/accelerated-computing-hub


GPU MODE ▷ #🍿 (54 messages🔥):

  • Creating an LLM for Efficient Kernels
  • CUDABench Proposal
  • Community Authorship and Contributions
  • Optimization Techniques for CUDA Kernels
  • Game Development Analogies in GPU Programming

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

  • Hermes 70B API
  • Revenue-sharing partnership
  • Hyperbolic launch
  • AI Inference Service

Link mentioned: Tweet from Hyperbolic (@hyperbolic_labs): Hyperbolic is making AI more accessible to everyone. Today, we’re launching a revenue-sharing partnership with @NousResearch, creators of the Hermes 3 large language model. Through our AI Inference S...


Nous Research AI ▷ #general (119 messages🔥🔥):

  • Nous Research Forge
  • Hermes AI Censorship
  • Claude Automation
  • AI Role-Playing
  • Grunt Work Opportunities

Links mentioned:


Nous Research AI ▷ #ask-about-llms (20 messages🔥):

  • Claude's new system prompt
  • Finetuning models like Gemma 2
  • Whisper-based translation frameworks
  • Llama model quantization
  • Support for JSON data parsing

Links mentioned:


Nous Research AI ▷ #research-papers (1 messages):

feffy: p-hacking :P


Nous Research AI ▷ #interesting-links (5 messages):

  • ZK Proofs and ChatGPT ownership
  • Advancements in ZK technology
  • Genmo AI motion quality
  • Prompt adherence in AI generation
  • Human action generation

Links mentioned:


Nous Research AI ▷ #research-papers (1 messages):

feffy: p-hacking :P


Eleuther ▷ #general (35 messages🔥):

  • Chess Move Explainability
  • Accusations of Cheating in Chess
  • LLMs Self-Explanation Accuracy
  • Molmo Vision Models
  • DINOv2 Understanding

Links mentioned:


Eleuther ▷ #research (45 messages🔥):

  • RoPE 2D Encoding
  • Building Open Source Datasets
  • Transformers and LayerNorms
  • Softmax Attention Adaptations

Links mentioned:


Eleuther ▷ #interpretability-general (2 messages):

  • AI-Driven Observability Interface
  • Interpretability Startups
  • Technical Demonstrations
  • Jacob Steinhardt
  • Sarah Schwettmann

Link mentioned: Transluce: no description found


Eleuther ▷ #lm-thunderdome (7 messages):

  • simple_evaluate tasks
  • evaluating small models
  • Pile tasks error

Latent Space ▷ #ai-general-chat (49 messages🔥):

  • Anthropic Mouse Generator
  • Ideogram Canvas Features
  • AI and Loneliness Epidemic
  • Speculative Decoding in vLLM
  • New Meeting Automation Tools

Links mentioned:


LlamaIndex ▷ #blog (4 messages):

  • Llama Impact Hackathon
  • Box AI integration
  • Multi-agent concierge system
  • Building LLM-powered web apps with LlamaIndex.TS

Link mentioned: Adapters: LlamaIndex: Learn how to use LlamaIndex with the Vercel AI SDK.


LlamaIndex ▷ #general (31 messages🔥):

  • Persistent Context in Workflows
  • Analyzing Long YouTube Lectures
  • Using Anthropic LLM
  • Progress Stats in SimpleDirectoryReader
  • LlamaIndex Workflow Compatibility with Redpanda

Link mentioned: Gemini 1.5 Pro 2M context window, code execution capabilities, and Gemma 2 are available today: no description found


tinygrad (George Hotz) ▷ #general (5 messages):

  • Tensor int64/uint64 support
  • SHA3 Implementation
  • Tensor.ones and numpy

tinygrad (George Hotz) ▷ #learn-tinygrad (29 messages🔥):

  • Action Chunking Transformers training
  • TinyJIT challenges
  • JIT function input/output requirements
  • Running faster kernels
  • Reverse engineering byte code

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (14 messages🔥):

  • Claude AI Experience
  • Continuous Pretraining Discussion
  • Character.AI User Tragedy
  • Safety Features Implementation
  • MIT Technology Review Podcast

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (1 messages):

xeophon.: https://milesbrundage.substack.com/p/why-im-leaving-openai-and-what-im


Interconnects (Nathan Lambert) ▷ #memes (5 messages):

  • Jeremy Howard's tweet
  • Tek's energy levels
  • Hermes 3 performance

Link mentioned: Tweet from Jeremy Howard (@jeremyphoward): CEO of Microsoft getting ratioed by an anime pfp account...


Interconnects (Nathan Lambert) ▷ #posts (11 messages🔥):

  • Anthropic's focus on B2B
  • Cost comparison between Anthropic and Simular
  • Automation and AI agents
  • Microsoft vs OpenAI in AI demonstrations

Links mentioned:


OpenInterpreter ▷ #general (29 messages🔥):

  • Screenpipe tool
  • Claude 3.5 model
  • Open Interpreter's development
  • AI integration on different OS
  • Efficient data extraction for AI

Links mentioned:


OpenInterpreter ▷ #ai-content (1 messages):

facelessman: https://youtu.be/VgJ0Cge99I0 -- Love this episode -- love these folks!!!


Cohere ▷ #discussions (8 messages🔥):

  • Multimodal Command model
  • Global connection
  • Aya Expanse
  • Complex bots

Cohere ▷ #questions (7 messages):

  • Cohere API trial and production keys
  • Using Command-Night in Portuguese
  • Feedback integration in fine-tuning LLM
  • Ollama Mistral performance issues

Links mentioned:


Cohere ▷ #api-discussions (9 messages🔥):

  • Cohere V2 API Errors
  • Finetuned Models Issues
  • Vercel AI SDK Integration
  • Tool Use and Function Calling

Links mentioned:


Cohere ▷ #projects (1 messages):

  • Agentic Builder Day
  • Cohere Models
  • AI Agent hackathon

Link mentioned: OpenSesame | Build Better AI Agents: OpenSesame simplifies the entire AI agent lifecycle, from building to evaluating. Our platform empowers businesses to easily create, share, and implement AI agents and detect hallucinations, making AI...


Modular (Mojo 🔥) ▷ #mojo (20 messages🔥):

  • Community Meeting Discussions
  • Serial Communication in Mojo
  • C/C++ Support in Mojo
  • LED Matrix Communication
  • Framework Laptop 16

Modular (Mojo 🔥) ▷ #max (3 messages):

  • C API for MAX Engine
  • Mojo's advantages
  • Graph Builder API in C

Torchtune ▷ #general (6 messages):

  • Debugging run command
  • Testing on multiple GPUs
  • Fine-tuning custom models
  • Environment comparison

Torchtune ▷ #dev (6 messages):

  • Linters and Pre-commit Hooks
  • CI Issues
  • Tokenizer Tests

LangChain AI ▷ #general (3 messages):

  • Developer Project Survey
  • Location in Manila
  • FunctionMessages and LLM Responses

Link mentioned: From Lightbulb to Launch: Developer Challenges in Project Realization : Hello fellow developers! Are you full of innovative ideas but struggle to bring them to life? You're not alone. We're conducting research to understand the challenges developers face when turn...


LangChain AI ▷ #share-your-work (3 messages):

  • AI Coding Assistant Study
  • AI-Powered Funding Tool
  • ApeBrains Trader Specialization

Links mentioned:


LangChain AI ▷ #tutorials (1 messages):

  • GeoGuessr AI Bot
  • Vision LLMs
  • LangChain
  • Multimodal AI

Link mentioned: Coding a Vision LLM Agent that plays GeoGuessr by itself (GPT-4o, Claude 3.5 and Gemini 1.5): How to code an AI bot that plays autonomously the GeoGuessr game using Multimodal Vision LLMs that take screenshots of the game with Python + LangGhain and c...


DSPy ▷ #show-and-tell (2 messages):

  • Advanced Workflow System
  • Upgrade Process

DSPy ▷ #general (5 messages):

  • DSPy Funding Potential
  • Synthetic Data Generation Metrics

LLM Agents (Berkeley MOOC) ▷ #mooc-questions (7 messages):

  • LLM Agents MOOC Signup
  • Hackathon Project Open Sourcing
  • Agents Development Tutorials

LLM Finetuning (Hamel + Dan) ▷ #general (2 messages):

  • Axolotl configurations
  • LangSmith Prompt Hub
  • Kaggle Solutions
  • Hugging Face datasets
  • Discord bot for message scraping

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (2 messages):

  • Experimental Triton FA support
  • Mixtral vs. Llama 3.2

Link mentioned: [Feature]: Memory Efficient Flash Attention for gfx1100 (7900xtx) · Issue #16 · ROCm/aotriton: Suggestion Description Started using torchlearn to train models in pytorch using my gfx1100 card but get a warning that 1toch was not compiled with memory efficient flash attention. I see there is ...


Gorilla LLM (Berkeley Function Calling) ▷ #discussion (2 messages):

  • Troubleshooting model evaluation
  • Handler registration
  • Model generation command






{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}