Frozen AI News archive

Llama 3.2: On-device 1B/3B, and Multimodal 11B/90B (with AI2 Molmo kicker)

**Meta** released **Llama 3.2** with new multimodal versions including **3B** and **20B** vision adapters on a frozen Llama 3.1, showing competitive performance against **Claude Haiku** and **GPT-4o-mini**. **AI2** launched multimodal **Molmo 72B** and **7B** models outperforming Llama 3.2 in vision tasks. Meta also introduced new **128k-context 1B and 3B models** competing with **Gemma 2** and **Phi 3.5**, with collaborations hinted with **Qualcomm**, **Mediatek**, and **Arm** for on-device AI. The release includes a **9 trillion token count** for Llama 1B and 3B. Partner launches include **Ollama**, **Together AI** offering free 11B model access, and **Fireworks AI**. Additionally, a new **RAG++ course** from **Weights & Biases**, **Cohere**, and **Weaviate** offers systematic evaluation and deployment guidance for retrieval-augmented generation systems based on extensive production experience.

Canonical issue URL

AI News for 9/24/2024-9/25/2024. We checked 7 subreddits, 433 Twitters and 31 Discords (223 channels, and 3218 messages) for you. Estimated reading time saved (at 200wpm): 316 minutes. You can now tag @smol_ai for AINews discussions!

Big news from Mira Murati and FB Reality Labs today, but the actual technical news you can use today is Llama 3.2:

image.png

As teased by Zuck and previewed in the Llama 3 paper (our coverage here), the Multimodal versions of Llama 3.2 released as anticipated, adding a 3B and a 20B vision adapter on a frozen Llama 3.1:

image.png

The 11B is comparable/slightly better than Claude Haiku, and the 90B is comparable/slightly better than GPT-4o-mini, though you will have to dig a lot harder to find out how far it trails behind 4o, 3.5 Sonnet, 1.5 Pro, and Qwen2-VL with a 60.3 on MMMU.

Meta is being praised for their open source here, but don't miss the multimodal Molmo 72B and 7B models from AI2 also releasing today. It has not escaped /r/localLlama's attention that Molmo is outperforming 3.2 in vision:

image.png

The bigger/pleasant/impressive surprise from Meta are the new 128k-context 1B and 3B models, which noew compete with Gemma 2 and Phi 3.5:

image.png

The release notes hint at some very tight on device collaborations with Qualcomm, Mediatek, and Arm:

The weights being released today are based on BFloat16 numerics. Our teams are actively exploring quantized variants that will run even faster, and we hope to share more on that soon.

Don't miss:

Partner launches:


This issue sponsored by RAG++: a new course from Weights & Biases. Go beyond RAG POCs and learn how to evaluate systematically, use hybrid search correctly and give your RAG system access to tool calling. Based on 18 months of running a customer support bot in production, industry experts at Weights & Biases, Cohere, and Weaviate show how to get to a deployment-grade RAG app. Includes free credits from Cohere to get you started!

image.png

Swyx commentary: Whoa, 74 lessons in 2 hours. I've worked on this kind of very tightly edited course content before and it's amazing that this is free! Chapters 1-2 cover some necessary RAG table stakes, but then it was delightful to see Chapter 3 teach important ETL and IR concepts, and learn some new things on cross encoding, rank fusion, and query translation in 4 and 5. We shall have to cover this on livestream soon!


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

Advanced Voice Model Release

The new voice model features lower latency, the ability to interrupt long responses, and support for memory to personalize responses. It also includes new voices and improved accents.

Google's Gemini 1.5 Pro and Flash Updates

Google announced significant updates to their Gemini models:

The models can now process 1000-page PDFs, 10K+ lines of code, and hour-long videos. Outputs are 5-20% shorter for efficiency, and safety filters are customizable by developers.

AI Model Performance and Benchmarks

AI Development and Research

AI Tools and Applications

Memes and Humor


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. High-Speed Inference Platforms: Cerebras and MLX

Theme 2. Qwen 2.5: Breakthrough Performance on Consumer Hardware

Theme 3. Gemini 1.5 Pro 002: Google's Latest Model Impresses

Theme 4. Apple Silicon vs NVIDIA GPUs for LLM Inference

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

AI Model Releases and Improvements

AI Capabilities and Benchmarks

AI Development Tools and Interfaces

AI Impact on Society and Work

Emerging AI Research and Applications


AI Discord Recap

A summary of Summaries of Summaries by O1-mini

Theme 1. New AI Model Releases and Multimodal Enhancements

Theme 2. Model Performance, Quantization, and Optimization

Theme 3. API Pricing, Integration, and Deployment Challenges

Theme 4. AI Safety, Censorship, and Licensing Issues

Theme 5. Hardware Infrastructure and GPU Optimization for AI


PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord


HuggingFace Discord


Nous Research AI Discord


aider (Paul Gauthier) Discord


OpenRouter (Alex Atallah) Discord


OpenAI Discord


LM Studio Discord


Eleuther Discord


Perplexity AI Discord


Interconnects (Nathan Lambert) Discord


GPU MODE Discord


OpenAccess AI Collective (axolotl) Discord


Cohere Discord


LlamaIndex Discord


Latent Space Discord


Stability.ai (Stable Diffusion) Discord


Torchtune Discord


Modular (Mojo 🔥) Discord


OpenInterpreter Discord


LLM Agents (Berkeley MOOC) Discord


DSPy Discord


tinygrad (George Hotz) Discord


LangChain AI Discord


LAION Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Finetuning (Hamel + Dan) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Gorilla LLM (Berkeley Function Calling) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Unsloth AI (Daniel Han) ▷ #general (510 messages🔥🔥🔥):

  • Llama 3.2 Release
  • Fine-tuning Models
  • Model Performance Comparisons
  • Vision Models
  • Model Inference and Compatibility

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (42 messages🔥):

  • Model Costs Comparison
  • API vs Local GPU Usage
  • OpenAI's Corporate Transition
  • Token Usage in Generation

Link mentioned: Tweet from Mira Murati (@miramurati): I shared the following note with the OpenAI team today.


Unsloth AI (Daniel Han) ▷ #help (52 messages🔥):

  • KTO and ORPO methods
  • OpenAI's Feedback Mechanism
  • Llama fine-tuning inquiries
  • Issues with Llama model
  • Spider Dataset and SQL Assistants

Links mentioned:


HuggingFace ▷ #announcements (1 messages):

  • New Model Releases
  • Gradio 5 Launch
  • FinePersonas Data Set
  • HF Hub Google Cloud Integration
  • Wikimedia Dataset Release

Links mentioned:


HuggingFace ▷ #general (432 messages🔥🔥🔥):

  • Llama 3.2 release
  • vLLM memory management
  • Quantization techniques
  • Hugging Face libraries
  • Machine learning optimization

Links mentioned:


HuggingFace ▷ #today-im-learning (7 messages):

  • Fine-tuning LLMs
  • Stable Diffusion Models
  • Diffusion Models from Scratch
  • Optimizing Embeddings with PCA

HuggingFace ▷ #cool-finds (12 messages🔥):

  • Comic Sans FLUX Model
  • Neural Network Learning Constraints
  • Digital Processing of Postal Codes
  • DHL Manual Sorting
  • Deutsche Post Operations

Links mentioned:


HuggingFace ▷ #i-made-this (125 messages🔥🔥):

  • Google Gemini demo
  • Zeromagic API platform
  • Transcription tools
  • PCA optimization project
  • Game development and quantum programming

Links mentioned:


HuggingFace ▷ #computer-vision (3 messages):

  • Molmo models
  • Multimodal comparison
  • Graphs and resources

Link mentioned: no title found: no description found


HuggingFace ▷ #NLP (12 messages🔥):

  • Training Topic Clusters
  • Setfit Serialization Issue
  • vLLM Deployment for NER
  • Fine-tuning Token Embeddings

Link mentioned: dslim/bert-base-NER · Hugging Face: no description found


HuggingFace ▷ #diffusion-discussions (4 messages):

  • Google Colab
  • Diffusion Models
  • Flux Model
  • SDXL Lightning
  • Würstchen Model

Nous Research AI ▷ #announcements (1 messages):

  • Hermes 3
  • Llama-3.1 8B
  • HuggingChat

Link mentioned: HuggingChat: Making the community's best AI chat models available to everyone.


Nous Research AI ▷ #general (214 messages🔥🔥):

  • Llama 3.2 Model Release
  • Quantization vs. Distillation
  • New Innovations in AI Training
  • Hermes 3 Update
  • Localization of Models

Links mentioned:


Nous Research AI ▷ #ask-about-llms (25 messages🔥):

  • Sample Packing Techniques
  • Tokenizers in Model Training
  • LLaMA 3.2 Vision Encoder

Nous Research AI ▷ #research-papers (2 messages):

  • Character Video Synthesis
  • Resume ATS Builder
  • Job Recommendation Systems

Link mentioned: Paper page - MIMO: Controllable Character Video Synthesis with Spatial Decomposed Modeling: no description found


Nous Research AI ▷ #interesting-links (2 messages):

  • Opus Insight updates
  • Sonnet 3.5 review process
  • ThinkLab exploration
  • O1-preview functionality

Links mentioned:


Nous Research AI ▷ #research-papers (2 messages):

  • Character video synthesis
  • Resume ATS builder
  • Job recommendation systems

Link mentioned: Paper page - MIMO: Controllable Character Video Synthesis with Spatial Decomposed Modeling: no description found


aider (Paul Gauthier) ▷ #general (95 messages🔥🔥):

  • Llama 3.2 Release
  • Aider Functionality
  • Vector Databases
  • Meta Connect
  • Sonnet 3.5 Errors

Links mentioned:

Then stay with us for the... | By Meta for DevelopersFacebook: Join Mark Zuckerberg as he shares Meta’s vision for AI and the metaverse, including Meta’s newest product announcements. Then stay with us for the.../llms.txt—a proposal to provide information to help LLMs use websites – Answer.AI: We propose that those interested in providing LLM-friendly content add a /llms.txt file to their site. This is a markdown file that provides brief background information and guidance, along with links...no title found: no description foundReddit - Dive into anything: no description foundElevated Errors on Claude 3.5 Sonnet: no description foundGitHub - mendableai/firecrawl: 🔥 Turn entire websites into LLM-ready markdown or structured data. Scrape, crawl and extract with a single API.: 🔥 Turn entire websites into LLM-ready markdown or structured data. Scrape, crawl and extract with a single API. - mendableai/firecrawl


aider (Paul Gauthier) ▷ #questions-and-tips (104 messages🔥🔥):

  • Aider Model Issues
  • Using Aider with PDFs
  • Testing Aider with Various Models
  • Aider's Token Usage
  • Integrating External Libraries and Analyzers

Links mentioned:


aider (Paul Gauthier) ▷ #links (4 messages):

  • par_scrape tool
  • shot-scraper utility
  • Llama 3.2 Multimodal
  • o1-preview bug fixing

Links mentioned:


OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

  • Database Upgrade
  • API Completion Response Changes
  • Gemini Model Updates
  • New Vision Language Models
  • Cohere Models Discount

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (169 messages🔥🔥):

  • GPTs Agents Performance
  • OpenRouter Model Availability
  • Mistral Image Recognition Issues
  • Llama 3.2 Release
  • OpenRouter API Rate Limits

Links mentioned:


OpenRouter (Alex Atallah) ▷ #beta-feedback (2 messages):

  • Local Server Support
  • API Access Conditions

OpenAI ▷ #ai-discussions (110 messages🔥🔥):

  • Advanced Voice Mode Availability
  • Meta AI Restrictions
  • Llama 3.2 Licensing
  • Game Development with AI
  • AI Character Interactions

Links mentioned:


OpenAI ▷ #gpt-4-discussions (8 messages🔥):

  • Voice Functionality with GPT
  • Complex Tasks Handling by GPT
  • Comparative Efficiency of AI Models

OpenAI ▷ #prompt-engineering (24 messages🔥):

  • Minecraft API prompting
  • Feedback mechanisms in AI
  • Essay grading challenges
  • Model output variability
  • Prompt engineering best practices

OpenAI ▷ #api-discussions (24 messages🔥):

  • Prompt Engineering for Minecraft Questions
  • Challenges with Feedback Models
  • Fine-Tuning Models for Essay Feedback

LM Studio ▷ #general (145 messages🔥🔥):

  • Llama 3.2 Models
  • SillyTavern Integration
  • Multimodal Capabilities
  • Model Performance Benchmarks
  • Support for New Models

Links mentioned:


LM Studio ▷ #hardware-discussion (17 messages🔥):

  • Price Discrepancies in Tech
  • Dual GPU Setup in LM Studio
  • Performance of RTX 3090
  • Vulkan and Mixed GPU Types

Links mentioned:


Eleuther ▷ #general (19 messages🔥):

  • yt-dlp tool
  • PyTorch training attribute issue
  • Local LLM benchmark suite
  • Foundation model engineer hiring
  • Controlling diffusion models

Links mentioned:


Eleuther ▷ #research (70 messages🔥🔥):

  • NAIV3 Technical Report
  • BERT Masking Rates
  • Model Training Techniques
  • Crowd-Sourced Datasets
  • SUNDAE Model Insights

Links mentioned:


Eleuther ▷ #scaling-laws (9 messages🔥):

  • Decomposing models
  • Scaling laws datapoints
  • Broken neural scaling laws dataset
  • Google research findings

Link mentioned: google-research/revisiting_neural_scaling_laws/README.md at master · google-research/google-research: Google Research. Contribute to google-research/google-research development by creating an account on GitHub.


Eleuther ▷ #interpretability-general (1 messages):

  • Chain-of-thought responses
  • Filler tokens efficacy
  • Human-like task decomposition
  • Transformers and algorithmic tasks

Link mentioned: Let's Think Dot by Dot: Hidden Computation in Transformer Language Models: Chain-of-thought responses from language models improve performance across most benchmarks. However, it remains unclear to what extent these performance gains can be attributed to human-like task deco...


Eleuther ▷ #lm-thunderdome (40 messages🔥):

  • Pile model formatting issues
  • lm_eval usage with OpenAI API
  • Aexams task evaluation
  • Exact match metrics
  • Debugging sequence length errors

Links mentioned:


Eleuther ▷ #multimodal-general (1 messages):

  • Non-casual masking
  • AR masks and positional information

Perplexity AI ▷ #general (123 messages🔥🔥):

  • Perplexity AI updates
  • ChatGPT voice mode capabilities
  • O1 preview access
  • Issues with context retention
  • Usage of AI tools for school

Links mentioned:


Perplexity AI ▷ #sharing (6 messages):

  • Air Fryers Worth It
  • 4-Day Workout Plan
  • Jony Ive's OpenAI Project
  • Antarctica's Ozone Hole
  • Fashion Industry Research

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (9 messages🔥):

  • Wolfram Alpha integration
  • Perplexity API capabilities
  • Webhook Zap structure
  • Scripting web interface usage
  • External models availability

Interconnects (Nathan Lambert) ▷ #news (32 messages🔥):

  • Anthropic Financials
  • OpenAI Training Data Access
  • Molmo Model Performance
  • Molmo Model Capabilities
  • Molmo License Changes

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (15 messages🔥):

  • Twitter and Blogging
  • OpenAI's Recent Announcements
  • Hugging Face Collections
  • Late Fusion in Visual Language Models
  • Playing with GPT-4

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (24 messages🔥):

  • OpenAI Team Changes
  • Drama in AI Companies
  • Anthropic's Position
  • Investor Confidence
  • Latest OpenAI News

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (22 messages🔥):

  • UV Python Tool Enhancement
  • FTC AI Crackdown Announcement
  • Data Dependencies in Scripts
  • Docker Integration with UV
  • LLaMA Stack Discussion

Links mentioned:


Interconnects (Nathan Lambert) ▷ #memes (8 messages🔥):

  • Artificial General Intelligence (AGI)
  • Social Media Impacts
  • Xeophon Quotes
  • Transformative Opportunities
  • Public Reactions

Links mentioned:


Interconnects (Nathan Lambert) ▷ #rl (2 messages):

  • Reinforcement Learning
  • Curriculum Learning
  • Demonstration Data
  • Simulation in Robotics

Link mentioned: Reverse Forward Curriculum Learning for Extreme Sample and Demonstration Efficiency in Reinforcement Learning: Reinforcement learning (RL) presents a promising framework to learn policies through environment interaction, but often requires an infeasible amount of interaction data to solve complex tasks from sp...


Interconnects (Nathan Lambert) ▷ #posts (25 messages🔥):

  • Llama 3.2 Release
  • Multimodal Models
  • Molmo Model Performance
  • Meta's Edge Device Integration
  • Community Reactions

Links mentioned:


GPU MODE ▷ #torch (22 messages🔥):

  • Diffusion Transformer Policy for Imitation Learning
  • Torch Profiler File Size Issues
  • Pytorch Kernel Performance Improvement
  • Understanding torch.float8_e4m3fn
  • Embedding Data Loading from GPU

Links mentioned:


GPU MODE ▷ #cool-links (17 messages🔥):

  • New Course Release
  • Pranav Marla's O1 Alternative
  • NVIDIA ASR Model Optimization
  • Llama 3.2 Announcement
  • Local AI Model Performance

Links mentioned:


GPU MODE ▷ #jobs (2 messages):

  • Luma Engineering Roles
  • Meta Community Projects

Link mentioned: Software Engineer, Systems ML - PyTorch Performance and Engagement: Meta's mission is to give people the power to build community and bring the world closer together. Together, we can help people build stronger communities - join us.


GPU MODE ▷ #beginner (12 messages🔥):

  • CUDA code in Python
  • Custom Ops in PyTorch
  • Beginner Projects

Link mentioned: ao/torchao/csrc at main · pytorch/ao: PyTorch native quantization and sparsity for training and inference - pytorch/ao


GPU MODE ▷ #torchao (1 messages):

  • RoPE cache in FP32
  • Torchao Llama

Link mentioned: ao/torchao/_models/llama/model.py at 7dff17a0e6880cdbeed1a14f92846fac33717b75 · pytorch/ao: PyTorch native quantization and sparsity for training and inference - pytorch/ao


GPU MODE ▷ #off-topic (16 messages🔥):

  • GPU Mode Name Ideas
  • Lambda Labs Credits
  • Mascot Suggestions

GPU MODE ▷ #triton-puzzles (3 messages):

  • CUDA puzzles
  • Machine Learning in PMP book

Link mentioned: GitHub - srush/GPU-Puzzles: Solve puzzles. Learn CUDA.: Solve puzzles. Learn CUDA. Contribute to srush/GPU-Puzzles development by creating an account on GitHub.


GPU MODE ▷ #llmdotc (13 messages🔥):

  • RoPE implementation
  • Fused RMSNorm integration
  • Llama 3.1 and 3.2 support
  • CUDA kernel development

Links mentioned:


GPU MODE ▷ #rocm (6 messages):

  • MI250x
  • GroupedGemm examples
  • AMD Instinct™ GPU Training
  • Architecture Development

Link mentioned: AMD Instinct™ GPU Training: no description found


GPU MODE ▷ #bitnet (2 messages):

  • TorchAO PR for BitNet
  • Optimized Inference Kernel
  • Model Training Plans

Link mentioned: BitNet b1.58 training by gau-nernst · Pull Request #930 · pytorch/ao: This PR adds training code for BitNet b1.58 (ternary weights - 1.58 bit. The first version of BitNet is binary weights). This is implemented as tensor subclass and integrate nicely with the quantiz...


GPU MODE ▷ #arm (4 messages):

  • iPhone 16 performance
  • SME on A18 chip
  • Metal benchmarks on GitHub

Link mentioned: GitHub - philipturner/metal-benchmarks: Apple GPU microarchitecture: Apple GPU microarchitecture. Contribute to philipturner/metal-benchmarks development by creating an account on GitHub.


GPU MODE ▷ #webgpu (1 messages):

  • Compute Pass Pipeline Behavior
  • Workgroup Dispatch Visibility

GPU MODE ▷ #liger-kernel (23 messages🔥):

  • Jensen-Shannon Divergence Benchmark
  • Lambda Labs Usage
  • SyncBatchNorm Implementation

Links mentioned:


GPU MODE ▷ #metal (2 messages):

  • Metal Atomics
  • Memory Barrier Semantics
  • Workgroup Scheduling

GPU MODE ▷ #self-promotion (1 messages):

  • GPU Performance Optimization
  • Deep Learning Applications with GPUs
  • NVIDIA Nsight
  • GPU Programming Languages

Link mentioned: An Introduction to GPU Performance Optimization for Deep Learning | DigitalOcean: no description found


OpenAccess AI Collective (axolotl) ▷ #general (82 messages🔥🔥):

  • Run Pod Issues
  • Molmo 72B Announcement
  • OpenAI's Recent Developments
  • Llama 3.2 Release
  • Meta and EU Controversy

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (9 messages🔥):

  • Axolotl with Metaflow
  • Using rope_theta
  • Tool calling for fine-tuning

OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (5 messages):

  • Illegal CUDA Memory Access
  • CUDA Version Compatibility
  • NVIDIA Driver Updates
  • Compute Sanitizer Usage
  • Error Checking in CUDA

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


Cohere ▷ #discussions (52 messages🔥):

  • Recursive Iterative Models
  • Multi-step Tool Applications
  • RAG Models
  • Cohere Course on RAG
  • Using Cohere Models in Projects

Links mentioned:


Cohere ▷ #questions (6 messages):

  • Cohere API Key Pricing
  • Citations for Unstructured Text Data
  • Rerank Fine-tuning Expectations

Links mentioned:


Cohere ▷ #api-discussions (15 messages🔥):

  • NotFoundError with multilingual Embed
  • 404 error on documentation link
  • Switching to newer Embed model
  • Response from Cohere Support

Cohere ▷ #projects (11 messages🔥):

  • Cohere in Embedded Systems
  • Smart Telescope Mount/Tracker
  • Cohere Cookbook
  • Embedding Use Cases
  • GitHub Notebooks

Links mentioned:


LlamaIndex ▷ #announcements (1 messages):

  • LlamaParse Fraud Alert
  • LlamaIndex Product Clarification

LlamaIndex ▷ #blog (5 messages):

  • LlamaParse Scam Warning
  • AWS Gen AI Loft Event
  • Advanced RAG over Excel
  • Pixtral 12B Model Launch
  • LlamaIndex Hiring Announcement

Link mentioned: LlamaCloud: no description found


LlamaIndex ▷ #general (74 messages🔥🔥):

  • Using LlamaIndex with Ollama
  • Accessing VectorStoreIndex
  • Handling LlamaTrace Project Errors
  • Error Resolution in Notebooks
  • Passing Messages to ReAct Agent

Links mentioned:


Latent Space ▷ #ai-general-chat (76 messages🔥🔥):

  • Gemini Pricing Strategies
  • Anthropic Revenue Breakdown
  • Llama 3.2 Launch
  • Mira Murati Departure
  • Meta's Orion Glasses

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

swyxio: new meetup in <@&1284244976024424630> led by <@656968717883670570> ! https://lu.ma/i8ulstlw


Stability.ai (Stable Diffusion) ▷ #general-chat (66 messages🔥🔥):

  • Installing Basicsr
  • ComfyUI vs Forge
  • 3D Model Generators
  • Stable Diffusion Usage
  • ControlNet OpenPose

Torchtune ▷ #announcements (1 messages):

  • Llama 3.2 Release
  • Multimodal Support
  • Long-context Models
  • Finetuning Options

Torchtune ▷ #general (29 messages🔥):

  • Green Card Concerns
  • Pen and Paper Return
  • Llama 3.2 Access
  • MetaAI Restrictions
  • Legal Department Reactions

Torchtune ▷ #dev (9 messages🔥):

  • TF32 usage in PyTorch
  • KV-cache toggling RFC
  • Handling Tensor sizing

Links mentioned:


Modular (Mojo 🔥) ▷ #general (4 messages):

  • MOToMGP error
  • Linux box sizing for MAX
  • Mojo birthday celebrations

Modular (Mojo 🔥) ▷ #announcements (1 messages):

  • Mojo GitHub Discussions
  • Community Q&A on Magic

Link mentioned: Community Magic Questions: Please share your questions related to Magic! Zac will answer them during the Magic Q&A in our community meeting on September 30th. As always, the recording will be posted to YouTube.


Modular (Mojo 🔥) ▷ #mojo (21 messages🔥):

  • Mojo and C communication speed
  • Implementation of associated aliases
  • Potential for chainable iterables in Mojo
  • Ownership mechanisms in Mojo
  • UnsafeMaybeUninitialized wrapper

Link mentioned: [stdlib] Use ownership ops in Uninit to remove pointer indirection. by helehex · Pull Request #3453 · modularml/mojo: The ownership ops provide a more clear implementation than pop.array, and allows the pointer indirection to be removed. This makes the inner field of UnsafeMaybeUninitialized just a ElementType ins...


OpenInterpreter ▷ #general (12 messages🔥):

  • o1 Preview and Mini API access
  • Llama 3.2 experiment results
  • Design choices for logos
  • Broken Open Interpreter

OpenInterpreter ▷ #ai-content (5 messages):

  • Llama 3.2 Release
  • Tool Use YouTube Episode
  • GroqCloud Availability

Links mentioned:


LLM Agents (Berkeley MOOC) ▷ #mooc-questions (2 messages):

  • Quiz 3 Information

LLM Agents (Berkeley MOOC) ▷ #mooc-readings-discussion (13 messages🔥):

  • Gemini 1.5 Flash performance
  • New MMLU-Pro dataset
  • Chain of Thought effectiveness
  • AutoGen usage in research

Links mentioned:


DSPy ▷ #show-and-tell (2 messages):

  • Langtrace Updates
  • DSPy Features
  • Automatic Experiment Tracking
  • Checkpoint State Tracking

Link mentioned: Tweet from Karthik Kalyanaraman (@karthikkalyan90): Some DSPy specific native features on @langtrace_ai launching this week. - New project type - DSPy - Automatic experiment tracking (inspired from MLFlow) - Automatic checkpoint state tracking - Eval ...


DSPy ▷ #general (5 messages):

  • DSPy text classification
  • Claude models
  • DSPy for conversational agents
  • Orchestrating user queries
  • DSPy memory and conversation history

DSPy ▷ #examples (3 messages):

  • Text Classification
  • Complex Classes in Classification
  • Tutorial on Classification Tasks

tinygrad (George Hotz) ▷ #general (9 messages🔥):

  • tinygrad resources
  • debugging tinygrad
  • contributing to tinygrad

Link mentioned: Tutorials on Tinygrad: Tutorials on tinygrad


tinygrad (George Hotz) ▷ #learn-tinygrad (1 messages):

  • Training speed issues
  • Sampling code bug
  • Model output quality

LangChain AI ▷ #general (8 messages🔥):

  • Open Source Chat UI for Programming Tasks
  • Thumbs Up/Down Review Options for Chatbots
  • Azure Chat OpenAI Integration

LangChain AI ▷ #share-your-work (1 messages):

  • Agentic RAG application
  • LangGraph
  • Lightning Studios
  • Streamlit
  • Research Assistant

Link mentioned: LangGraph-AgenticRAG with Streamlit - a Lightning Studio by maxidiazbattan: This studio provides a guide on integrating and utilizing an Agentic-RAG, with Langgraph, Ollama and Streamlit.


LAION ▷ #general (1 messages):

  • GANs
  • CNNs
  • ViTs
  • Image tasks
  • Algorithm comparison

LAION ▷ #research (7 messages):

  • MaskBit image generation
  • MonoFormer multimodal transformer
  • Sliding window attention
  • VQGAN modernization
  • Embedding-free generation







{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}