Frozen AI News archive

not much happened today + AINews Podcast?

**Glean** doubled its valuation again. **Dan Hendrycks' Superforecaster AI** generates plausible election forecasts with interesting prompt engineering. A **Stanford** study found that **LLM-generated research ideas** are statistically more novel than those by expert humans. **SambaNova** announced faster inference for **llama-3** models, surpassing **Cerebras**. **Benjamin Clavie** gave a notable talk on retrieval-augmented generation techniques. **Strawberry** is reported to launch in two weeks. **Google Illuminate** offers AI-generated podcast discussions about papers and books. **Apple** unveiled new AI features in iOS 18, including visual intelligence and improved Siri, with on-device and cloud processing for camera-based event additions. The **Reflection 70B** model sparked controversy over performance claims. Experts highlighted the unreliability of traditional benchmarks like MMLU and HumanEval, recommending alternative evaluation methods such as LMSys Chatbot Arena and Hugging Face's open-sourced **Lighteval** suite. The AI research community continues to explore AI's role in generating novel research ideas and improving benchmarking.

Canonical issue URL

AI News for 9/9/2024-9/10/2024. We checked 7 subreddits, 433 Twitters and 30 Discords (215 channels, and 2311 messages) for you. Estimated reading time saved (at 200wpm): 247 minutes. You can now tag @smol_ai for AINews discussions!

Let's see:

Yesterday, folks were also excited about Google Illuminate, AI generated podcast discussions about papers and books. It is gated behind a waitlist, but we at Smol AI are exploring doing the same. Check out our first attempt here!


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

Apple's AI Announcements and Industry Reactions

AI Model Developments and Controversies

AI in Research and Innovation

AI Tools and Applications

AI Ethics and Safety

Memes and Humor


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Reflection 70B: From Hype to Controversy

Theme 2. AMD's UDNA: Unifying RDNA and CDNA to Challenge CUDA

Theme 3. DeepSeek V2.5: Quietly Released Powerhouse Model

Theme 4. Innovative Approaches to Model Efficiency and Deployment

Theme 5. Advancements in Specialized AI Models and Techniques

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

AI Model Releases and Improvements

AI Tools and Interfaces

AI Ethics and Societal Impact

Humor and Memes


AI Discord Recap

A summary of Summaries of Summaries by Claude 3.5 Sonnet

1. AI Model Releases and Benchmarks

2. LLM Fine-tuning and Optimization Techniques

3. Open Source AI Developments and Collaborations

4. Multimodal AI and Tool Integrations

GPT4O (gpt-4o-2024-05-13)

1. DeepSeek 2.5 Launch

2. Model Fine-Tuning Challenges

3. Hardware and Model Performance

4. AI Model Innovations

5. Open Source AI Developments


PART 1: High level Discord summaries

HuggingFace Discord


Unsloth AI (Daniel Han) Discord


LM Studio Discord


OpenAI Discord


OpenRouter (Alex Atallah) Discord


CUDA MODE Discord


Cohere Discord


OpenInterpreter Discord


Modular (Mojo 🔥) Discord


Nous Research AI Discord


Torchtune Discord


Perplexity AI Discord


Latent Space Discord


LlamaIndex Discord


Interconnects (Nathan Lambert) Discord


Stability.ai (Stable Diffusion) Discord


LAION Discord


OpenAccess AI Collective (axolotl) Discord


LangChain AI Discord


DSPy Discord


LLM Finetuning (Hamel + Dan) Discord


Gorilla LLM (Berkeley Function Calling) Discord


tinygrad (George Hotz) Discord


MLOps @Chipro Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ▷ #announcements (1 messages):

  • DeepSeek 2.5
  • Yi Coder 1.5B+9B
  • OLMoE
  • Multi-agent systems support
  • Semantic Dataset Search

Links mentioned:


HuggingFace ▷ #general (455 messages🔥🔥🔥):

  • Whisper Model Usage
  • Korean Lemmatizer Development
  • Model Structured Output
  • Quantization and Dataset Calibration
  • Hugging Face Community Dynamics

Links mentioned:


HuggingFace ▷ #today-im-learning (2 messages):

  • OpenSSL 3.3.2
  • Post Quantum Cryptography
  • TLS Handshakes

HuggingFace ▷ #cool-finds (1 messages):

cakiki: Is it open?


HuggingFace ▷ #i-made-this (21 messages🔥):

  • Synthetic Data Creation with GANs
  • Quantized GraphRAG Systems
  • Local-First Vector Database
  • Resume Roaster Project
  • LLM Responses and Formatting

Links mentioned:


HuggingFace ▷ #reading-group (1 messages):

  • Instruction-tuned Models
  • DPO/RLHF-tuning
  • LLaMA 3.1
  • Fine-tuning Guardrails

HuggingFace ▷ #NLP (12 messages🔥):

  • PDF Document Analysis
  • ColPali Embeddings Issue
  • Amazon ML Challenge 2023
  • Korean Lemmatizer with AI
  • Building NLP Models from Scratch

Links mentioned:


HuggingFace ▷ #diffusion-discussions (5 messages):

  • Diffusers and Latent Space Manipulation
  • Image-to-Image Generation with Diffusers
  • Using CLIP Text Embeddings
  • Denoising Latent Images

Link mentioned: Image-to-image: no description found


Unsloth AI (Daniel Han) ▷ #general (333 messages🔥🔥):

  • Model Fine-Tuning
  • MLC Deployment Issues
  • Unsloth Updates
  • Inference Problems
  • Llama-3.1-SuperNova-Lite

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (7 messages):

  • Kaggle Housing Price Challenge
  • Unsloth Fine-tuned Model Deployment
  • MOE Model Performance

Link mentioned: House Prices - Advanced Regression Techniques | Kaggle: no description found


Unsloth AI (Daniel Han) ▷ #help (27 messages🔥):

  • Full Fine-Tuning Inquiry
  • Loss Spiking Issue
  • Flash Attention 2 Usage
  • Optimal GPU Size for LLAMA 3.1
  • Metric Computation Support in SFTTrainer

Unsloth AI (Daniel Han) ▷ #research (9 messages🔥):

  • WizardMath fine-tuning
  • Collaboration on RAG
  • Experience in machine learning
  • Mechanical engineering background

LM Studio ▷ #general (81 messages🔥🔥):

  • Model Training Parameters
  • Multi-GPU Support in LM Studio
  • Availability of Older Versions
  • Optimal Models for Running AI
  • Performance on Limited Hardware

Links mentioned:


LM Studio ▷ #hardware-discussion (93 messages🔥🔥):

  • GPU capabilities
  • AMD vs NVIDIA performance
  • Mistral model operations
  • Surface Studio Pro upgrades
  • Building custom models

Link mentioned: Nvidia bans using translation layers for CUDA software — previously the prohibition was only listed in the online EULA, now included in installed files [Updated]: Translators in the crosshairs.


OpenAI ▷ #ai-discussions (89 messages🔥🔥):

  • Apple Silicon GPU capabilities
  • Gemini model functionalities
  • Llama 3 and free models
  • Video analysis AI projects
  • OpenCV limitations

Links mentioned:


OpenAI ▷ #gpt-4-discussions (8 messages🔥):

  • Driver's Bro GPT
  • Voice features in GPTs
  • Memory feature feedback
  • Using DALLE-3
  • Image creation through ChatGPT

OpenAI ▷ #prompt-engineering (13 messages🔥):

  • Stock Evaluation
  • Universal Evaluator Prompt
  • Accessing Prompt Library

OpenAI ▷ #api-discussions (13 messages🔥):

  • Using OAI Models for Stock Analysis
  • Universal Evaluator Prompt Persona
  • Accessing Prompt Library

OpenRouter (Alex Atallah) ▷ #announcements (1 messages):

  • Hermes 3 transition
  • Paid model announcement

Link mentioned: Hermes 3 405B Instruct (free) - API, Providers, Stats: Hermes 3 is a generalist language model with many improvements over Hermes 2, including advanced agentic capabilities, much better roleplaying, reasoning, multi-turn conversation, long context coheren...


OpenRouter (Alex Atallah) ▷ #app-showcase (2 messages):

  • Eggu Dataset
  • Open Source Multilingual Models
  • Cost of Usage

OpenRouter (Alex Atallah) ▷ #general (102 messages🔥🔥):

  • DeepSeek Models and Performance
  • Google Gemini Flash Rate Limits
  • Sonnet 3.5 Beta Issues
  • Costs of Hermes 3 and Llama 3 Models
  • AI Programming Tools Explore

Links mentioned:


CUDA MODE ▷ #general (5 messages):

  • Opus API Integration
  • Model Uploading Challenges

CUDA MODE ▷ #triton (6 messages):

  • Batch Performance Optimization
  • Triton Atomic Operations
  • Triton Compilation Process

CUDA MODE ▷ #torch (7 messages):

  • PyTorch Autotuning
  • Triton Autotuner
  • Functional Optimizer in PyTorch
  • Open Source Models Adaptation
  • Tulu Project Announcement

Link mentioned: ml-sigmoid-attention/optorch at main · apple/ml-sigmoid-attention: Contribute to apple/ml-sigmoid-attention development by creating an account on GitHub.


CUDA MODE ▷ #algorithms (24 messages🔥):

  • Sigmoid Attention Paper
  • FlashSigmoid vs FA3
  • Bias in Sigmoid Attention
  • Elementwise Sigmoid vs Rowwise Softmax
  • LayerScale

Link mentioned: Theory, Analysis, and Best Practices for Sigmoid Self-Attention: Attention is a key part of the transformer architecture. It is a sequence-to-sequence mapping that transforms each sequence element into a weighted sum of values. The weights are typically obtained as...


CUDA MODE ▷ #beginner (5 messages):

  • Tiling concept for matrix multiplication
  • Pragma unroll usage
  • Matrix multiplication resources

CUDA MODE ▷ #pmpp-book (3 messages):

  • Tiling concept for matrix multiplication
  • Matrix multiplication optimization resources

Link mentioned: How to Optimize a CUDA Matmul Kernel for cuBLAS-like Performance: a Worklog: In this post, I’ll iteratively optimize an implementation of matrix multiplication written in CUDA.My goal is not to build a cuBLAS replacement, but to deepl...


CUDA MODE ▷ #off-topic (1 messages):

pauleonix: Any suckerpinch fans around here? 😆 https://youtu.be/Ae9EKCyI1xU


CUDA MODE ▷ #llmdotc (7 messages):

  • Activation Value Saving
  • Activation Checkpointing
  • Memory Optimization Techniques
  • Liger Kernel Memory Management

CUDA MODE ▷ #cudamode-irl (11 messages🔥):

  • CUDA-MODE IRL Event Details
  • Quantization and Sparsity Projects
  • GPU Availability for Hacking

Link mentioned: Quantization and Sparsity Projects: Quantization and Sparsity Projects for IRL High Performance Implementation Projects: 1. Develop an A16W3 (mixed fp16 x 3-bit) Fused Matmul Kernel: Why? Currently, there is no available kernel for 3-...


CUDA MODE ▷ #liger-kernel (7 messages):

  • Benchmarking phi3
  • GPU utilization concerns
  • OOM issues with sequence length
  • GPU CI failures

Links mentioned:


Cohere ▷ #discussions (31 messages🔥):

  • Cohere's Acceptable Use Policy
  • Fine-tuning Models
  • Community Introductions
  • Bot Maintenance Updates

Links mentioned:


Cohere ▷ #questions (2 messages):

  • Embedding documents
  • Fine-tuning LLMs

Cohere ▷ #api-discussions (1 messages):

  • Temperature settings in outputs

Cohere ▷ #projects (42 messages🔥):

  • Advanced Computer Vision Projects
  • Multimodal Learning
  • Pokedex Project
  • Google Vision API
  • Team Collaboration

OpenInterpreter ▷ #general (14 messages🔥):

  • Windows usage
  • Desktop beta
  • Android mobile devices
  • Open Interpreter product discussion
  • Project issues

Links mentioned:


OpenInterpreter ▷ #O1 (57 messages🔥🔥):

  • 01 Light Discontinuation
  • Refund Process
  • 01 App Launch
  • Testing and Beta Feedback
  • Community Support for Open Source

Links mentioned:


OpenInterpreter ▷ #ai-content (5 messages):

  • Tool Use Episode Release
  • YouTube Links

Link mentioned: Activity Tracker and Calendar Automator - Ep 4 - Tool Use: Time is our most precious resource, let's use AI to optimize it!In this episode of Tool Use, Mike Bird and Ty Fiero discuss the importance of time management...


Modular (Mojo 🔥) ▷ #general (10 messages🔥):

  • Windows Native Version
  • Focus on Linux Support
  • WSL Support
  • Community Meeting
  • User Feedback Opportunity

Link mentioned: Appointments: no description found


Modular (Mojo 🔥) ▷ #mojo (61 messages🔥🔥):

  • Mojo language capabilities
  • DLHandle in Mojo
  • GStreamer bindings
  • Variant type in Mojo
  • SDL bindings in Mojo

Links mentioned:


Nous Research AI ▷ #general (65 messages🔥🔥):

  • DisTro confusion
  • AI training pitfalls
  • OCTAV algorithm implementation
  • Repetition in AI responses
  • Performance of various AI models

Link mentioned: Tweet from anton (@abacaj): Could not reproduce the 91% humaneval score for reflection (ref_70_e3), run locally using bf16 with vLLM. Used the "recommended" system prompt + extracting from output tags: 81.1% meta-llama-...


Nous Research AI ▷ #ask-about-llms (6 messages):

  • Scaling in AI
  • Quality of Data
  • Rich Sutton's Bitter Lesson
  • AI Research Trends

Link mentioned: The Bitter Lesson: no description found


Torchtune ▷ #general (53 messages🔥):

  • Tokenizer eos issue
  • Eleuther_Eval recipe loading
  • ChatML format for datasets
  • Checkpointing in training
  • Hugging Face TRL library

Links mentioned:


Torchtune ▷ #dev (17 messages🔥):

  • Mixed Precision Training
  • Liger vs Compile Speed
  • Dynamic seq_len Challenges
  • Chunked CE Memory Usage
  • FP8 Integration Ideas

Link mentioned: Reduce compile time for single-device and multi-device recipes by yf225 · Pull Request #1445 · pytorch/torchtune: Context What is the purpose of this PR? Is it to add a new feature fix a bug update tests and/or documentation other (improve compile time) Improvements in compile time (on my A100 machine): ...


Perplexity AI ▷ #announcements (1 messages):

  • Jim Harbaugh
  • Perplexity Playbook
  • Social Media Updates

Links mentioned:


Perplexity AI ▷ #general (57 messages🔥🔥):

  • Reflection LLM addition
  • Perplexity Pro rewards issue
  • Claude 3.5 performance concerns
  • Search functionality problems
  • User prompts and formatting

Links mentioned:


Perplexity AI ▷ #sharing (6 messages):

  • Apple iPhone event
  • AI detecting fake science
  • Nvidia Q2 earnings
  • Artistic journalism
  • Top IDEs for programming

Perplexity AI ▷ #pplx-api (2 messages):

  • search_domain_filter API
  • API functionality

Latent Space ▷ #ai-general-chat (47 messages🔥):

  • Apple Intelligence updates
  • ColPali model advancements
  • Superforecasting AI release
  • Strawberry OpenAI model
  • Expand.ai launch

Links mentioned:


LlamaIndex ▷ #blog (2 messages):

  • Agentic RAG
  • LlamaIndex
  • Search For RAG in the LLM era
  • Maven course
  • RAG strategies

LlamaIndex ▷ #general (45 messages🔥):

  • LlamaIndex and Llama 3 examples
  • Pandas DataFrame querying
  • Integration issues with MLflow
  • Kapa.ai usage and troubleshooting
  • Similarity search methods in LlamaIndex

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (39 messages🔥):

  • Deception 70B
  • OpenAI's Strawberry release
  • Otherside AI scams
  • AI forecasting systems
  • Exiting OpenAI employees

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (2 messages):

  • Gemini and Cursor integration
  • User experiences with Cursor

Stability.ai (Stable Diffusion) ▷ #general-chat (41 messages🔥):

  • Image Generation Hardware
  • Deep Dream Machine Alternatives
  • Training Tips for SDXL
  • Understanding CLIP Models
  • Discord Bot for AI Services

Links mentioned:


LAION ▷ #general (28 messages🔥):

  • Open Source AI Panel
  • Performance of AI Models
  • Private Machine Learning Solutions
  • Multiparty Computation in AI
  • Security in Machine Learning Deployment

Link mentioned: GitHub Presents: Open Source AI - Access, Democratization, and Responsibility · Luma: AI is rapidly transforming industries from software development, content creation, agentic workflows and beyond. Central to this transformation is open source…


LAION ▷ #research (1 messages):

chad_in_the_house: wow that's annoying lol


OpenAccess AI Collective (axolotl) ▷ #general (8 messages🔥):

  • AI Research Fraud
  • Reasoner Dataset
  • iChip Technology
  • Hugging Face Multi-Packing

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (3 messages):

  • OpenAI Fine-Tuning API
  • Chat Template Importer Changes
  • Weight Parameter for Training Data

OpenAccess AI Collective (axolotl) ▷ #general-help (4 messages):

  • BNB Issue Thread
  • H100 Performance without 8-bit
  • Fine-tuning Mistral NeMo
  • Errors with Padding Token in Fine-tuning

LangChain AI ▷ #general (4 messages):

  • Claude 3.5 audio capability
  • Token counting in langchain4j
  • Whisper as an alternative for transcription

LangChain AI ▷ #share-your-work (4 messages):

  • Chat AI Lite
  • EDA-GPT
  • Pilerbot

Links mentioned:


DSPy ▷ #general (8 messages🔥):

  • Emotion classification code
  • AdalFlow GitHub library
  • Llama AI model prompt
  • MIPRO prompt optimizer

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #general (3 messages):

  • LLM observability platforms
  • Anthropic API performance

Gorilla LLM (Berkeley Function Calling) ▷ #leaderboard (3 messages):

  • Merge Conflicts Resolution
  • Test Results Storage

tinygrad (George Hotz) ▷ #general (1 messages):

kimchiking7364: 🏄


MLOps @Chipro ▷ #events (1 messages):

  • Open Source AI Event
  • Panelists from Industry
  • Event Registration

Link mentioned: GitHub Presents: Open Source AI - Access, Democratization, and Responsibility · Luma: AI is rapidly transforming industries from software development, content creation, agentic workflows and beyond. Central to this transformation is open source…





{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}