Frozen AI News archive

not much happened today

This week in AI news highlights **Ollama 0.4** supporting **Meta's Llama 3.2 Vision** models (11B and 90B), with applications like handwriting recognition. **Self-Consistency Preference Optimization (ScPO)** was introduced to improve model consistency without human labels. Discussions on **model scaling**, **neural networks resurgence**, and **AMD's multi-GPU bandwidth** challenges were noted. The importance of **skip connections** in **Transformers** was emphasized. In healthcare, **less regulation plus AI** could revolutionize disease treatment and aging. Tools like **LlamaParse** and **Gemini** aid automated resume insights. **Gitpod Flex** demonstrated zero-trust architecture for secure development environments. Research includes surveys on **Small Language Models (SLMs)**, **number understanding** in LLMs, and **DTrOCR** using a **GPT-2 decoder** for OCR. Multi-agent systems in prediction markets were discussed by **TogetherCompute** and **LangChainAI**. Community events include **NeurIPS Happy Hour**, **NLP seminars**, and courses on **Agent Memory** with LLMs as operating systems.

Canonical issue URL

a quiet week is all we need.

AI News for 11/6/2024-11/7/2024. We checked 7 subreddits, 433 Twitters and 30 Discords (217 channels, and 1985 messages) for you. Estimated reading time saved (at 200wpm): 222 minutes. You can now tag @smol_ai for AINews discussions!

Anon on reddit thinks he has figured out AGI but ends up writing a kind of coherent literature review of Liquid Neural Networks and related work. The comments are mandatory.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

AI Models and Architectures

AI Tools and Applications

AI Research and Publications

AI Community and Events

AI in Business and Industry

Memes/Humor

Miscellaneous


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. LLM Selector: Analyzing Models Across 12 Benchmarks for Optimal Use

Theme 2. Integration of Liquid Time Constant Networks with Spiking Dynamics

Theme 3. Qwen 2.5 Coder: Stealth Updates & Future Directions

Theme 4. WebRL: Evolving Agents via Self-Developed Curriculum Reinforcement Learning

Theme 5. Open Source Models Revealing Significantly Lower Refusal Rates

Other AI Subreddit Recap

/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT

Theme 1. Claude 3.5 Sonnet New Update Effect on Code and Text Output

Theme 2. Nvidia's New GPUs: Reduced VRAM Limits Local AI Training

Theme 3. Anthropic's Secretive ClaudeAI Prompt Management Exposed

Theme 4. ChatGPT and ClaudeAI's New Limitations on Code Output


AI Discord Recap

A summary of Summaries of Summaries by O1-mini

1. AI Model Innovations and Releases

2. Performance Optimization and Resource Management

3. Integration with Platforms and Tools

4. AI Applications in Diverse Domains

5. AI Fine-Tuning and Customization


PART 1: High level Discord summaries

Nous Research AI Discord


Perplexity AI Discord


OpenRouter (Alex Atallah) Discord


Eleuther Discord


Unsloth AI (Daniel Han) Discord


HuggingFace Discord


OpenAI Discord


GPU MODE Discord


Notebook LM Discord Discord


Interconnects (Nathan Lambert) Discord


LM Studio Discord


Stability.ai (Stable Diffusion) Discord


Latent Space Discord


Modular (Mojo 🔥) Discord


Cohere Discord


LlamaIndex Discord


DSPy Discord


OpenInterpreter Discord


tinygrad (George Hotz) Discord


OpenAccess AI Collective (axolotl) Discord


Torchtune Discord


LLM Agents (Berkeley MOOC) Discord


Gorilla LLM (Berkeley Function Calling) Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Finetuning (Hamel + Dan) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LAION Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Nous Research AI ▷ #announcements (1 messages):

  • Nous Chat
  • Hermes 3
  • User Interface Enhancements

Link mentioned: NOUS CHAT | Talk to Hermes: Experience natural, intelligent conversations with Hermes, the open-source LLM by Nous Research.


Nous Research AI ▷ #general (215 messages🔥🔥):

  • TEE HEE HE wallet updates
  • Nous Research overview
  • Hermes 405B performance
  • Development discussions
  • Future of AI training data

Links mentioned:


Nous Research AI ▷ #ask-about-llms (17 messages🔥):

  • Haiku 3.5 Performance
  • Haiku 3.5 Pricing
  • Open Source AI Interfaces
  • Evaluating PyTorch Models

Nous Research AI ▷ #research-papers (1 messages):

  • Ferret-UI
  • Gemma-2B MLLM
  • Llama-3-8B MLLM
  • UI-centered reasoning
  • Mobile UI comprehension

Links mentioned:


Nous Research AI ▷ #research-papers (1 messages):

  • Ferret-UI
  • Gemma-2B
  • Llama-3-8B
  • Multimodal language models
  • Mobile UI comprehension

Links mentioned:


Nous Research AI ▷ #rag-dataset (6 messages):

  • YouTube summarizer project
  • Interactive chat sessions
  • Model recommendations
  • Hugging Face resources

Perplexity AI ▷ #general (221 messages🔥🔥):

  • Perplexity Pro subscription
  • Claude model capabilities
  • Image saving on Mac OS
  • Mobile device specifications
  • Mac OS login issues

Links mentioned:


Perplexity AI ▷ #sharing (13 messages🔥):

  • Chernobyl's Radiation-Eating Fungi
  • Physical buttons return
  • Michigan as a Climate Sanctuary
  • Future of AI
  • Best value speakers and amp

OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

  • Completion API migration
  • Scheduled downtime for database upgrade

OpenRouter (Alex Atallah) ▷ #general (224 messages🔥🔥):

  • Hermes Resurgence
  • Claude API Changes
  • Mistral's New Features
  • OpenRouter API Issues
  • Chinese AI Models Pricing

Links mentioned:

: no description foundModel Deprecations - Anthropic: no description foundMistral Moderation API: We are introducing our new moderation service enabling our users to detect undesirable text content along several policy dimensions.Mistral Batch API: Lower cost API for AI builders.Mistral AI API | Mistral AI Large Language Models: Our Chat Completion and Embeddings APIs specification. Create your account on La Plateforme to get access and read the docs to learn how to use...


OpenRouter (Alex Atallah) ▷ #beta-feedback (5 messages):

  • Customer Provider Keys
  • Integration Beta Features

Eleuther ▷ #general (57 messages🔥🔥):

  • Research opportunities in Switzerland
  • Jazz piano generation models
  • Music evaluation methods
  • AI and musical upscaling
  • Magenta's Music Transformer

Link mentioned: Listen to Transformer: An app to make it easier to explore and curate output from a music transformer


Eleuther ▷ #research (116 messages🔥🔥):

  • Flash Attention Techniques
  • Momentum Decay in Optimizations
  • NaNoGPT Updates
  • Evaluation Data Contamination
  • Advanced Attention Mechanisms

Links mentioned:


Eleuther ▷ #gpt-neox-dev (1 messages):

  • NeoX vs LitGPT benchmarks
  • Pretraining setups

Unsloth AI (Daniel Han) ▷ #general (40 messages🔥):

  • Fine-tuning Smollm2
  • Feature contribution analysis
  • Gemma2 model errors
  • 8bit and 4bit support
  • AI Unplugged newsletter

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (51 messages🔥):

  • NIM API Payment Options
  • AI Model Usage Feedback
  • Discord Scam Discussions
  • Exam Preparation
  • Mathematics Interview Questions

Unsloth AI (Daniel Han) ▷ #help (53 messages🔥):

  • Model Output Handling
  • VRAM Consumption Comparison
  • Fine-Tuning Approach
  • Token Count Mismatch in Models
  • Concurrent Inference Setup

Unsloth AI (Daniel Han) ▷ #community-collaboration (2 messages):

  • torch.compile and gradient checkpointing
  • product metrics tracking
  • software distribution in container ecosystem

Unsloth AI (Daniel Han) ▷ #research (1 messages):

  • Training signal from LLM
  • Judge evaluation in training

HuggingFace ▷ #general (119 messages🔥🔥):

  • Serverless Inference Endpoint for Hermes3
  • Using Hugging Face Models for Arabic Language
  • Quantization Techniques for Vision Models
  • Suggestions for Token Classification Architectures
  • Using API for Hugging Face Spaces

Links mentioned:


HuggingFace ▷ #today-im-learning (1 messages):

  • Building AI model for cybersecurity
  • Challenges in AI for cybersecurity

HuggingFace ▷ #cool-finds (2 messages):

  • Hunyuan3D-1 Framework
  • Grape Leaf Disease Detection App

Links mentioned:


HuggingFace ▷ #i-made-this (13 messages🔥):

  • Formula 1 telemetry chatbot
  • TinyLlama model conversion
  • Harmony questionnaire harmonization
  • USDA FoodData Central dataset
  • Text-to-3D generation

Links mentioned:


HuggingFace ▷ #computer-vision (4 messages):

  • OmniParser model
  • User engagement
  • Model performance
  • UI Interaction datasets

Link mentioned: microsoft/OmniParser · Hugging Face: no description found


HuggingFace ▷ #NLP (1 messages):

  • MaskGCT
  • F5-TTS
  • AI phone caller
  • audio chunk streaming

HuggingFace ▷ #diffusion-discussions (2 messages):

  • Integration Discussions
  • Automatic1111 SD3.5 Support

OpenAI ▷ #ai-discussions (85 messages🔥🔥):

  • SearchGPT performance
  • Best OpenAI models for coding
  • AI self-awareness discussions
  • Role of AI in jobs
  • Using AI as a study tool

OpenAI ▷ #gpt-4-discussions (4 messages):

  • Canvas document deletion
  • Loss of saved GPTs
  • Sidebar pinning issues
  • Custom GPT features improvement

OpenAI ▷ #prompt-engineering (3 messages):

  • File Handling Issues
  • Direct Messaging for Support

OpenAI ▷ #api-discussions (3 messages):

  • File format issues
  • Direct messaging for troubleshooting

GPU MODE ▷ #general (32 messages🔥):

  • Installing PyTorch
  • Kernel Development on Windows
  • CUDA Compiling Issues
  • Using Visual Studio for CUDA
  • Runtime Errors in CUDA

GPU MODE ▷ #triton (4 messages):

  • FP16 x FP16 Performance
  • A100 GPU Efficiency
  • CUDA Cores and GEMVs

GPU MODE ▷ #torch (6 messages):

  • Torch Script Debugging
  • Performance Overhead of .tolist() Call
  • C++ API Limitations

GPU MODE ▷ #cool-links (1 messages):

marksaroufim: https://x.com/_seemethere/status/1838319643347554756


GPU MODE ▷ #beginner (8 messages🔥):

  • Compute heavy operations in LLMs
  • Resources for GEMM optimization
  • CUDA and GPU learning materials

Links mentioned:


GPU MODE ▷ #off-topic (3 messages):

  • Digital Audio Primer
  • Discord AV1 Embed Tool
  • Video Embedding Constraints

Links mentioned:


GPU MODE ▷ #arm (1 messages):

marksaroufim: https://discord.gg/zCkRcp6e?event=1304137506014629969


GPU MODE ▷ #liger-kernel (19 messages🔥):

  • Generalized JSD kernel improvements
  • Engineering blog acknowledgments
  • S1ro nickname origin
  • LinkedIn profiles

GPU MODE ▷ #🍿 (6 messages):

  • Automated Deployments via Heroku
  • Utilizing Raspberry Pi for Discord Bots
  • Goals for Project Popcorn
  • Server Connectivity
  • GPU Implementation Plans

Link mentioned: Project Popcorn: Generate SOTA kernels with LLMs in public: Project Popcorn: Generate SOTA kernels with LLMs in public - 🍿.md


GPU MODE ▷ #thunderkittens (6 messages):

  • ThunderKittens Contribution List
  • Beginner Contributions to Kernels

Links mentioned:


Notebook LM Discord ▷ #use-cases (24 messages🔥):

  • Using Google Docs for Research
  • Podcast Reuse Policy
  • Fun Prompts for Podcast Hosts
  • Dyslexia and Day Trading
  • Understanding Terms of Service

Link mentioned: GitHub - robbiemu/llama-gguf-optimize: Scripts and tools for optimizing quantizations in llama.cpp with GGUF imatrices.: Scripts and tools for optimizing quantizations in llama.cpp with GGUF imatrices. - robbiemu/llama-gguf-optimize


Notebook LM Discord ▷ #general (59 messages🔥🔥):

  • NotebookLM Performance Issues
  • PDF Integration Challenges
  • Sharing Options and Bugs
  • Saved Notes Accessibility
  • Project Innovations for NotebookLM

Link mentioned: Mastering the Digital SAT: Expert Tips for Reading & Writing with Special Guest Riley | Episode 9: Visit our website for free SAT and GRE test preparation: https://campusgoals.com/Welcome to Episode 9 of "Mastering the SAT with Alex and Taylor," your go-to...


Interconnects (Nathan Lambert) ▷ #news (15 messages🔥):

  • Anthropic and Palantir partnership
  • U.S. government classification levels
  • Concerns about defense collaboration
  • Reddit community for security clearances

Links mentioned:


Interconnects (Nathan Lambert) ▷ #other-papers (5 messages):

  • 8-bits pretraining
  • Tim Dettmers' paper
  • YouTube talks on pretraining

Interconnects (Nathan Lambert) ▷ #ml-questions (1 messages):

  • Synthetic Data Generation
  • SFT Data Scaling
  • Instruction Data Usage
  • T0 Comparisons

Link mentioned: Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent: In this paper, we introduce Hunyuan-Large, which is currently the largest open-source Transformer-based mixture of experts model, with a total of 389 billion parameters and 52 billion activation param...


Interconnects (Nathan Lambert) ▷ #random (3 messages):

  • MLST podcast episode with Chollet
  • Chollet's views on memorization
  • Frustrations with Tim's style

Interconnects (Nathan Lambert) ▷ #memes (8 messages🔥):

  • Election polling as prompt engineering
  • Bias in polling methodologies
  • Monetary gains from polling insights

Link mentioned: Tweet from Siqi Chen (@blader): so the french guy who yolo’d $30m on trump on polymarket noticed polls asking “who are your neighbors voting for?” instead of “who are you voting for?” had trump overperforming. so he commissioned hi...


Interconnects (Nathan Lambert) ▷ #posts (32 messages🔥):

  • Tim's move to CMU
  • Quantization techniques
  • Character.AI's inference optimization
  • GPU efficiency
  • Community discussions on quantization

Link mentioned: Optimizing AI Inference at Character.AI: At Character.AI, we're building toward AGI. In that future state, large language models (LLMs) will enhance daily life, providing business productivity and entertainment and helping people with e...


LM Studio ▷ #general (41 messages🔥):

  • Ollama's llama 3.2 Vision
  • MLX Engine updates
  • NVIDIA's Call for Feedback
  • LM Studio CPU Mode
  • Unsupported Models in LM Studio

Links mentioned:


LM Studio ▷ #hardware-discussion (20 messages🔥):

  • Single Slot RTX 4090
  • Mac M2 Pro Memory Usage
  • Large Model Performance
  • Experimental Context Settings
  • MacBook M4 Reviews

Stability.ai (Stable Diffusion) ▷ #general-chat (55 messages🔥🔥):

  • Stable Diffusion Models
  • Outpainting Techniques
  • User Interface Generation
  • Local Installation Guides
  • Discord Community Interaction

Links mentioned:


Latent Space ▷ #ai-general-chat (38 messages🔥):

  • Llama 3.2 Vision
  • Aide Open Source IDE
  • Claude AI Limitations
  • Training Open Language Models
  • Codebuff CLI Tool

Links mentioned:


Modular (Mojo 🔥) ▷ #mojo (28 messages🔥):

  • No Bounds Check Decorator
  • Mojo Standard Library Development
  • Mojo Interoperability with Python and C/C++
  • Mojo as a Python Superset
  • Timeline for Mojo Features

Links mentioned:


Cohere ▷ #discussions (21 messages🔥):

  • Cohere Reranker API
  • Command-R-Plus Issues
  • AWS Bedrock with SpringAI

Cohere ▷ #announcements (1 messages):

  • Open-source fine-tuning
  • Hugging Face integration
  • SageMaker deployment

Link mentioned: GitHub - cohere-ai/cohere-finetune: A tool that facilitates easy, efficient and high-quality fine-tuning of Cohere's models: A tool that facilitates easy, efficient and high-quality fine-tuning of Cohere's models - cohere-ai/cohere-finetune


Cohere ▷ #questions (1 messages):

mrdragonfox: <@1303804989629534333> [email protected]


Cohere ▷ #api-discussions (3 messages):

  • AWS Bedrock
  • SpringAI Embeddings

LlamaIndex ▷ #blog (2 messages):

  • Automated Resume Insights
  • Context Refinement in RAG Systems

LlamaIndex ▷ #general (23 messages🔥):

  • Ollama Llama Vision Integration
  • Open Source Chatbot UI
  • Chainlit with LlamaIndex for Multimodal RAG
  • Resources for Llama-Parse
  • Isolated Instrumentation in Workflows

Link mentioned: GitHub - mckaywrigley/chatbot-ui: Come join the best place on the internet to learn AI skills. Use code "chatbotui" for an extra 20% off.: Come join the best place on the internet to learn AI skills. Use code "chatbotui" for an extra 20% off. - mckaywrigley/chatbot-ui


DSPy ▷ #show-and-tell (2 messages):

  • Future of Dott.ai
  • Steve's vision

Links mentioned:


DSPy ▷ #general (19 messages🔥):

  • DSPy Docstring Issues
  • Understanding DSPy Framework
  • EMNLP 2024 Presentation
  • Modular Language Model Optimization

Links mentioned:


OpenInterpreter ▷ #general (11 messages🔥):

  • Understanding OS Mode
  • Discord Event Timing
  • Viewer Limitations
  • OmniParser Tool
  • Using OS Mode with gpt-4o

Links mentioned:


OpenInterpreter ▷ #ai-content (3 messages):

  • Python 3.13 Compatibility Issues
  • Conda Environment Solutions

tinygrad (George Hotz) ▷ #general (9 messages🔥):

  • Dedicated Transformer ASIC
  • Sohu Hardware Architecture
  • Discussion on Availability of Custom Hardware

Link mentioned: Tweet from Rohan Paul (@rohanpaul_ai): MASSIVE 🤯 First dedicated transformer ASIC (Application-Specific Integrated Circuit) just dropped Custom chip burns transformer architecture directly into silicon, making AI models run 10x faster t...


tinygrad (George Hotz) ▷ #learn-tinygrad (3 messages):

  • Using multiple GPUs
  • Sharding models
  • ThreadPoolExecutor issues

OpenAccess AI Collective (axolotl) ▷ #general (8 messages🔥):

  • ScheduleFree SOAP
  • Optimizer Hyperparameters
  • Model Merging and MOEs
  • CAME Optimizer

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (3 messages):

  • Adding special tokens
  • Fine-tuning with LORA

Torchtune ▷ #general (4 messages):

  • LR Scheduler Usage
  • Ichigo Project Implementation
  • Adding Scheduler to Recipes

Links mentioned:


LLM Agents (Berkeley MOOC) ▷ #mooc-questions (2 messages):

  • LLM Course Cohorts

Gorilla LLM (Berkeley Function Calling) ▷ #discussion (1 messages):

  • Function extraction
  • Dataset files






{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}