Frozen AI News archive

not much happened this weekend

**Nous Research** announced **DisTrO**, a new optimizer that drastically reduces inter-GPU communication by 1000x to 10,000x enabling efficient training on slow networks, offering an alternative to **GDM's DiLoCo**. **Cursor AI** gained viral attention from an 8-year-old user and announced a new fundraise, with co-host Aman returning to their podcast. **George Hotz** launched **tinybox** for sale. In robotics, **AGIBOT** revealed 5 new humanoid robots with open-source plans, and **Unitree** showcased its G1 humanoid robot nearing mass production at $16,000. **ETH Zurich** and **Disney** developed an AI system for physics-based robot motion generation from text or images. **UC San Diego** released **ACE**, an open-source teleoperation system for controlling multiple robots. AI21 Labs unveiled **Jamba 1.5**, a multilingual model with 256k context length and permissive licensing. **Luma Labs** released **Dream Machine 1.5** for improved text-to-video generation. **Ideogram** launched **v2** of its text-to-image model with near-perfect text generation. **Nvidia** and **Mistral** released **Mistral-NeMo-Minitron 8B**, a small model outperforming **Mistral-7B** and **llama-3-8b** on the Open LLM leaderboard.

Canonical issue URL

AI News for 8/23/2024-8/26/2024. We checked 7 subreddits, 384 Twitters and 30 Discords (214 channels, and 5673 messages) for you. Estimated reading time saved (at 200wpm): 639 minutes. You can now tag @smol_ai for AINews discussions!

A few news items:

Since the newsflow is light, why not give Box feedback on Box AI's new beta?


[Sponsored by Box] You are building things with AI. So is Box. Imagine if you built your things using Box’s things. Actually, don’t imagine it, try it yourself in the Box AI Developer Zone.

Swyx's comment: thanks to Box (via Freeman & Forrest) for supporting AI News this August (1, 2, 3)!


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

AI and Robotics Developments

AI Models and Tools

AI Applications and Research

AI Development and Tools

AI Research and Frameworks

AI Ethics and Regulation

Memes and Humor

This summary captures the key developments, research, and discussions in AI and robotics from the provided tweets, focusing on aspects relevant to AI engineers and developers.


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Hardware Optimization for Local LLM Inference

Theme 2. Advancements in Long-Context LLM Generation

Theme 3. Anthropic's Controversial Stance on AI Regulation

Theme 4. Emerging Chinese LLMs Challenging Western Models

All AI Reddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

Robotics and AI Hardware

Biotechnology and Food Tech

AI Model Development


AI Discord Recap

A summary of Summaries of Summaries by Claude 3.5 Sonnet

1. LLM Advancements and Benchmarking

2. LLM Optimization Techniques

3. Open Source AI Developments

4. AI Industry and Community Updates


PART 1: High level Discord summaries

Nous Research AI Discord


Unsloth AI (Daniel Han) Discord


Stability.ai (Stable Diffusion) Discord


HuggingFace Discord


OpenAI Discord


Perplexity AI Discord


OpenRouter (Alex Atallah) Discord


Eleuther Discord


Latent Space Discord


tinygrad (George Hotz) Discord


Cohere Discord


LlamaIndex Discord


Torchtune Discord


LangChain AI Discord


OpenAccess AI Collective (axolotl) Discord


Modular (Mojo 🔥) Discord


OpenInterpreter Discord


DSPy Discord


Gorilla LLM (Berkeley Function Calling) Discord


LAION Discord


Interconnects (Nathan Lambert) Discord


Alignment Lab AI Discord


LLM Finetuning (Hamel + Dan) Discord


MLOps @Chipro Discord


DiscoResearch Discord


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Nous Research AI ▷ #announcements (1 messages):

  • DisTrO
  • Distributed Optimization
  • LLM Training
  • Inter-GPU Communication
  • AdamW

Link mentioned: DisTrO/A_Preliminary_Report_on_DisTrO.pdf at main · NousResearch/DisTrO: Distributed Training Over-The-Internet. Contribute to NousResearch/DisTrO development by creating an account on GitHub.


Nous Research AI ▷ #general (786 messages🔥🔥🔥):

  • Hermes 2
  • Mistral struggles
  • Model Merging
  • Open Empathic
  • GPTs Agents

Links mentioned:


Nous Research AI ▷ #ask-about-llms (25 messages🔥):

  • Co-Writer
  • Discord Bots
  • LLM Model
  • LLM Quantization
  • MoE

Links mentioned:


Nous Research AI ▷ #research-papers (8 messages🔥):

  • Pints LLM
  • Medical AI Research
  • LLM Benchmarking
  • LLM Training Efficiency
  • Multimodal LLMs

Links mentioned:


Nous Research AI ▷ #interesting-links (3 messages):

  • Pints-AI 1.5
  • Sparse-Marlin
  • Text Generation Webui

Links mentioned:


Nous Research AI ▷ #research-papers (8 messages🔥):

  • 1.5-Pints LLM
  • Medical AI Research
  • LLaVA-Surg
  • HIBOU
  • RuleAlign

Links mentioned:


Nous Research AI ▷ #rag-dataset (1 messages):

  • Whisper Diarization
  • Whisper v3

Unsloth AI (Daniel Han) ▷ #general (512 messages🔥🔥🔥):

  • Unsloth
  • Ollama
  • Llama.cpp
  • LinkedIn
  • Model Merging

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (14 messages🔥):

  • Liger Kernel
  • Triton
  • EAGLE
  • LLM Training

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (67 messages🔥🔥):

  • Unsloth CPU Compatibility
  • Unsloth on Tesla P100
  • Unsloth SFTTrainer Override
  • Unsloth Steps vs Epochs
  • Unsloth Phi-3.5-MoE Availability

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (1 messages):

  • Replete-LLM V2
  • Llama 3.1
  • Replete-AI/The_Living_AI_Dataset
  • System Prompts
  • Quantizations

Links mentioned:


Unsloth AI (Daniel Han) ▷ #community-collaboration (3 messages):

  • LLM Fine-Tuning for Question Answering
  • Unsloth and Llama Model

Stability.ai (Stable Diffusion) ▷ #general-chat (570 messages🔥🔥🔥):

  • Stable Diffusion Online
  • Stable Diffusion Seed
  • ComfyUI
  • Stable Diffusion Image Upscaling
  • Flux

Links mentioned:


HuggingFace ▷ #general (366 messages🔥🔥):

  • Codegemma
  • Hermes
  • Mistral
  • Model Merging
  • Open Empathic

Links mentioned:


HuggingFace ▷ #today-im-learning (18 messages🔥):

  • Model Quantization
  • Model Distillation
  • Diffusion Model Training
  • GPT-2 Rewriting
  • Prompt Weights

Links mentioned:


HuggingFace ▷ #cool-finds (6 messages):

  • Stale bot
  • LLM Merging
  • Medical AI Papers
  • 1-bit LLMs

Links mentioned:


HuggingFace ▷ #i-made-this (14 messages🔥):

  • Tau LLM
  • database match command
  • TauAgent class
  • reward signals
  • TinyLlama

Links mentioned:


HuggingFace ▷ #reading-group (14 messages🔥):

  • AWS RAGCHECKER
  • DPO
  • KTO
  • LLMs Alignment

HuggingFace ▷ #computer-vision (2 messages):

  • FSRCNN implementation
  • PyTorch implementation

Link mentioned: notebooks/0_dummy.ipynb · master · Amri Rasyidi / Paper Replication · GitLab: GitLab.com


HuggingFace ▷ #NLP (7 messages):

  • VLLM memory usage
  • VLLM system memory utilization
  • Accelerate trainer not working on Google Colab TPU
  • Speculative decoding with tool calling

Links mentioned:


HuggingFace ▷ #diffusion-discussions (8 messages🔥):

  • Hugging Face iOS App Model Configuration
  • Stable Diffusion Fine-Tuning
  • Quantization Changes in Stable Diffusion

OpenAI ▷ #ai-discussions (275 messages🔥🔥):

  • GPTs
  • AGI
  • LLMs
  • Model Scaling
  • DeepMind

Link mentioned: What runs GPT-4o? Inside Microsoft's 2024 AI supercomputer with Mark Russinovich: Microsoft has built the world’s largest cloud-based AI supercomputer that is already exponentially bigger than it was just 6 months ago, paving the way for a...


OpenAI ▷ #gpt-4-discussions (8 messages🔥):

  • GPT sharing and community building
  • GPT Builder issues
  • OpenAI API pricing and subscription models

OpenAI ▷ #prompt-engineering (12 messages🔥):

  • GPTs custom builders
  • Prompt engineering templates
  • GPT jailbreak
  • Meta robot tag

OpenAI ▷ #api-discussions (12 messages🔥):

  • GPT-4 Jailbreak
  • Meta Robot Tag
  • Prompt Engineering

Perplexity AI ▷ #announcements (2 messages):

  • Perplexity Creator Community
  • Kale partnership
  • Perplexity Pro for LinkedIn Premium

Link mentioned: Perplexity AI Creator Community: Earn rewards to post about Perplexity AI!


Perplexity AI ▷ #general (216 messages🔥🔥):

  • Perplexity Pro
  • GPT-4o
  • Perplexity's performance
  • Claude 3.5 Sonnet
  • Image generation in Perplexity

Links mentioned:


Perplexity AI ▷ #sharing (24 messages🔥):

  • Perplexity Onboarding
  • Perplexity Page Features
  • Teaching Seniors
  • Steve Jobs Jacket
  • Model Motivation

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (6 messages):

  • Perplexity API Rate Limits
  • Pro Search API
  • Perplexity API Responsiveness

OpenRouter (Alex Atallah) ▷ #announcements (1 messages):

  • Database Outage

OpenRouter (Alex Atallah) ▷ #general (245 messages🔥🔥):

  • Grok-2
  • Grok-mini
  • Mistral
  • Claude
  • OpenRouter

Links mentioned:


Eleuther ▷ #general (67 messages🔥🔥):

  • OMI
  • LLM failure mode
  • Anthropic Interpretability Work
  • Sparse MoE
  • Model Interpretability

Links mentioned:


Eleuther ▷ #research (131 messages🔥🔥):

  • GNN Research
  • Chinchilla Scaling Laws
  • ARC Research
  • SAE Papers
  • GPT4o

Links mentioned:


Eleuther ▷ #scaling-laws (2 messages):

  • Scaling Laws
  • Model Scaling Limits

Link mentioned: Inverse Scaling: When Bigger Isn't Better: Work on scaling laws has found that large language models (LMs) show predictable improvements to overall loss with increased scale (model size, training data, and compute). Here, we present evidence f...


Eleuther ▷ #lm-thunderdome (17 messages🔥):

  • OpenAI ChatGPT4o and Anthropic external APIs multiple choice eval
  • llama 3.1 eval
  • lm-eval library output_type
  • lm-eval library doc_to_choice
  • lm-eval library chat_template

Link mentioned: lm-evaluation-harness/lm_eval/evaluator.py at aab42ba836b4af28cc1c5c1e697ea334c6ea7ced · EleutherAI/lm-evaluation-harness: A framework for few-shot evaluation of language models. - EleutherAI/lm-evaluation-harness


Eleuther ▷ #gpt-neox-dev (3 messages):

  • GPU Utilization Measurement
  • NVIDIA-SMI Overreporting
  • DCGM-Exporter
  • PyNVML
  • TFLOPs and HFU/MFU Logging

Link mentioned: GitHub - EleutherAI/cookbook: Deep learning for dummies. All the practical details and useful utilities that go into working with real models.: Deep learning for dummies. All the practical details and useful utilities that go into working with real models. - EleutherAI/cookbook


Latent Space ▷ #ai-general-chat (54 messages🔥):

  • BERTopic
  • OpenAI's sidebars
  • GPTs Agents
  • OpenEmpathic project
  • Model Merging

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

  • AI Engineer London Meetup
  • AI Engineer World's Fair
  • Meetup speakers

Link mentioned: Tweet from Damien C. Tanner (@dctanner): We're brining a slice of @swyx's AI Engineer World's Fair to London! Evening of 12 September is the first AI Engineer London Meetup. Hear from 4 amazing speakers: @maximelabonne, @rovio...


Latent Space ▷ #ai-in-action-club (78 messages🔥🔥):

  • Structured Outputs
  • LLM Deduplication
  • TaxonomySynthesis
  • BERTopic
  • GPT Researcher

Links mentioned:


tinygrad (George Hotz) ▷ #announcements (1 messages):

georgehotz: buy your tinybox here https://tinycorp.myshopify.com/


tinygrad (George Hotz) ▷ #general (78 messages🔥🔥):

  • E-graphs
  • Tinygrad
  • Tinybox
  • AMD
  • BERT

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (13 messages🔥):

  • GPU=1 and OpenCL
  • Tensor Loading
  • Training Speed and Data Types
  • Performance Differences Between dtypes.half and dtypes.float
  • RoPE Approximation and Model Matching

Links mentioned:


Cohere ▷ #discussions (37 messages🔥):

  • Command-R pricing
  • Durov's trial
  • Command-R free trial

Links mentioned:


Cohere ▷ #questions (9 messages🔥):

  • Cohere free trial
  • Invite Feature Bug
  • Fine Tuning Cohere Embeddings
  • Rasa Chatbot

Cohere ▷ #api-discussions (32 messages🔥):

  • Cohere API Rate Limits
  • Multiple API Keys
  • Async Requests
  • Rerank 3 Pricing
  • Token Limits

Link mentioned: Pricing: Access our models directly through our API to create scalable production workloads.


LlamaIndex ▷ #blog (5 messages):

  • LlamaIndex
  • Create Llama
  • GraphRAG
  • Data Silos
  • LLMs for newsletters

LlamaIndex ▷ #general (71 messages🔥🔥):

  • tqdm
  • LlamaIndex API changes
  • PineconeVectorStore
  • LlamaIndex and OpenAI models
  • Updating VectorStoreIndex

Links mentioned:


Torchtune ▷ #dev (60 messages🔥🔥):

  • Eager vs compile discrepancy
  • Triton RNG
  • In-place operations and Torch.compile
  • KV cache allocation
  • Cudagraphs and memory consumption

Links mentioned:


LangChain AI ▷ #general (41 messages🔥):

  • LangChain Document Loading
  • LangChain Model Usage
  • LLMChain vs LCEL
  • LangChain with Postgres
  • GenAI in Data Science

Links mentioned:


LangChain AI ▷ #share-your-work (3 messages):

  • AI agents
  • LangGraph framework
  • Retrieval Agents
  • ParDocs
  • RAG

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (15 messages🔥):

  • GPT-4 fine-tuning
  • Mistral
  • LLM Benchmarks
  • lm-eval-harness
  • Glianorex Benchmark

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (17 messages🔥):

  • LIGER
  • LLM Training Efficiency
  • Single GPU Training
  • Training Time

Link mentioned: GitHub - linkedin/Liger-Kernel: Efficient Triton Kernels for LLM Training: Efficient Triton Kernels for LLM Training. Contribute to linkedin/Liger-Kernel development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #general-help (5 messages):

  • Phi-3-medium-128k-instruct training config
  • training config for new tokens

OpenAccess AI Collective (axolotl) ▷ #community-showcase (1 messages):

  • Data Curation
  • LLM Judge
  • Model Rating
  • Data Curation Setup

Modular (Mojo 🔥) ▷ #general (16 messages🔥):

  • Mojo JIT
  • Mojo Script Mode
  • Max/Mojo Development Pace
  • GPU Support
  • Community Meeting

Link mentioned: [Public] Mojo Community Meeting: Mojo Community Meeting This doc link: https://modul.ar/community-meeting-doc This is a public document; everybody is welcome to view and comment / suggest. All meeting participants must adhere to th...


Modular (Mojo 🔥) ▷ #mojo (18 messages🔥):

  • Mojo's memory usage
  • Modverse 42 release schedule
  • Mojo's struct parameters
  • Mojo's UnsafePointer

Modular (Mojo 🔥) ▷ #max (1 messages):

guidorice: not sure if anyone shared it: https://youtu.be/7TnkqfX84gI?si=sqHI4BLGOwneLarH 👍


OpenInterpreter ▷ #general (23 messages🔥):

  • OpenInterpreter Profiles
  • OpenInterpreter --vision on Windows
  • OpenInterpreter OS Mode
  • Prebuilt OpenInterpreter
  • OpenInterpreter Documentation

OpenInterpreter ▷ #O1 (6 messages):

  • OpenInterpreter Brand Guidelines

OpenInterpreter ▷ #ai-content (4 messages):

  • Zed AI
  • Open Source
  • Cursor
  • AI Code Editor
  • Claude-3.5

Links mentioned:


DSPy ▷ #papers (2 messages):

  • Apple's ML-Superposition Prompting

Link mentioned: GitHub - apple/ml-superposition-prompting: Contribute to apple/ml-superposition-prompting development by creating an account on GitHub.


DSPy ▷ #general (18 messages🔥):

  • Typed outputs
  • DSPy's prompt length
  • Truncated outputs

Links mentioned:


DSPy ▷ #colbert (1 messages):

  • ColBERT model for German language
  • ColBERTv2 data format

Gorilla LLM (Berkeley Function Calling) ▷ #leaderboard (5 messages):

  • Hugging Face Leaderboard
  • BFCL V2-Live Dataset
  • Model Evaluation and Uploads to BFCL

Link mentioned: [BFCL] How should we calculate the overall accuracy for V2-Live dataset? · ShishirPatil/gorilla · Discussion #602: Would love to hear the thoughts from the community on this matter: BFCL V2 • Live dataset features a diverse set of 2,251 question-function-answer pairs. It comprises of 258 simple, 7 multiple, 16 ...


Gorilla LLM (Berkeley Function Calling) ▷ #discussion (2 messages):

  • Gorilla Function Calling
  • Gorilla Leaderboard
  • Contributing to Gorilla
  • Executable Test Pairs

Link mentioned: gorilla/berkeley-function-call-leaderboard at main · ShishirPatil/gorilla: Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls) - ShishirPatil/gorilla


LAION ▷ #general (4 messages):

  • Anthropic's mechanistic interpretability
  • Llama 8b
  • Mistral
  • Open Source Implementations
  • AI Engineer London Meetup

Link mentioned: Tweet from Damien C. Tanner (@dctanner): We're brining a slice of @swyx's AI Engineer World's Fair to London! Evening of 12 September is the first AI Engineer London Meetup. Hear from 4 amazing speakers: @maximelabonne, @rovio...


Interconnects (Nathan Lambert) ▷ #random (3 messages):

  • OpenAI DevRel
  • Logan's Departure
  • Romain Huet

Link mentioned: Tweet from undefined: no description found


Alignment Lab AI ▷ #general (1 messages):

  • AI Engineer London Meetup

Link mentioned: Tweet from Damien C. Tanner (@dctanner): We're brining a slice of @swyx's AI Engineer World's Fair to London! Evening of 12 September is the first AI Engineer London Meetup. Hear from 4 amazing speakers: @maximelabonne, @rovio...


LLM Finetuning (Hamel + Dan) ▷ #general (1 messages):

rolandtannous: hello is Hamel around?


MLOps @Chipro ▷ #events (1 messages):

  • CUDA Hackathon
  • NVIDIA Engineers
  • Accelerated Computing

DiscoResearch ▷ #general (1 messages):

  • Together API pricing change
  • Together API 8B and 70B price increases
  • Together API pricing versus OpenAI pricing
  • Together AI Funding

Link mentioned: Announcement of pricing changes to Together's serverless API: Together serverless API pricing changes



{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}