Frozen AI News archive

DocETL: Agentic Query Rewriting and Evaluation for Complex Document Processing

**UC Berkeley's EPIC lab** introduces innovative LLM data operators with projects like **LOTUS** and **DocETL**, focusing on effective programming and computation over large data corpora. This approach contrasts GPU-rich big labs like **Deepmind** and **OpenAI** with GPU-poor compound AI systems. **Microsoft** open-sourced **BitNet b1.58**, a 1-bit ternary parameter LLM enabling **4-20x faster training** and on-device inference at human reading speeds. Nvidia released **Llama-3.1-Nemotron-70B-Instruct**, a fine-tuned open-source model outperforming **GPT-4o** and **Claude-3.5-sonnet**. These developments highlight advances in **model-optimization**, **on-device-ai**, and **fine-tuning**.

Canonical issue URL

AI News for 10/18/2024-10/21/2024. We checked 7 subreddits, 433 Twitters and 32 Discords (231 channels, and 6066 messages) for you. Estimated reading time saved (at 200wpm): 791 minutes. You can now tag @smol_ai for AINews discussions!

We usually reserve the feature story of AINews for the single biggest impactful news item of the day, but that usually results in a heavy bias towards recapping press releases from big model labs. Other stories of the year develop gradually, more of a swell than a splash, and may not be as big but are still useful as part of a well diversified diet. We use quieter days like these to shed some cumulative light on community tools like DSPy and AI price cut stories.

UC Berkeley has been a leader in many of the biggest waves in tech - per David Patterson, the 40 year history of UCB research labs have spawned everything from RISC, RAID, and massive companies like Databricks. The latest lab in this tradition is EPIC — focused on Effective Programming, Interaction, and Computation with Data. We were fortunate to attend their recent conference and were particularly impressed by two similar papers, LOTUS and DocETL the latter of has been the subject of notable hype and was finally published today. Both offer some very well thought through LLM operators over large corpuses of data.

The github docs give more of idea of the proposed APIs and concepts, and at the limit this could be viewed as "just another LLM framework" similar to DSPy, but the big data focus at an institution known for successfully thinking about commercially relevant big data problems makes this one worth a closer look than the average twitter anon:

At the very highest level this is just the latest front in the ongoing battle between GPU Rich Big Labs (Deepmind, OpenAI) and GPU Poor Compound AI approaches to AI. The DocETL demo site helps you compare results and approaches between using their framework and "sticking it all in context". There will likely not be a clear winner here for a long time and AI Engineers will simply have to be familiar with both.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

AI Acceleration

AI Model Developments and Research

AI Applications and Tools

AI Ethics and Societal Impact

Memes and Humor


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Advancements in LLM Architecture and Training

Theme 2. Innovative LLM Frameworks and Tools for Developers

Theme 3. Local LLMs Outperforming Cloud Alternatives

Theme 4. IBM Granite 3.0: Open-Source LLMs with Full Commercial Use

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

AI Research and Techniques

AI Model Releases and Improvements

AI Applications and Implications

AI Safety and Ethics Concerns

AI Industry Developments


AI Discord Recap

A summary of Summaries of Summaries by O1-preview

Theme 1: AI Model Advances and New Releases

Theme 2: AI Safety and Ethical Concerns

Theme 3: Model Training Challenges and Optimization

Theme 4: AI Agent Frameworks and Applications

Theme 5: AI in Creative Content Generation


PART 1: High level Discord summaries

HuggingFace Discord


Notebook LM Discord Discord


Eleuther Discord


Unsloth AI (Daniel Han) Discord


Nous Research AI Discord


OpenAI Discord


Perplexity AI Discord


Modular (Mojo 🔥) Discord


Latent Space Discord


LM Studio Discord


OpenRouter (Alex Atallah) Discord


aider (Paul Gauthier) Discord


GPU MODE Discord


Interconnects (Nathan Lambert) Discord


Stability.ai (Stable Diffusion) Discord


LlamaIndex Discord


tinygrad (George Hotz) Discord


Cohere Discord


OpenAccess AI Collective (axolotl) Discord


Torchtune Discord


DSPy Discord


LLM Agents (Berkeley MOOC) Discord


LAION Discord


OpenInterpreter Discord


LangChain AI Discord


MLOps @Chipro Discord


Mozilla AI Discord


DiscoResearch Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Finetuning (Hamel + Dan) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Gorilla LLM (Berkeley Function Calling) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ▷ #announcements (1 messages):

  • HelpingAI2 Demo
  • Protein Structure Prediction
  • AI in Nuclear Research
  • WorldMedQA-V Release
  • Books Mixer AI

Links mentioned:


HuggingFace ▷ #general (839 messages🔥🔥🔥):

  • Hugging Face Issues
  • AI Model Capabilities
  • GPU Usage
  • Kaggle vs Colab
  • Synthetic Data Generation

Links mentioned:


HuggingFace ▷ #today-im-learning (27 messages🔥):

  • Modifying models
  • Learning Python
  • Nvidia L4 GPU cooling solutions
  • Deep Reinforcement Learning
  • Multihead Attention

Links mentioned:


HuggingFace ▷ #cool-finds (10 messages🔥):

  • LightRAG
  • CGPO
  • Min-p Sampling
  • Medical AI Research Highlights
  • Visual Question Answering Models

Links mentioned:


HuggingFace ▷ #i-made-this (33 messages🔥):

  • Text Classification Overview
  • AI Energy Consumption and Nuclear Power
  • OmniBench Benchmark Introduction
  • Emotional AI Interaction
  • Dataset Releases by Recursal

Links mentioned:


HuggingFace ▷ #core-announcements (1 messages):

  • Advanced Dreambooth LoRA Training Script
  • Flux Features
  • Community Contributions
  • Pivotal Tuning
  • Experimental Resource Updates

Link mentioned: Advanced Flux Dreambooth LoRA Training with 🧨 diffusers: no description found


HuggingFace ▷ #computer-vision (1 messages):

shan_raja: Website bounding box


HuggingFace ▷ #NLP (52 messages🔥):

  • NLP Resources
  • Model Performance Issues
  • Text Classification Feedback
  • Inference Speed Optimization

HuggingFace ▷ #diffusion-discussions (28 messages🔥):

  • NozyIO UI Project
  • Yolo Integration
  • Module Modularity in Diffusers
  • Diffuser Error Resolution

Links mentioned:


Notebook LM Discord ▷ #use-cases (250 messages🔥🔥):

  • Use cases for NotebookLM
  • Podcast generation
  • Analysis of texts
  • AI in education
  • Discord chat scraping

Links mentioned:


Notebook LM Discord ▷ #general (613 messages🔥🔥🔥):

  • NotebookLM functionality
  • AI podcast generation
  • User feedback on NotebookLM
  • Translation and language support
  • Creative uses of NotebookLM

Links mentioned:


Eleuther ▷ #general (201 messages🔥🔥):

  • Data Requirements in Open Source AI
  • Copyright Law and AI Training
  • AAAI vs ICLR Workshops
  • Open Source Model Definitions
  • Community Projects and Contributions

Links mentioned:


Eleuther ▷ #research (320 messages🔥🔥):

  • Selective Attention in Transformers
  • Diff Transformer
  • Weight Sharing in Attention Mechanisms
  • RWKV-7 Training Speed Record
  • Research Practices in Literature Review

Links mentioned:


Eleuther ▷ #interpretability-general (16 messages🔥):

  • SAE feature interpretations
  • Distribution shifts
  • Oversampling in SAE training
  • Language model explanations
  • Variability in OpenAI API models

Links mentioned:


Eleuther ▷ #lm-thunderdome (45 messages🔥):

  • Integration of eval harness
  • Challenges with custom models
  • Finding lm-evaluation-harness datasets
  • Open LLM leaderboard resources

Links mentioned:


Eleuther ▷ #gpt-neox-dev (32 messages🔥):

  • FP16 Hysteresis
  • Dynamic Loss Scaling in Pythia
  • Rotary Percent Configurations
  • Allgather and Reduce Bucket Sizes
  • BF16 and FP16 Training in Pythia Models

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (317 messages🔥🔥):

  • Unsloth AI
  • Gradient Accumulation Bug Fix
  • Training LLMs
  • Multimodal Support
  • Knowledge Graphs

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (1 messages):

foxhop.: https://x.com/RussellBal/status/1847989964992139699


Unsloth AI (Daniel Han) ▷ #help (102 messages🔥🔥):

  • Model Fine-Tuning Issues
  • Layer Freezing
  • Tokenization Errors
  • CUDA Memory Management
  • Multiple Target Column Predictions

Links mentioned:


Unsloth AI (Daniel Han) ▷ #community-collaboration (3 messages):

  • Training LLMs on new dataset formats
  • Freezing embeddings for special tokens
  • Challenges with memory efficiency in LLM training
  • Custom autograd functions for selective training

Link mentioned: no title found): no description found


Unsloth AI (Daniel Han) ▷ #research (11 messages🔥):

  • ReAct Agent Tool Calling
  • LayerSkip Inference
  • Self-Taught Evaluator
  • Meta Lingua Efficient Training
  • SPIRIT-LM Multimodal Model

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

  • Nous Video on Safety
  • Nous Blog Post on Safety

Link mentioned: Tweet from Nous Research (@NousResearch): no description found


Nous Research AI ▷ #general (229 messages🔥🔥):

  • AI Safety Concerns
  • Crypto Scams
  • Deepfake Issues
  • Nous Research Developments
  • Voice Generation Technology

Links mentioned:


Nous Research AI ▷ #ask-about-llms (90 messages🔥🔥):

  • Misguided Attention Prompts
  • Monty Hall Problem
  • Cognitive Biases in LLMs
  • LLM Training and Reasoning
  • Hermes Performance

Links mentioned:


Nous Research AI ▷ #research-papers (31 messages🔥):

  • Model Efficiency in LLMs
  • Importance of Token-Level Learning
  • Recent Advances in Optimizers
  • Medical AI Developments
  • Exploration of New Dataset Models

Links mentioned:


Nous Research AI ▷ #interesting-links (18 messages🔥):

  • SCP Generator
  • OpenAI AGI Loophole
  • Meta FAIR Research
  • Segment Anything Model
  • Hermes AI Model

Links mentioned:


Nous Research AI ▷ #research-papers (31 messages🔥):

  • Model Efficiency in Language Models
  • Medical AI Research Highlights
  • Synthetic Data for AI Training
  • Advancements in Optimizers
  • Cross-lingual Sentence Encoders

Links mentioned:


Nous Research AI ▷ #reasoning-tasks (2 messages):

  • MarketAgents Project
  • Multi-Agent Market Simulation

Link mentioned: GitHub - marketagents-ai/MarketAgents: A distributed agent orchestration framework for market agents: A distributed agent orchestration framework for market agents - marketagents-ai/MarketAgents


OpenAI ▷ #ai-discussions (260 messages🔥🔥):

  • O1 Preview Performance
  • AI in Programming
  • OpenAI Alternatives
  • Understanding AI Predictions
  • Challenges with Current AI Models

Links mentioned:


OpenAI ▷ #gpt-4-discussions (21 messages🔥):

  • ChatGPT Memory Issues
  • Custom GPT Activation
  • YouTube GPT API Errors

OpenAI ▷ #prompt-engineering (27 messages🔥):

  • Tips for Using AI Prompts
  • Improving Realism in AI Conversations
  • Weights in Prompts for AI Responses
  • Performance of ChatGPT Models
  • User Experience in AI Role-Playing

OpenAI ▷ #api-discussions (27 messages🔥):

  • Improving AI realism
  • Prompt techniques for API
  • Adjusting AI responses
  • Role-playing with AI
  • Parameter weighting in prompts

Perplexity AI ▷ #general (286 messages🔥🔥):

  • Perplexity Pro Limitations
  • User Experiences with Perplexity
  • AI Model Discussions
  • Collaboration Tools
  • Pricing and Subscription Issues

Links mentioned:


Perplexity AI ▷ #sharing (24 messages🔥):

  • Best Consoles of 2023
  • Oldest City in Russia
  • AI Content Identification by YouTube
  • Cool Hangout Spots in Kuala Lumpur
  • Reliance Industries Stock Recommendation

Perplexity AI ▷ #pplx-api (6 messages):

  • Sonar-online models performance
  • API credits issues
  • API for spaces feature

Modular (Mojo 🔥) ▷ #general (19 messages🔥):

  • Mojo Programming Language
  • Mojo vs C++ and Python
  • Carbon Programming Language
  • GPU Architecture Video
  • Using TensorFlow and PyTorch with Mojo

Links mentioned:


Modular (Mojo 🔥) ▷ #mojo (248 messages🔥🔥):

  • Mojo Reference Handling
  • Performance Optimization in Mojo
  • Tuple Lengths at Compile Time
  • Error Handling in Mojo
  • Usage of Async/Await in Mojo

Links mentioned:


Modular (Mojo 🔥) ▷ #max (2 messages):

  • Graph Training Support
  • C-API Model Execution

Latent Space ▷ #ai-general-chat (133 messages🔥🔥):

  • DeepSeek Janus
  • Meta Spirit LM
  • Microsoft Copilot Agents
  • AI Reply Bots
  • IBM Granite 3.0

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

  • AI policies in Singapore
  • Government adoption of AI
  • Sovereign AI approaches
  • AI in election season

Link mentioned: Tweet from swyx (@swyx): 🆕 @latentspacepod is proud to present: Building the AI Engineer Nation https://latent.space/p/josephine-teo A special conversation with @joteo_ylm, our first with a sitting member of Cabinet �...


Latent Space ▷ #ai-in-action-club (133 messages🔥🔥):

  • AST vs DSL
  • Code Transformation Techniques
  • BAML DSL
  • Compiler Education
  • Leveraging LLMs for Programming

Links mentioned:


LM Studio ▷ #general (221 messages🔥🔥):

  • Model Performance Comparisons
  • Troubleshooting LM Studio
  • Vision Model Capabilities
  • Settings for Image Input
  • Backup and Recovery in LM Studio

Links mentioned:


LM Studio ▷ #hardware-discussion (30 messages🔥):

  • Xeon Processor Configurations
  • RX 7900 XTX Performance
  • RX 6600 Vulkan vs ROCm
  • M4 Ultra Chip for AI Tasks

OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

  • Inflection Payment Issues
  • Grok Beta Rename
  • Grok Pricing Increase
  • Liquid LFM Pricing Updates

OpenRouter (Alex Atallah) ▷ #app-showcase (3 messages):

  • AI powered text summarizer
  • Vercel function timeout
  • OpenRouter API response time
  • Streaming responses
  • Alternative models

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (225 messages🔥🔥):

  • OpenRouter model issues
  • Grok 2
  • Hermes 3
  • Billing problems
  • AI model capabilities

Links mentioned:


OpenRouter (Alex Atallah) ▷ #beta-feedback (3 messages):

  • Custom Provider Keys
  • Self-Service Integration Sign Up

aider (Paul Gauthier) ▷ #general (126 messages🔥🔥):

  • Durable Execution Concepts
  • Aider and VSCode Integration
  • Mistral API Usage
  • CEDARScript Runtime
  • Hello World Refactoring Issues

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (56 messages🔥🔥):

  • Aider Usage with Sonnet and Claude
  • Managing Aider's Auto Commit Features
  • File Creation and Existence Issues in Aider
  • Leveraging Aider History for Context
  • Setting Main and Weak Models in Aider

Links mentioned:


aider (Paul Gauthier) ▷ #links (1 messages):

  • bitnet.cpp
  • 1-bit LLMs
  • Inference performance on ARM and x86 CPUs

GPU MODE ▷ #general (30 messages🔥):

  • TensorRT-LLM Code Sharing
  • Unsloth Lecture and Resources
  • GPU MODE Talk Recording
  • Event Scheduling Inquiry
  • Distributed Training Framework Comparisons

Links mentioned:


GPU MODE ▷ #triton (5 messages):

  • Triton and PTX compatibility
  • Windows support for Triton
  • Interaction between torch.compile and Triton

Links mentioned:


GPU MODE ▷ #torch (11 messages🔥):

  • Torch Distributions in LibTorch
  • New PyTorch Environment Variable
  • Clearing Compiled Ops in PyTorch
  • Behavior of Autocast in PyTorch
  • DDP Training Issues with ResNet50

Links mentioned:


GPU MODE ▷ #announcements (1 messages):

  • Han Brothers
  • Unsloth Presentation
  • Triton Tricks
  • CUDA Techniques

GPU MODE ▷ #cool-links (6 messages):

  • Domino Communication Scheme
  • Torchtitan Library
  • Large Language Model Training

Links mentioned:


GPU MODE ▷ #jobs (1 messages):

  • Hiring GPU Programmers
  • Decentralized AI
  • Tokens per second improvement

GPU MODE ▷ #beginner (40 messages🔥):

  • Flash Attention Multiplication
  • LlamaCPP and GGML library usage
  • Raspberry Pi graphics performance
  • Triton and CUDA compatibility
  • Debugging Multihead Attention in models

Links mentioned:


GPU MODE ▷ #pmpp-book (4 messages):

  • Chapter 4 exercises
  • Occupancy calculation

Link mentioned: PMPP_notes/4_GPU_Architecture/exercises.md at main · mandliya/PMPP_notes: Notes and code for Programming Massively Parallel Processors - mandliya/PMPP_notes


GPU MODE ▷ #youtube-recordings (1 messages):

gau.nernst: https://www.youtube.com/watch?v=hfb_AIhDYnA


GPU MODE ▷ #torchao (2 messages):

  • GitHub Issues
  • PyTorch Release Performance

Link mentioned: torch 2.5 slower than 2.4.1 ? · Issue #138386 · pytorch/pytorch: 🐛 Describe the bug I noticed that the latest stable release 2.5.0 is slower than 2.4.1 when using torch.compile (reduce-overhead), I tried on different machines with a 4090 RTX and it's pretty mu...


GPU MODE ▷ #off-topic (6 messages):

  • Daniel Hanchen's talk recording
  • Inventec CXL Box
  • Micron GPU architecture
  • Iceberg lettuce salad recipe
  • Lecture 32: Unsloth

Links mentioned:


GPU MODE ▷ #irl-meetup (6 messages):

  • Sydney Meetup Coordination
  • NeurIPS Conference Participation
  • NeurIPS Location

GPU MODE ▷ #triton-puzzles (1 messages):

seahorse0180: Also ran into this issue just now.


GPU MODE ▷ #llmdotc (2 messages):

  • Parallel Prefix Sum Algorithm
  • Mamba Training
  • SSMs
  • Linear RNNs
  • llm.c Repository

GPU MODE ▷ #rocm (2 messages):

  • Improvement sources for MI300X
  • RCCL tuning issues
  • Performance in single node RCCL
  • Async TP kernels challenges

Link mentioned: rccl/src/graph/tuning.cc at develop · ROCm/rccl: ROCm Communication Collectives Library (RCCL). Contribute to ROCm/rccl development by creating an account on GitHub.


GPU MODE ▷ #sparsity-pruning (3 messages):

  • Activation Sparsity Tools
  • PyTorch Sparse Functionality
  • PowerInfer Research
  • Sparse Kernel Implementations

Link mentioned: Issues · pytorch/ao: PyTorch native quantization and sparsity for training and inference - Issues · pytorch/ao


GPU MODE ▷ #liger-kernel (7 messages):

  • Acknowledgement of Contributors in Liger Arxiv Whitepaper
  • Gradient Accumulation Bug in Liger Kernel
  • Memory Issues with Triton and Liger Operations
  • Calls for Code Review on Gradient Accumulation
  • Updates on Liger Kernel Documentation

Links mentioned:


GPU MODE ▷ #metal (9 messages🔥):

  • Objective-C language server
  • C/C++ memory management with PyTorch
  • MPS stream in PyTorch
  • Unified memory on Apple Silicon
  • MTLCommandQueue functionality

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (12 messages🔥):

  • Human Data Labeling for Geospatial Data
  • Platforms for Data Labeling
  • Offshore Vendors for Visual Data
  • Scale AI and Alternatives

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (74 messages🔥🔥):

  • Emoji Requests
  • RLHF Book Development
  • OpenAI Token Usage
  • CARDS Alignment Method
  • Dark Mode Discussion

Links mentioned:


Interconnects (Nathan Lambert) ▷ #rlhf (16 messages🔥):

  • Interconnects Emojis
  • Discord Support
  • Emoji Uploading
  • AI Company Logos

Interconnects (Nathan Lambert) ▷ #reads (2 messages):

  • LLM Reasoning Debate
  • OpenAI's GPT Releases
  • Training Data Limitations

Link mentioned: The LLM Reasoning Debate Heats Up : Three recent papers examine the robustness of reasoning and problem-solving in large language models


Stability.ai (Stable Diffusion) ▷ #general-chat (98 messages🔥🔥):

  • Performance of RTX GPUs
  • Creating Images with Different Perspectives
  • Using Loras in Prompts
  • Stable Diffusion API Access Issues
  • Need for Assistance in Image Editing

Links mentioned:


LlamaIndex ▷ #blog (7 messages):

  • 3-day hackathon
  • LlamaParse Premium
  • Agentic System for Automated Sales Outreach
  • Advanced RAG Workflow
  • Multimodal RAG Pipeline

LlamaIndex ▷ #general (81 messages🔥🔥):

  • Ollama Integration in LlamaIndex
  • Evaluating Retrieval Methods
  • Event Streaming in Workflows
  • Document Summarization Techniques
  • Deployment Platforms for ML Models

Links mentioned:


LlamaIndex ▷ #ai-discussion (3 messages):

  • Multilingual Embedding Models
  • Creating API for Proprietary Materials

tinygrad (George Hotz) ▷ #general (55 messages🔥🔥):

  • Multihead Attention Standardization
  • Tinygrad Development Updates
  • WebGPU Support
  • Local LLM Usage Trends
  • Benchmark CI Testing

Link mentioned: Big graph · Issue #7044 · tinygrad/tinygrad: LazyBuffer.view becomes UOps.VIEW #7077 #7078 #7007 #7090 big graph SINK #7122, #7178, #7170 #7134, #7175 #7132, #7188 #7190 #7149 ASSIGN and toposort become graph_rewrite deciding when to realize ...


tinygrad (George Hotz) ▷ #learn-tinygrad (35 messages🔥):

  • FrozenBatchNorm2d
  • Auto-generating classes from .safetensors
  • Action Chunking Transformers in Tinygrad
  • Implementing nonzero in Tinygrad
  • Context Manager for default float in Tinygrad

Links mentioned:


Cohere ▷ #discussions (35 messages🔥):

  • Mystery model
  • Agent assist APIs
  • Connection issues with Google Drive
  • Community introductions
  • General chat discussions

Link mentioned: Tweet from UltraIA (@Ultra_IA): LOL


Cohere ▷ #announcements (2 messages):

  • Aya Community Project
  • Cohere Developer Office Hours

Cohere ▷ #questions (35 messages🔥):

  • OpenRouter Benefits
  • Cohere API Usage
  • Langchain SSL Issues

Link mentioned: Chat — Cohere: Generates a message from the model in response to a provided conversation. To learn more about the features of the Chat API follow our [Text Generation guides](https://docs.cohere.com/v2/docs/chat-api...


Cohere ▷ #api-discussions (6 messages):

  • API read timeout issues
  • Getting citations from the API
  • Chat API documentation

Links mentioned:


Cohere ▷ #projects (2 messages):

  • JavaScript Implementations
  • Direct API Requests

OpenAccess AI Collective (axolotl) ▷ #general (44 messages🔥):

  • Liger Kernel Installation
  • Axolotl Layer Freezing Issue
  • SNR Results for Spectrum
  • AGI House Events

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (2 messages):

  • Qwen2 support for DoRA/QDoRA
  • Answer.AI's QDoRA repo

Link mentioned: GitHub - AnswerDotAI/fsdp_qlora: Training LLMs with QLoRA + FSDP: Training LLMs with QLoRA + FSDP. Contribute to AnswerDotAI/fsdp_qlora development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #general-help (1 messages):

  • Training Domain-Specific LLMs
  • Fine-tuning LLMs
  • Instruct Models

Torchtune ▷ #general (38 messages🔥):

  • Meta's FAIR Research
  • Attention Mask Issues in Torch
  • Flex Attention Challenges
  • Performance Warnings in PyTorch
  • Mask Construction Discussions

Links mentioned:


Torchtune ▷ #dev (1 messages):

  • v0.4.0 code freeze
  • New features in v0.4.0
  • Release timeline

Link mentioned: v0.4.0 Tracker · Issue #1747 · pytorch/torchtune: Estimated release date: Tuesday, November 5th Estimated branch cut date (aka code freeze): Tuesday, October 29th Release owner: @joecummings New features: #1645 #1847 (@felipemello1) #1835 (@Optimo...


DSPy ▷ #show-and-tell (11 messages🔥):

  • Pydantic All-in-One
  • DSPy GPTs
  • AI Agents in Production Event
  • Streaming and Bot Notifications
  • HotpotQA Alternate History Generator

Links mentioned:


DSPy ▷ #papers (2 messages):

  • LightRAG tutorial
  • GraphRAG observations
  • Ollama integration
  • R2R insights
  • Microsoft's local search

Link mentioned: Local LightRAG: A GraphRAG Alternative but Fully Local with Ollama: In this video, we explore how to set up and run LightRAG—a retrieval augmented generation (RAG) system that combines knowledge graphs with embedding-based re...


DSPy ▷ #general (18 messages🔥):

  • DSPy with Hugging Face models
  • Ollama usage for model deployment
  • AGI House hackathons
  • Building LLMs using DSPy
  • SGLang for model inference

Links mentioned:


DSPy ▷ #examples (3 messages):

  • Hosting models on Hugging Face
  • Running DSPy modules

DSPy ▷ #colbert (3 messages):

  • AcgNDCG pseudo-function
  • BM25 retriever inquiry
  • AvgNDCG DSPy Metric
  • PATH first author outreach

LLM Agents (Berkeley MOOC) ▷ #mooc-announcements (1 messages):

  • Lecture 7
  • TapeAgents framework
  • WorkArena++ benchmark
  • Nicolas Chapados

Link mentioned: CS 194/294-196 (LLM Agents) - Lecture 7: no description found


LLM Agents (Berkeley MOOC) ▷ #mooc-questions (33 messages🔥):

  • Certification for Course Completion
  • Project Development Strategies
  • Hackathon Participation
  • Written Article Assignment
  • Local LLM Running Options

Links mentioned:


LLM Agents (Berkeley MOOC) ▷ #mooc-lecture-discussion (3 messages):

  • Orchestration of agents
  • Lecture timing

LAION ▷ #general (4 messages):

  • LibreFLUX release
  • FLUX.1-schnell comparison
  • Open source characteristics
  • Community reactions

Link mentioned: jimmycarter/LibreFLUX · Hugging Face: no description found


LAION ▷ #research (12 messages🔥):

  • Open-MUSE training issues
  • Microsoft LLM breakthrough
  • BitNet model
  • Training logs for MUSE project

Links mentioned:


OpenInterpreter ▷ #general (10 messages🔥):

  • Aider incremental improvements
  • Open Interpreter equivalent of /functions folder
  • Custom tools support
  • Python virtual environments
  • Integrating voice assistants

Link mentioned: Tweet from Jacob@AIwithBenefits (@AIwithBenefits): Added a @hume_ai voice assistant to the @phidatahq generalist agent, and a little help from the @OpenInterpreter system message. Quoting Jacob@AIwithBenefits (@AIwithBenefits) Loving the new @phid...


OpenInterpreter ▷ #O1 (1 messages):

  • OpenInterpreter Mac setup
  • Interaction issues
  • LiveKit Meet link concerns

LangChain AI ▷ #general (9 messages🔥):

  • LangGraph Code Assistant
  • Role-based RAG Models
  • Expected Context Issues
  • Techstars Startup Weekend Event
  • Code Generation Approaches

Links mentioned:


LangChain AI ▷ #share-your-work (1 messages):

  • OpenAI Swarm
  • LangChain LangGraph
  • Multi-Agent Frameworks

Link mentioned: OpenAI Swarm vs LangChain LangGraph: A Detailed Look at Multi-Agent Frameworks: Ankush k Singal


MLOps @Chipro ▷ #events (1 messages):

huikang: https://app.agihouse.org/events/agi-thon-werewolf-agents-tournament-20241109


Mozilla AI ▷ #announcements (1 messages):

  • AI access challenges
  • Competition in AI
  • External researcher access
  • Big Tech and AI
  • Open AI ecosystem

DiscoResearch ▷ #general (1 messages):

huunguyen: has anyone tried q-galora?





{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}