Frozen AI News archive

Claude 3.5 Sonnet (New) gets Computer Use

**Anthropic** announced new Claude 3.5 models: **3.5 Sonnet** and **3.5 Haiku**, improving coding performance significantly, with Sonnet topping several coding benchmarks like **Aider** and **Vectara**. The new **Computer Use API** enables controlling computers via vision, scoring notably higher than other AI systems, showcasing progress in AI-driven computer interaction. **Zep** launched a cloud edition for AI agents memory management, highlighting challenges in **multimodal memory**. The update also mentions **Llama 3.1** and **Nemotron** models from **NVIDIA**.

Canonical issue URL

AI News for 10/21/2024-10/22/2024. We checked 7 subreddits, 433 Twitters and 32 Discords (232 channels, and 3347 messages) for you. Estimated reading time saved (at 200wpm): 341 minutes. You can now tag @smol_ai for AINews discussions!

Instead of the widely anticipated (and now indefinitely postponed) Claude 3.5 Opus, Anthropic announced a new 3.5 Sonnet, and 3.5 Haiku, bringing a bump to each model.

image.png

3.5 Sonnet, already delivers significant gains in coding. The new 3.5 Haiku (with benchmarks on the model card) matches the performance of Claude 3 Opus “on many evaluations for the same cost and similar speed to the previous generation of Haiku”.

image.png Notably on coding, it improves performance on SWE-bench Verified from 33.4% to 49.0%, scoring HIGHER than o1-preview's own 41.4% without any fancy reasoning steps. However, on math on 3.5 Sonne's 27.6% high water mark still pales in comparison to o1-preview's 83%.

Other Benchmarks:

Computer Use

Anthropic's new Computer Use API (docs here, demo here) point to OSWorld for their relevant screen manipulation benchmark - scoring 14.9% in the screenshot-only category—notably better than the next-best AI system's score of 7.8%.

image.png

When afforded more steps to complete the task, Claude scored 22.0%. This is still substantially under human performance in the 70’s, but is notable because this is essentially the functionality that Adept previously announced with its Fuyu models but never widely released. In a reductive sense "computer use" (controlling a computer via vision) is contrasted against standard "tool use" (controlling computers via API/function calling).

Example Videos:

Vendor Request Form, Coding Via Vision, Google Searches and Google Maps

Simon Willison kicked the tires on the github quickstart further with tests including compile and run hello world in C (it has gcc already so this just worked) and installing missing Ubuntu packages.

Replit was also able to plug in Claude as a human feedback replacement for @Replit Agent.


[Sponsored by Zep] Zep just launched their cloud edition today! Zep is a low-latency memory layer for AI agents and assistants that can reason about facts that change over time. Jump into the Discord to chat the future of knowledge graphs and memory!

swyx commentary: with computer use now officially blessed by Claude's upgraded vision model, how will agent memory storage need to change? You can see the simplistic image memory implementation from Anthropic but there's no answer for multimodal memory yet... one hot topic for the Zep Discord.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

AI Model Updates and Releases

AI Research and Techniques

AI Tools and Applications

AI Ethics and Societal Impact

Memes and Humor


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Moonshine: New Open-Source Speech-to-Text Model Challenges Whisper

Theme 2. Allegro: New State-of-the-Art Open-Source Text-to-Video Model

Theme 3. AI Sabotage Incident at ByteDance Raises Security Concerns

Theme 4. PocketPal AI: Open-Source App for Local Models on Mobile

Theme 5. Trend Towards More Restrictive Licenses for Open-Weight AI Models

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

AI Model Developments and Releases

AI Research and Industry Insights

AI Ethics and Societal Impact

Robotics Advancements

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries by O1-preview

Theme 1. Claude 3.5 Breaks New Ground with Computer Use

Theme 2. Stable Diffusion 3.5 Lights Up AI Art

Theme 3. AI Video Generation Heats Up with Mochi 1 and Allegro

Theme 4. Cohere Embeds Images into Multimodal Search

Theme 5. Hackathon Fever: Over $200k in Prizes from Berkeley


PART 1: High level Discord summaries

HuggingFace Discord


OpenRouter (Alex Atallah) Discord


aider (Paul Gauthier) Discord


Stability.ai (Stable Diffusion) Discord


Unsloth AI (Daniel Han) Discord


Nous Research AI Discord


LM Studio Discord


Latent Space Discord


Notebook LM Discord Discord


Perplexity AI Discord


Eleuther Discord


Interconnects (Nathan Lambert) Discord


GPU MODE Discord


OpenAI Discord


Cohere Discord


Modular (Mojo 🔥) Discord


tinygrad (George Hotz) Discord


OpenInterpreter Discord


DSPy Discord


LlamaIndex Discord


LLM Agents (Berkeley MOOC) Discord


Torchtune Discord


LangChain AI Discord


OpenAccess AI Collective (axolotl) Discord


Gorilla LLM (Berkeley Function Calling) Discord


LAION Discord


Mozilla AI Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Finetuning (Hamel + Dan) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ▷ #general (586 messages🔥🔥🔥):

  • AI DJ Software
  • Hugging Face Model Queries
  • OCR Tools
  • Training TTS Models
  • Structured Output in LLMs

Links mentioned:


HuggingFace ▷ #today-im-learning (18 messages🔥):

  • 2021 lecture series
  • Creating virtual characters
  • Path to becoming an ML Engineer
  • 3blue1brown's educational resources
  • Manim animation engine

Links mentioned:


HuggingFace ▷ #cool-finds (1 messages):

capetownbali: Nice find...


HuggingFace ▷ #i-made-this (10 messages🔥):

  • Granite 3.0 model release
  • Webinar on LLM Best Practices
  • Evolution of Contextual Embeddings
  • ZK Proofs for Chat History Ownership
  • PR Merged for HuggingFace.js

Links mentioned:


HuggingFace ▷ #core-announcements (1 messages):

sayakpaul: <@&1014517792550166630> enjoy:
https://huggingface.co/blog/sd3-5


HuggingFace ▷ #NLP (8 messages🔥):

  • Tensor conversion bottleneck
  • Dataset device bottleneck
  • CPU and GPU usage during inference
  • Evaluating fine-tuned LLMs
  • Managing evaluation results

Link mentioned: Annoyed Cat GIF - Annoyed cat - Discover & Share GIFs: Click to view the GIF


HuggingFace ▷ #diffusion-discussions (29 messages🔥):

  • Kaggle and GPU Usage
  • Model Downloading Techniques
  • Learning Rate and Training Insights
  • Diffusers Callbacks for Image Generation
  • Cultural Connections in AI

Links mentioned:


OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

  • Claude 3.5 Sonnet
  • Llama 3.1 Nitro
  • Ministral updates
  • Grok Beta
  • Claude self-moderated endpoints

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (455 messages🔥🔥🔥):

  • New Claude 3.5 Sonnet
  • OpenRouter API
  • Computer Use feature
  • Model pricing
  • Haiku 3.5 release

Links mentioned:


aider (Paul Gauthier) ▷ #general (290 messages🔥🔥):

  • Claude 3.5 Sonnet
  • DeepSeek as Editor Model
  • Pricing for Models
  • Model Performance and Benchmarks
  • Integration of Local Models

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (20 messages🔥):

  • Aider Configuration
  • Aider Command Issues
  • Architect Mode Functionality
  • Mistral API Authentication
  • Reddit and Claude AI Discussions

Links mentioned:


aider (Paul Gauthier) ▷ #links (5 messages):

  • Claude 3.5 Sonnet upgrades
  • Claude 3.5 Haiku introduction
  • Computer use capability
  • DreamCut AI video editor

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #announcements (1 messages):

  • Stable Diffusion 3.5 Launch
  • Performance of Stable Diffusion 3.5 Large
  • Stable Diffusion 3.5 Large Turbo
  • Community Feedback
  • Accessibility of New Models

Link mentioned: Stable Diffusion 3.5 — Stability AI: Today we are introducing Stable Diffusion 3.5. This open release includes multiple model variants, including Stable Diffusion 3.5 Large and Stable Diffusion 3.5 Large Turbo.


Stability.ai (Stable Diffusion) ▷ #general-chat (280 messages🔥🔥):

  • Stable Diffusion 3.5 Release
  • Performance Comparisons with Flux
  • Model Licensing
  • Technical Support for SD3.5
  • Applications of LoRA in AI Art

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (234 messages🔥🔥):

  • Gradient Accumulation Bug Fix
  • LLM Training Efficiency
  • Model Performance and Benchmarking
  • Collaboration with Meta
  • Finetuning Strategies

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (34 messages🔥):

  • Grad School Application Editor
  • LLaMA Model Fine-Tuning
  • Unsloth Installation Issues
  • Multi-GPU Support in Unsloth
  • CUDA and Library Errors

Links mentioned:


Unsloth AI (Daniel Han) ▷ #community-collaboration (1 messages):

  • Unsloth Studio Fixes
  • GitHub Pull Request
  • Discord Issue Reporting

Link mentioned: Fix/studio by Erland366 · Pull Request #1 · unslothai/unsloth-studio: There are several issue in the studio. The issue was issued by user in the discord. This issue is trigger when importing unsloth, but somehow the issue didn't happen inside finetune notebook....


Nous Research AI ▷ #general (137 messages🔥🔥):

  • Catastrophic Forgetting in Fine-Tuning
  • Performance of LLMs on Benchmarks
  • Nous Research Video and Projects
  • Claude Model Updates
  • Token as a Service Providers

Links mentioned:


Nous Research AI ▷ #ask-about-llms (5 messages):

  • Hermes 3 availability
  • Claude system prompt enhancements
  • Claude's problem-solving capabilities

Nous Research AI ▷ #research-papers (11 messages🔥):

  • Research Paper Trustworthiness
  • Fine-Tuning Models
  • Falsification in Scientific Research
  • Simple Arithmetic for Language Models
  • AdamW Optimization Techniques

Links mentioned:


Nous Research AI ▷ #interesting-links (15 messages🔥):

  • Poe multi-model access
  • Machine Talks on generative art
  • New ASR model release
  • ZK proofs for chat history ownership
  • Mina vs. OpenBlock comparison

Links mentioned:


Nous Research AI ▷ #research-papers (11 messages🔥):

  • Research Paper Trustworthiness
  • Falsification of Scientific Data
  • Peer Review Concerns
  • Fine-tuning Models
  • Efficient Computation in Neural Networks

Links mentioned:


LM Studio ▷ #announcements (1 messages):

  • LM Studio v0.3.5 features
  • Headless mode
  • On-demand model loading
  • Pixtral support on Apple MLX
  • New CLI command to download models

Links mentioned:


LM Studio ▷ #general (171 messages🔥🔥):

  • GPU Offloading Issues
  • Model Loading Errors
  • AI Model Performance Metrics
  • ML Studio Features and Settings
  • Game Image Enhancers

Links mentioned:


Latent Space ▷ #ai-general-chat (155 messages🔥🔥):

  • Anthropic Claude 3.5
  • Mochi 1 Video Generation
  • CrewAI Series A Fundraising
  • Stable Diffusion 3.5 Release
  • Outlines Library Rust Port

Links mentioned:


Notebook LM Discord ▷ #use-cases (45 messages🔥):

  • NotebookLM Experiments
  • Podcast Creation
  • Language Learning
  • WallStreetBets Analysis
  • AI-generated Content

Links mentioned:


Notebook LM Discord ▷ #general (103 messages🔥🔥):

  • Issues with NotebookLM Language Settings
  • Sharing and Collaboration Challenges
  • Multilingual Audio Overviews
  • Podcasting Experiences
  • Document Upload Issues

Links mentioned:


Perplexity AI ▷ #general (103 messages🔥🔥):

  • New AI models and features
  • API functionality and user concerns
  • Perplexity competition and market presence
  • User experiences and feedback
  • Support and functionality inquiries

Link mentioned: Tweet from Anthropic (@AnthropicAI): Introducing an upgraded Claude 3.5 Sonnet, and a new model, Claude 3.5 Haiku. We’re also introducing a new capability in beta: computer use. Developers can now direct Claude to use computers the way ...


Perplexity AI ▷ #sharing (11 messages🔥):

  • College Paths and Degrees
  • Snapdragon 8 Elite Overview
  • Galaxy Z Fold Special Edition
  • Gold Prices Rise
  • AI-Driven Fact-Checking

Eleuther ▷ #announcements (1 messages):

  • SAE interpretation pipeline
  • Evaluation techniques for explanations
  • Causal feature explanation
  • Feature alignment using Hungarian algorithm
  • Open-source tools for LLMs

Links mentioned:


Eleuther ▷ #general (28 messages🔥):

  • Non-archival workshops
  • Chess AI model integration
  • Chess move explainability
  • Stockfish analysis speed
  • Research goals in AI development

Link mentioned: How many positions per second should my homemade engine calculate?: My program prints time spent on executing a function for doing/retracting move, and both take together an average of 00.0002 seconds. That means my engine can analyze at most 5000 positions per sec...


Eleuther ▷ #research (59 messages🔥🔥):

  • 1B context length with RAG
  • Robustness of SAEs
  • LayerNorm in transformer models
  • Independent research publication experiences
  • Ethics in research sharing

Links mentioned:


Eleuther ▷ #interpretability-general (2 messages):

  • Mech Interp Paper Ratings
  • Sharing Research
  • Twitter Presence

Link mentioned: Tweet from Alice Rigg (@woog09): I rated ALL mech interp papers submitted to ICLR 2025: https://docs.google.com/spreadsheets/d/1TTHbONFo4OV35Bv0KfEFllnkP-aLGrr_fmzwfdBqBY0/edit?gid=0#gid=0. The ratings are calibrated: 3 - outstanding...


Eleuther ▷ #lm-thunderdome (8 messages🔥):

  • Batch Size Configuration
  • Model Initialization Handling

Interconnects (Nathan Lambert) ▷ #news (74 messages🔥🔥):

  • Allegro Model Launch
  • Stability AI's Stable Diffusion 3.5
  • Anthropic's Claude 3.5 Haiku
  • Computer Use API
  • New Video Generation Models

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (7 messages):

  • AI-generated papers
  • Viral content on social media
  • Feedback mechanisms in tech communities

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (7 messages):

  • Factor 64
  • Blog readership
  • Reasoning tokens
  • CARDS method in LLMs

Link mentioned: Tweet from Ruqi Zhang (@ruqi_zhang): Introducing CARDS, a new method for LLM decoding-time alignment: ✨5x faster in text generation and 99% win-ties in GPT-4/Claude-3 evaluation ✨provably generates high-reward high-likelihood text ✨no r...


Interconnects (Nathan Lambert) ▷ #memes (2 messages):

  • Jeremy Howard's tweet
  • Tek's angry man arc

Link mentioned: Tweet from Jeremy Howard (@jeremyphoward): CEO of Microsoft getting ratioed by an anime pfp account...


GPU MODE ▷ #general (7 messages):

  • Unsloth Lecture Release
  • Gradient Accumulation Insights
  • GitHub AI Project
  • Engineering Tips Discussion

Links mentioned:


GPU MODE ▷ #triton (19 messages🔥):

  • Torch Compile Interpretation
  • Softplus Triton Kernel Optimization
  • Kernel Compilation Sources

GPU MODE ▷ #torch (1 messages):

  • Meta HOTI 2024
  • Llama 3 Infrastructure

Link mentioned: Day 2 10:00: Keynote: Powering Llama 3: Peek into Meta’s Massive Infrastructure for Generative AI: Speaker: Pavan Balaji (Meta)


GPU MODE ▷ #beginner (3 messages):

  • FA2 padded inputs
  • CUDA project ideas
  • CUDA accelerated regression

GPU MODE ▷ #torchao (14 messages🔥):

  • torchao v0.6.1 Release
  • Compatibility of torchao Optimizers with HF Trainer
  • Implementing Quantization Aware Training with Older Torch Versions
  • Dynamic Masking during Training in torchao

Links mentioned:


GPU MODE ▷ #llmdotc (1 messages):

apaz: They're working on llama3, but yeah.


GPU MODE ▷ #rocm (3 messages):

  • ROCm 6.2 Docker Image
  • GitHub Actions for AMD Cluster
  • Torch + ROCm with Poetry
  • Difference between ROCm and Official PyTorch Images
  • Job Queue Setup

Links mentioned:


GPU MODE ▷ #bitnet (7 messages):

  • Bitnet Implementation Weights
  • Packed Weights
  • Ternary Weights

GPU MODE ▷ #sparsity-pruning (1 messages):

  • TorchAO Sparsity Future Plans
  • Advancements in Sparsity & Pruning
  • Collaborative Opportunities

Link mentioned: [RFC] Sparsity Future Plans · Issue #1136 · pytorch/ao: I had a chance to reflect after PTC / CUDA-MODE and wanted to share some thoughts on future plans for sparsity in torchao. Current State There are two components of sparsity, accuracy and accelerat...


GPU MODE ▷ #liger-kernel (11 messages🔥):

  • Liger Kernel Inference
  • Cross Entropy Issues
  • Pull Request for Batch Norm
  • Transformers Monkey Patching
  • Loss Function References

Links mentioned:


GPU MODE ▷ #self-promotion (1 messages):

  • Model Quantization
  • IEEE IPTA Conference

Link mentioned: IPTA 2024 - Quantization Tutorial: Model Quantization Techniques for Efficient Transformer Models Dr. Hicham Badri Principal Research Scientist @Mobius Labs GmbH IEEE IPTA 2024 - Rabat, Morocco


GPU MODE ▷ #project-popcorn (6 messages):

  • LLM for Efficient Kernels
  • Scaling Test Time Compute
  • Kernel Dataset Competition
  • HidetScript DSL

Links mentioned:


OpenAI ▷ #ai-discussions (47 messages🔥):

  • AGI Challenges
  • Custom GPT Confusion
  • Quantum Computing Perspectives
  • Anthropic AI Releases
  • TV Show Recognition Issues

OpenAI ▷ #gpt-4-discussions (4 messages):

  • o1-preview usage limits
  • User onboarding in discussions

OpenAI ▷ #prompt-engineering (11 messages🔥):

  • Order of Context in Prompting
  • Error Correction on CSV Data
  • Problem-Solving with GPT
  • Emphasizing Prompt Details
  • Structured Prompts for Independent Thinking

OpenAI ▷ #api-discussions (11 messages🔥):

  • Importance of Context Order in Instructions
  • Using Table of Contents for Emphasis
  • Error Correction on CSV from Menu Photos
  • Structured Prompts for Problem Solving

Cohere ▷ #discussions (38 messages🔥):

  • Cohere models usage
  • Multi-modal embeddings
  • Cohere for AI + Embed - V3
  • Event scheduling issues
  • Performance tuning in LLMs

Link mentioned: Vsauce Michael GIF - Vsauce Michael Or Is It - Discover & Share GIFs: Click to view the GIF


Cohere ▷ #announcements (1 messages):

  • Multimodal Embed 3 Release
  • RAG Systems Integration
  • API Changes in Embed 3
  • Image Processing Enhancements

Links mentioned:


Cohere ▷ #questions (10 messages🔥):

  • LLM Model Fine-Tuning
  • Parallel Request Handling
  • Cohere Command R Features
  • Playground Usage in Professional Settings

Link mentioned: Starting the Chat Fine-Tuning — Cohere: Learn how to fine-tune a Command model for chat with the Cohere Web UI or Python SDK, including data requirements, pricing, and calling your model.


Cohere ▷ #api-discussions (19 messages🔥):

  • Multilingual Model Latency
  • API Token Usage
  • Read Timeout Issues

Link mentioned: incident.io - Status pages: no description found


Cohere ▷ #projects (1 messages):

  • Agentic Builder Day
  • OpenSesame collaboration
  • Cohere Models competition

Link mentioned: OpenSesame | Build Better AI Agents: OpenSesame simplifies the entire AI agent lifecycle, from building to evaluating. Our platform empowers businesses to easily create, share, and implement AI agents and detect hallucinations, making AI...


Modular (Mojo 🔥) ▷ #general (5 messages):

  • Mojo Language Structure of Arrays
  • Mojo Language Slices
  • Community Reflections on Collections

Modular (Mojo 🔥) ▷ #mojo (40 messages🔥):

  • Performance of Binary Stripping
  • Comptime Variables in Mojo
  • Using Tuple Arguments
  • BigInt Operations Comparison
  • Arbitrary Width Integer Libraries

tinygrad (George Hotz) ▷ #general (15 messages🔥):

  • LLVM Renderer Refactor
  • Tinygrad Performance Improvements
  • Gradient Clipping Integration
  • ACT Training Progress
  • TinyJit Decorator Queries

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (15 messages🔥):

  • Tensor Indexing Techniques
  • Python Compatibility with MuJoCo
  • Intermediate Representation Inspection
  • Custom Compiler Development

OpenInterpreter ▷ #general (24 messages🔥):

  • Hume AI Voice Assistant
  • Claude 3.5 Sonnet Release
  • Open Interpreter and Claude Integration
  • Screenpipe Tool
  • Open Source Monetization Models

Links mentioned:


OpenInterpreter ▷ #ai-content (1 messages):

facelessman: https://youtu.be/VgJ0Cge99I0 -- Love this episode -- love these folks!!!


DSPy ▷ #show-and-tell (2 messages):

  • New version creation
  • Upgrade process
  • Current system functionality

DSPy ▷ #general (22 messages🔥):

  • AI Helper Documentation
  • Broken Links
  • Docs Bot Return
  • General Vibe of 3.0

LlamaIndex ▷ #blog (4 messages):

  • VividNode desktop app
  • Serverless RAG app
  • Knowledge Management for RFPs
  • Llama Impact Hackathon
  • Document indexing in vector databases

LlamaIndex ▷ #general (14 messages🔥):

  • CondensePlusContextChatEngine Memory Initialization
  • Limiting TPM and RPM in LlamaIndex
  • Using GraphRag with Dynamic Data
  • Parsing .docx Files with LlamaIndex API
  • Persisting Context in Workflows

LlamaIndex ▷ #ai-discussion (1 messages):

  • LaBSE performance
  • sentence-transformers/multilingual models

LLM Agents (Berkeley MOOC) ▷ #hackathon-announcements (2 messages):

  • LLM Agents MOOC Hackathon
  • Hackathon FAQ
  • Sponsorship
  • Tracks of Hackathon
  • Sign-Up Details

Links mentioned:


LLM Agents (Berkeley MOOC) ▷ #mooc-announcements (2 messages):

  • Lecture 7 Announcement
  • LLM Agents MOOC Hackathon
  • TapeAgents Framework
  • WorkArena++ Benchmark
  • Hackathon Tracks and Sponsors

Links mentioned:


LLM Agents (Berkeley MOOC) ▷ #mooc-questions (7 messages):

  • Workflow understanding for AI agents
  • Assignment deadlines
  • Learning resources for auto gen

Link mentioned: AI Agentic Design Patterns with AutoGen - DeepLearning.AI: Use the AutoGen framework to build multi-agent systems with diverse roles and capabilities for implementing complex AI applications.


LLM Agents (Berkeley MOOC) ▷ #mooc-lecture-discussion (4 messages):

  • Lecture Start Time
  • YouTube Stream Issues

LLM Agents (Berkeley MOOC) ▷ #mooc-readings-discussion (4 messages):

  • Function Calling in LLMs
  • TapeAgents Framework
  • Agent Development
  • Model Distillation Techniques

Links mentioned:


Torchtune ▷ #general (16 messages🔥):

  • PyTorch Core Issue
  • Error in Distributed Training
  • Config File Format in Torchtune
  • Flex Performance on GPUs
  • Hardware Setup for Training

Links mentioned:


Torchtune ▷ #dev (1 messages):

  • Hermes 2.5.0 Release
  • Recommendations for nightlies

LangChain AI ▷ #general (6 messages):

  • Langchain Open Canvas compatibility
  • Agent orchestration with Langchain
  • Final output chain refactoring
  • Langchain 0.3.4 malicious warning
  • Local hosting for Enterprise applications

LangChain AI ▷ #share-your-work (2 messages):

  • NumPy Documentation Improvement
  • Transition to Consulting

Links mentioned:


LangChain AI ▷ #tutorials (1 messages):

  • Self-Attention
  • Dynamic Contextual Embeddings

OpenAccess AI Collective (axolotl) ▷ #general (2 messages):

  • Experimental Triton FA support
  • User Warning on Flash Attention

Link mentioned: [Feature]: Memory Efficient Flash Attention for gfx1100 (7900xtx) · Issue #16 · ROCm/aotriton: Suggestion Description Started using torchlearn to train models in pytorch using my gfx1100 card but get a warning that 1toch was not compiled with memory efficient flash attention. I see there is ...


OpenAccess AI Collective (axolotl) ▷ #general-help (6 messages):

  • Instruction-Tuned Models
  • Domain-Specific Instruction Data
  • Catastrophic Forgetting
  • Raw Domain Data
  • GPT-4 Generated Instruction Data

Gorilla LLM (Berkeley Function Calling) ▷ #discussion (2 messages):

  • Function Calling Model Fine-tuning
  • Benchmarking Custom Endpoints
  • Gorilla Project Documentation

Link mentioned: gorilla/berkeley-function-call-leaderboard at main · ShishirPatil/gorilla: Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls) - ShishirPatil/gorilla


LAION ▷ #resources (1 messages):

  • Webinar on LLM Best Practices
  • Prompt Engineering Techniques
  • Performance Optimization
  • Retrieval-Augmented Generation
  • Analytics Vidhya Blog Articles

Link mentioned: Explore the Future of AI with Expert-led Events: Analytics Vidhya is the leading community of Analytics, Data Science and AI professionals. We are building the next generation of AI professionals. Get the latest data science, machine learning, and A...


Mozilla AI ▷ #announcements (1 messages):

  • AI access challenges
  • Competition in AI
  • External researcher access
  • Big Tech and AI control





{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}