Frozen AI News archive

Llama 3.1 Leaks: big bumps to 8B, minor bumps to 70b, and SOTA OSS 405b model

**Llama 3.1** leaks reveal a **405B dense model** with **128k context length**, trained on **39.3M GPU hours** using H100-80GB GPUs, and fine-tuned with **over 25M synthetic examples**. The model shows significant benchmark improvements, especially for the 8B and 70B variants, with some evals suggesting the 70B outperforms **GPT-4o**. **GPT-4o Mini** launched as a cost-efficient variant with strong performance but some reasoning weaknesses. Synthetic datasets like **NuminaMath** enable models such as **Alibaba Qwen 2** to surpass GPT-4o and Claude 3.5 in math competitions. Discussions include reasoning task benchmarks and dataset building for improved reasoning.

Canonical issue URL

AI News for 7/19/2024-7/22/2024. We checked 7 subreddits, 384 Twitters and 30 Discords (474 channels, and 7039 messages) for you. Estimated reading time saved (at 200wpm): 765 minutes. You can now tag @smol_ai for AINews discussions!

We know it's coming tomorrow (with Soumith's ICML keynote), so we really tried to avoid discussing the leaks since we're going to be covering it tomorrow, but Llama 3.1 is leaking like a sieve (weights, evals, model card) that is leaky, so unfortunately it is all the community is talking about today, despite a lot of it being repeats of the first Llama 3 release in April.

Apart from the well telegraphed 405B dense model release, here are the diffs to Llama 3.1 as far as we can tell, mostly from the model card spells out the various priorities they had:

We made a diff spreadsheet to visualize - TLDR, HUGE bump for the 8B, across the board and instruct 70b is mildly better. 405B is still behind flagship models.:

image.png

However some independently run evals have Llama 3.1 70b doing better than GPT 4o - jury still out.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

GPT-4o Mini Release and Performance

Synthetic Data and Model Performance

Reasoning and Robustness Benchmarks

Memes and Humor in the AI Community


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. AI-Powered Mathematics Training

Theme 2. Local LLM Resource Optimization

All AI Reddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

Theme 1. LLaMA 3 405B Model Release and Implications

Theme 2. AI in Healthcare: Improving Cancer Detection

Theme 3. LocalLLaMA Advancements and Applications


AI Discord Recap

A summary of Summaries of Summaries

1. LLM Model Releases and Benchmarks

2. AI Infrastructure and Optimization

3. AI Model Performance and Efficiency

4. Knowledge Graphs and Retrieval-Augmented Generation (RAG)

5. Community Contributions and Open-Source Projects


PART 1: High level Discord summaries

Nous Research AI Discord


HuggingFace Discord


Modular (Mojo 🔥) Discord


Stability.ai (Stable Diffusion) Discord


CUDA MODE Discord


LM Studio Discord


Perplexity AI Discord


OpenAI Discord


OpenRouter (Alex Atallah) Discord


Cohere Discord


Eleuther Discord


Interconnects (Nathan Lambert) Discord


Latent Space Discord


OpenAccess AI Collective (axolotl) Discord


LangChain AI Discord


LAION Discord


LlamaIndex Discord


DSPy Discord


tinygrad (George Hotz) Discord


OpenInterpreter Discord


LLM Finetuning (Hamel + Dan) Discord


LLM Perf Enthusiasts AI Discord


Alignment Lab AI Discord


AI Stack Devs (Yoko Li) Discord


MLOps @Chipro Discord


DiscoResearch Discord


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Nous Research AI ▷ #research-papers (3 messages):

  • Implicit Chain-of-Thought (CoT)
  • UltraChat and Multi-turn Interactions
  • Multiagent Debate Models

Links mentioned:


Nous Research AI ▷ #datasets (1 messages):

fedorovist: https://huggingface.co/datasets/jdpressman/retroinstruct-mix-v0.2


Nous Research AI ▷ #off-topic (5 messages):

  • PPLX Pro Search AI
  • ReFT paper discussion
  • Greg Schoeninger's Reddit post on ReFT
  • YouTube video on ReFT
  • Oxen.ai community and Paper Club

Links mentioned:


Nous Research AI ▷ #interesting-links (1 messages):

alexanderlong_84476: https://pluralisresearch.substack.com/p/decentralized-ai-looms


Nous Research AI ▷ #general (434 messages🔥🔥🔥):

  • 1 bit quant results for DeepSeek-V2-Chat-0628
  • New AI projects and model updates
  • Llama 3.1 benchmarking
  • AI model legality concerns
  • Tech tools and deployment experiences

Links mentioned:


Nous Research AI ▷ #ask-about-llms (20 messages🔥):

  • Captcha Solving with LLMs
  • VRAM Estimates for LLama-3 405B
  • Consumer Level Hardware for AI Models

Nous Research AI ▷ #rag-dataset (40 messages🔥):

  • Triplex for knowledge graph construction
  • R2R platform for local LLM experimentation
  • Application of knowledge graphs in RAG/LLM
  • Microsoft's Graph RAG
  • Deeper adjacency matrix and symbolic reasoning

Links mentioned:


Nous Research AI ▷ #world-sim (3 messages):

  • World Sim
  • World Client
  • Nous Research

Link mentioned: worldsim: no description found


Nous Research AI ▷ #reasoning-tasks-master-list (691 messages🔥🔥🔥):

  • QuietStar
  • Auto-generate prompts
  • Type systems for LLMs
  • Intermediate representations
  • Task structure in Open-Reasoning-Tasks

Links mentioned:


HuggingFace ▷ #general (959 messages🔥🔥🔥):

  • LLM fine-tuning
  • GPU and hardware capabilities
  • Model deployment issues
  • Whisper Model for transcriptions
  • LLM code and architecture troubleshooting

Links mentioned:


HuggingFace ▷ #today-im-learning (4 messages):

  • Knowledge Graphs
  • News Reading Experience
  • Hugging Face Model Kwargs
  • Speaker Diarization & Whisper Transcription

HuggingFace ▷ #cool-finds (8 messages🔥):

  • Apple Intelligence
  • LoRA fine-tuning
  • AI paper on arXiv
  • AI's world transformation
  • Free online courses

Links mentioned:


HuggingFace ▷ #i-made-this (49 messages🔥):

  • Hermes 2.5
  • Model Merging
  • Open Empathic
  • Gary4live
  • SmolLM

Links mentioned:


HuggingFace ▷ #reading-group (2 messages):

  • Event Creation
  • Diagram Feedback

HuggingFace ▷ #core-announcements (1 messages):

  • SD3 training bugs
  • Diffusers repository

Link mentioned: [Training] SD3 training fixes by sayakpaul · Pull Request #8917 · huggingface/diffusers: What does this PR do? Fixes #8887 and #8708. Additionally, it adds an option to control the pre-conditioning behavior on the model outputs. Multiple folks have reported that for rectified-flows we ...


HuggingFace ▷ #computer-vision (19 messages🔥):

  • Hybrid Model with Inception and ViT
  • Scrabble Board Tile Detection
  • Binary Segmentation Projects

Links mentioned:


HuggingFace ▷ #NLP (10 messages🔥):

  • SQL RAG
  • Date Extraction
  • Open-source Text-to-HTML/CSS Model
  • Fine-tuning Language Models
  • Metrics in Transformer Fine-tuning

HuggingFace ▷ #diffusion-discussions (4 messages):

  • Open-source text-to-html/css generation
  • Artistic styles in SDv1.4
  • Evaluating diffusion models
  • Stable diffusion paper encoder methods

Modular (Mojo 🔥) ▷ #general (213 messages🔥🔥):

  • Mojo Socket Implementation
  • Dual Stack Sockets Issue
  • Mojo Integration with Existing Libraries
  • Interoperability between Mojo and Python
  • Production Readiness of Mojo

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (1 messages):

ModularBot: From Modular: https://twitter.com/Modular/status/1815463417391837596


Modular (Mojo 🔥) ▷ #📺︱youtube (1 messages):

  • Mojo 🔥 Community Meeting #4
  • Flat Buffers
  • Forge Tools
  • Mojo 🔥 Standard Library
  • Mojo 🔥 Gen

Link mentioned: Mojo 🔥 Community Meeting #4: Recording of the Mojo Community Meeting #4🫓 Flat Buffers: memory efficient serialization⚒️ Forge Tools: extending the Mojo 🔥 standard library🔄 Mojo 🔥 Gen...


Modular (Mojo 🔥) ▷ #mojo (185 messages🔥🔥):

  • Anti-Pattern Discussions
  • OpenSSL and Mojo Projects
  • Newton's Method for Float Literals
  • Future of Mojo's Async Scheduler
  • CPU Performance Comparisons

Links mentioned:


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (1 messages):

  • Matrix Multiplication in Mojo
  • Comparing Mojo to Numpy Performance

Link mentioned: Matrix multiplication in Mojo | Modular Docs: Learn how to leverage Mojo's various functions to write a high-performance matmul.


Modular (Mojo 🔥) ▷ #max (5 messages):

  • nightly/max feed reliability
  • getting started with MAX
  • open source contributions to MAX
  • guidance for new contributors

Links mentioned:


Modular (Mojo 🔥) ▷ #max-gpu (3 messages):

  • XLA Involvement
  • Mojo on GPU Availability

Modular (Mojo 🔥) ▷ #nightly (83 messages🔥🔥):

  • Nightly Mojo Compiler Updates
  • LegacyPointer and DTypePointer Discussion
  • Issues with memcpy Function
  • Changes in Mojo API
  • Community Interaction and Documentation

Links mentioned:


Modular (Mojo 🔥) ▷ #mojo-marathons (43 messages🔥):

  • NumPy Performance Testing
  • Understanding CPU FLops
  • Matrix Multiplication Benchmarks
  • Architecture-Specific Optimizations
  • Mojo's Generics Limitations

Link mentioned: GitHub - m-j-w/CpuId.jl: Ask the CPU for cache sizes, SIMD feature support, a running hypervisor, and more.: Ask the CPU for cache sizes, SIMD feature support, a running hypervisor, and more. - m-j-w/CpuId.jl


Stability.ai (Stable Diffusion) ▷ #general-chat (424 messages🔥🔥🔥):

  • ComfyUI recommendations
  • Forge vs. Easy Diffusion experience
  • Using Latent mode for Regional Prompter
  • Issues with VRAM and GPU compatibility
  • Upscaling errors in Forge

Links mentioned:


CUDA MODE ▷ #general (28 messages🔥):

  • Google Meet Troubleshooting
  • CUDA and SVD
  • Building with LLMs
  • ECC Memory in Workstations
  • Register Allocation in Flash Attention

Links mentioned:


CUDA MODE ▷ #triton (4 messages):

  • Profiling Triton Kernels
  • Memory Usage
  • CUDA tools for Profiling
  • Nsight Compute
  • Nsight Systems

CUDA MODE ▷ #torch (7 messages):

  • at::Tensor.mutable_data_ptr
  • torch.cond
  • torch control flow operators

Link mentioned: torch.cond — PyTorch 2.3 documentation: no description found


CUDA MODE ▷ #announcements (1 messages):

  • AMD ROCm
  • Composable Kernel library
  • AMD tech stack

CUDA MODE ▷ #algorithms (3 messages):

  • Similarity Search Algorithm
  • Schotastic Rounding in Quantization

CUDA MODE ▷ #cool-links (27 messages🔥):

  • CubeCL
  • FlashAttention2 Custom Mask
  • FLUTE Kernel

Links mentioned:


CUDA MODE ▷ #beginner (17 messages🔥):

  • Triton Block Size
  • Non-GPU part of NVCC Compiler and VLAs
  • FP16 vs FP32 Performance
  • Triton Multi-Stage Pipelining

Links mentioned:


CUDA MODE ▷ #pmpp-book (3 messages):

  • Model Weights in Shared Memory
  • CUDA Register Capacity

CUDA MODE ▷ #youtube-recordings (1 messages):

andreaskoepf: https://youtu.be/-732zELVbpU?si=HBXEE8t2fxCKhC5v


CUDA MODE ▷ #off-topic (8 messages🔥):

  • Request for Cuda/C++ Channel
  • Discussion on existing channel usage
  • LLM Deployment Cost Inquiry

CUDA MODE ▷ #irl-meetup (6 messages):

  • ICML Meetup at Conference
  • CUDA Learners in Berkeley

CUDA MODE ▷ #llmdotc (297 messages🔥🔥):

  • CUDA MODE IRL Event
  • Train GPT-2 to GPT-3 Progress
  • Multi-GPU and ZeRO Implementation
  • MuP Branch Progress
  • FP8 Training Optimizations

Links mentioned:


CUDA MODE ▷ #rocm (10 messages🔥):

  • ROCm hardware entry options
  • Differences between RDNA and CDNA
  • MI300 capabilities
  • Challenges with AMD GPUs in ROCm
  • FP8 MMA acceleration on MI300

CUDA MODE ▷ #lecture-qa (1 messages):

  • Arithmetic Intensity
  • GPU Performance Metrics

LM Studio ▷ #💬-general (168 messages🔥🔥):

  • GGUF file metadata in C#
  • Quantization process (q5, q6, f16)
  • Issues with LM Studio & Hugging Face
  • Running large models locally
  • Using local LLMs for profit

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (126 messages🔥🔥):

  • Open Model Test Results
  • Converting Models to GGUF
  • New Jail-Breaking Technique
  • DeepSeek-Coder Issues
  • Memory Usage with Qwen2 72B

Links mentioned:


LM Studio ▷ #announcements (2 messages):

  • Hugging Face API Networking Errors
  • Issue Resolution

LM Studio ▷ #🧠-feedback (11 messages🔥):

  • Failed model load
  • Flash Attention troubleshooting
  • HuggingFace API issues
  • Alternative model repositories

Link mentioned: Hugging Face – The AI community building the future.: no description found


LM Studio ▷ #📝-prompts-discussion-chat (2 messages):

  • Renaming Presets
  • Finding Models

LM Studio ▷ #🎛-hardware-discussion (40 messages🔥):

  • NVidia Tesla P40 usage
  • AMD GPU compatibility
  • Home NAS recommendations
  • Choosing GPUs for AI/ML workloads
  • Finetuning Large Language Models

Link mentioned: Getting started with LLM fine-tuning: Large Language Model (LLM) Fine-tuning is the process of adapting the pre-trained model to specific tasks. This process is done by updating its parameters on a new dataset. Specifically, the LLM is pa...


LM Studio ▷ #🧪-beta-releases-chat (18 messages🔥):

  • Nemo issues
  • Search functionality problems
  • Huggingface API

LM Studio ▷ #amd-rocm-tech-preview (1 messages):

captainpumpkinhead: "natively"


LM Studio ▷ #model-announcements (1 messages):

  • Athese by Nexusflow
  • Model benchmarks
  • Multilingual performance

LM Studio ▷ #🛠-dev-chat (2 messages):

  • LM Studio Discord bot
  • Private responses with LM Studio
  • GitHub tutorial link

Link mentioned: GitHub - mrdjohnson/lmstudio-discord-bot: A tutorial for creating a Discord bot that responds using LM Studio! This code is based on a blogpost found here: https://dev.to/mrdjohnson/i-made-a-discord-bot-with-lmstudiojs-4fd6: A tutorial for creating a Discord bot that responds using LM Studio! This code is based on a blogpost found here: https://dev.to/mrdjohnson/i-made-a-discord-bot-with-lmstudiojs-4fd6 - mrdjohnson/lm...


Perplexity AI ▷ #general (338 messages🔥🔥):

  • Image generation with Pro subscription
  • GPTs Agents
  • Issues with Pro search
  • Profile and collection prompts
  • Perplexity's context window

Links mentioned:


Perplexity AI ▷ #sharing (23 messages🔥):

  • YouTube tests AI conversational capabilities
  • OpenAI drops GPT-4.0 mini
  • Unraveling Chaos project
  • CrowdStrike global IT outage
  • Possibilities of developing software

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (9 messages🔥):

  • Feature Roadmap
  • Perplexity API Token Charges
  • Online Models Usage

OpenAI ▷ #ai-discussions (170 messages🔥🔥):

  • Sonnet 3.5 vs 4O Mini
  • Finetuning multimodal models
  • Cost-effective TTS solutions
  • Voice assistant apps
  • GPT-4o mini vs GPT-5

Link mentioned: System Prompt in Markdown Code Block | Arc Search: Arc Search read websites across the internet to make you this perfect tab.


OpenAI ▷ #gpt-4-discussions (32 messages🔥):

  • GPT-4o mini replacing GPT-3.5
  • Differences between GPT-4o and GPT-4o mini
  • New features for GPT-4o
  • API vs. ChatGPT features
  • GPT-4o mini for longform content

OpenAI ▷ #prompt-engineering (9 messages🔥):

  • Solving Mathematical or Logical Problems Accurately
  • Custom Instructions in ChatGPT
  • Using Custom Instructions Effectively
  • Guidance AI for Prompt Engineering

OpenAI ▷ #api-discussions (9 messages🔥):

  • Improving ChatGPT Accuracy
  • Modification of ChatGPT Responses
  • Custom Instructions for ChatGPT
  • Prompt Engineering Tools

OpenRouter (Alex Atallah) ▷ #announcements (1 messages):

  • Rankings page update
  • Infrastructure migration

OpenRouter (Alex Atallah) ▷ #app-showcase (2 messages):

  • OpenRouter provider for GPTScript
  • gptscript on command line
  • gptscript demo video

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (202 messages🔥🔥):

  • Hermes 2.5
  • GPTs Agents
  • OpenRouter Feature Requests
  • Model Merging
  • Dolphin Llama 70B

Links mentioned:


Cohere ▷ #general (156 messages🔥🔥):

  • Open source contributions
  • LLM use in RPG games
  • Nextjs app router update
  • Developer office hours format
  • Integrating Cohere models

Links mentioned:


Cohere ▷ #project-sharing (4 messages):

  • chat GUI with local LLMs
  • multiplayer text games Discord app

Link mentioned: Create 'n' Play - Create AI multiplayer text games for your Discord server! | Product Hunt: AI-powered Discord bot turns your server into a text game paradise! Craft ANY game with our /search-games command. Endless possibilities—it's the ultimate community engagement tool. Let AI fuel y...


Cohere ▷ #announcements (1 messages):

  • Developer Office Hours
  • Structured Generations in the API
  • Cohere Toolkit features
  • Cohere For AI research papers
  • Community events

Links mentioned:


Eleuther ▷ #general (13 messages🔥):

  • Softmax invariance and z loss
  • Understanding large language models (LLM)
  • GPT model example code
  • Scaling diarization pipelines
  • Vector quantization models

Link mentioned: easydist/benchmark/torch/model/gpt.py at 3dbb146812fddf6259590a0b4611a251f3e7cbe5 · alibaba/easydist: Automated Parallelization System and Infrastructure for Multiple Ecosystems - alibaba/easydist


Eleuther ▷ #research (39 messages🔥):

  • Self-Modeling in AI
  • Hybrid Post-Quantum Encryption
  • Feature Contamination in Neural Networks
  • Human-like Response Time in CNNs
  • Efficient Dictionary Learning with Switch SAE

Links mentioned:


Eleuther ▷ #scaling-laws (3 messages):

  • Scaling laws and hypernetworks
  • Brains and backprop
  • Scaling Exponents Across Parameterizations and Optimizers

Link mentioned: Tweet from main (@main_horse): Scaling Exponents Across Parameterizations and Optimizers [GDM] [nocode/weights] https://arxiv.org/abs/2407.05872 trains 10,000+ (!) models, varying * optim (SGD/Adam/Adafactor) * model size (1.1B ~...


Eleuther ▷ #interpretability-general (9 messages🔥):

  • MATS 7.0 Applications
  • **nnsight** Paper Release
  • Apollo's Mech Interp Projects List
  • Tokengrams Project
  • Suffix Arrays for Pile

Links mentioned:


Eleuther ▷ #lm-thunderdome (56 messages🔥🔥):

  • Zeno Upload Feature
  • Commit Branch Queries
  • Logging Issues
  • Multinode Inference Support
  • PSA: ICML Conference

Link mentioned: Refactor API models by baberabb · Pull Request #2008 · EleutherAI/lm-evaluation-harness: This PR introduces a new superclass for API request models, providing: Modularity for downstream classes Overloadable methods for request transformation, API requests and response parsing Tokeniza...


Eleuther ▷ #multimodal-general (4 messages):

  • Latent downscaling
  • Image classification performance
  • Generated latents

Eleuther ▷ #gpt-neox-dev (19 messages🔥):

  • Nemotron-340B specifics
  • Nathan's bounty for Nemotron-340B conversion
  • vLLM multinode inference
  • Multi-node performance for Nemotron-340B
  • Evaluation harness discussion

Link mentioned: Tweet from Nathan Lambert (@natolambert): I'm offering a paid bounty to successfully convert nvidia/Nemotron-4-340B-Instruct to HuggingFace / related libraries. Starting reward $75 We really need this to unlock synthetic permissive data...


Interconnects (Nathan Lambert) ▷ #news (61 messages🔥🔥):

  • Nemotron-4-340B conversion to HuggingFace
  • Llama-3 and 3.1 leaks
  • Meta AI's potential premium offerings
  • Distillation techniques for large models
  • SOC2 compliance for HuggingFace

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (4 messages):

  • WizardMath paper
  • Instruction Reward Model (IRM)
  • PRM by Uesato et al 2022 paper
  • Step-by-step reward labeling
  • UltraChat vs Zephyr paper

Interconnects (Nathan Lambert) ▷ #ml-drama (3 messages):

  • ICML 2024 Spotlight
  • Lit Review Concerns
  • Harvey Legal AI Criticism

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (52 messages🔥):

  • Interview Invitations
  • MosaicML Sword Tradition
  • Claude AI Text Restrictions
  • Reward Model Innovations
  • Stripe Settings Issue

Links mentioned:


Interconnects (Nathan Lambert) ▷ #nlp (2 messages):

  • Blog post on distillation
  • Lilian Wang
  • Surprise at lack of resources

Interconnects (Nathan Lambert) ▷ #reads (2 messages):

  • Yitay blog post on model architectures
  • Encoder vs. Encoder-Decoder models
  • Evolution of LLMs
  • @srush_nlp tweet

Link mentioned: What happened to BERT & T5? On Transformer Encoders, PrefixLM and Denoising Objectives — Yi Tay: A Blogpost series about Model Architectures Part 1: What happened to BERT and T5? Thoughts on Transformer Encoders, PrefixLM and Denoising objectives


Latent Space ▷ #ai-general-chat (74 messages🔥🔥):

  • Langfuse vs Langsmith
  • GPT-4o mini and AI-generated content
  • Rumors about Harvey AI
  • Elon Musk's Memphis Supercluster
  • LLaMA 3.1 leaks and evaluations

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

swyxio: big monthly recap is up: https://x.com/latentspacepod/status/1815411709085143197


Latent Space ▷ #ai-in-action-club (29 messages🔥):

  • Audio Issues
  • Layout Detection
  • Texify vs Mathpix
  • Presentation Feedback
  • Model Training Dataset

Link mentioned: VikParuchuri - Overview: VikParuchuri has 90 repositories available. Follow their code on GitHub.


OpenAccess AI Collective (axolotl) ▷ #general (64 messages🔥🔥):

  • Training Qwen2-7b
  • Triplex for Knowledge Graphs
  • Mistral 12b Issues
  • LLaMA 3 Inconsistencies
  • LLaMA 3.1 Benchmarks

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (22 messages🔥):

  • Knowledge Distillation Support
  • DPO Improvements
  • DeepSpeed Zero-3 Compatibility

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (5 messages):

  • Axolotl training error
  • GPU memory issues
  • Batch size adjustment
  • Mixed precision training
  • NCCL timeouts

Links mentioned:


LangChain AI ▷ #general (57 messages🔥🔥):

  • LangChain.js Token issues
  • Video and blog post on LLM lessons
  • Beginners guide on LangChain
  • Vector store filtering
  • Discussion on deploying RAG app to production

Links mentioned:


LangChain AI ▷ #share-your-work (6 messages):

  • Triplex model
  • Vector visualization for embeddings
  • Semantic search with LangChain
  • AI function builder for TypeScript
  • LangChain tutorial

Links mentioned:


LangChain AI ▷ #tutorials (3 messages):

  • LangChain article on Medium by Harshit Ambalia
  • Guide for deploying a RAG app
  • Scheduler Agent guide using Composio, LangChain, and ChatGPT

Links mentioned:


LAION ▷ #general (34 messages🔥):

  • sdxl vae latents
  • HF hosting advantages
  • new BUD-E demo
  • local LLM on Linux terminal
  • Kolors diffusion model

Links mentioned:


LAION ▷ #announcements (1 messages):

  • Bud-E voice assistant
  • Daily Online-Hackathons
  • BUD-E Discord Server

LAION ▷ #research (14 messages🔥):

  • Plotting Loss Curves
  • Mem0 AI Memory
  • Datadog Time Series Modeling
  • Research Recruitment

Links mentioned:


LlamaIndex ▷ #blog (8 messages🔥):

  • PostgresML for Reranking
  • LLMs as Judges
  • Merlinn: Open-source On-call Copilot
  • Multimodal RAG with Ollama and Qdrant
  • Deasie RAG Workshop

Links mentioned:


LlamaIndex ▷ #general (33 messages🔥):

  • llama-parse API issues
  • ReActAgent max iterations
  • VectorStoreIndex embedding model
  • LlamaIndex webinar
  • Extracting pictures from PDFs

Links mentioned:


LlamaIndex ▷ #ai-discussion (4 messages):

  • ETL for Video and Music Data
  • SycoPhancy in LLMs
  • Korvus for RAG Pipeline

Links mentioned:


DSPy ▷ #general (44 messages🔥):

  • Issues with GPT4o-mini Model
  • DSPy Tracing Release
  • TypedPredictors Compatibility
  • DSPy Paper Release
  • Optimizers Reliability in DSPy

Links mentioned:


tinygrad (George Hotz) ▷ #general (8 messages🔥):

  • OpenPilot Model Run Analysis
  • Bitcast Functionality in Tinygrad
  • Promising Pull Requests
  • Weekly Tinygrad Meeting

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (11 messages🔥):

  • Composing LazyBuffers
  • Shapetrackers Tutorial
  • Viability of Tinygrad versus PyTorch

Link mentioned: Tinygrad tutorial - shapetrackers and view merging - machine learning optimization: tinygrad fork with code tutorial: https://github.com/Zaffer/tinygrad/tree/tuturial-notebooktinygrad docs: https://docs.tinygrad.org/tinygrad notes: https://m...


OpenInterpreter ▷ #general (4 messages):

  • Crowdstrike update
  • Python subinterpreters
  • Meta Llama 3.1

Links mentioned:


OpenInterpreter ▷ #O1 (8 messages🔥):

  • Deepseek chat v2 6.28
  • 4o mini performance
  • Apple Watch support
  • Device shipping updates
  • Coqui model on MacOS

OpenInterpreter ▷ #ai-content (1 messages):

  • Augmentoolkit on GitHub
  • Pinokio Project Launch

Link mentioned: Pinokio: AI Browser


LLM Finetuning (Hamel + Dan) ▷ #general (4 messages):

  • Finetuning with GPT models
  • Issues with OpenAI credits

LLM Finetuning (Hamel + Dan) ▷ #jarvis-labs (1 messages):

vishnu9158: Nope


LLM Finetuning (Hamel + Dan) ▷ #east-coast-usa (1 messages):

karmapa: Yes how about late august meetup in NY?


LLM Finetuning (Hamel + Dan) ▷ #openpipe (1 messages):

  • Openpipe with other providers
  • Integration of Replicate models
  • Modal model compatibility

LLM Finetuning (Hamel + Dan) ▷ #openai (1 messages):

  • Issues with credit allocation
  • Course forms

LLM Perf Enthusiasts AI ▷ #general (3 messages):

  • OpenAI Scale Tier
  • TPS Calculation
  • GPT-4-o throughput

LLM Perf Enthusiasts AI ▷ #jobs (1 messages):

  • Websim platform
  • Founding AI Engineer role
  • AI-assisted software creation
  • Non-deterministic programs
  • Human-AI system

Link mentioned: websim.ai: no description found


Alignment Lab AI ▷ #general-chat (2 messages):

  • Building with LLMs
  • BUD-E Voice Assistant

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #team-up (1 messages):

ari991963: Hi all, I am Aria a 2D/3D artist, if you are interested to collaborate dm


MLOps @Chipro ▷ #general-ml (1 messages):

  • Target Audience Clarification
  • Communication Strategy

DiscoResearch ▷ #general (1 messages):

  • 1 year of building with LLMs
  • TLDR series on LLMs
  • Lessons from LLM practitioners

Link mentioned: TLDR: 1 year of building with LLMs – D-Squared: no description found



{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}