Frozen AI News archive

Qdrant''s BM42: "Please don''t trust us"

**Qdrant** attempted to replace BM25 and SPLADE with a new method called "BM42" combining transformer attention and collection-wide statistics for semantic and keyword search, but their evaluation using the Quora dataset was flawed. **Nils Reimers** from **Cohere** reran BM42 on better datasets and found it underperformed. Qdrant acknowledged the errors but still ran a suboptimal BM25 implementation. This highlights the importance of dataset choice and evaluation sanity checks in search model claims. Additionally, **Stripe** faced criticism for AI/ML model failures causing account and payment issues, prompting calls for alternatives. **Anthropic** revealed that **Claude 3.5 Sonnet** suppresses some answer parts with backend tags, sparking debate. **Gemma 2** model optimizations allow 2x faster fine-tuning with 63% less memory and longer context windows, running up to 34B parameters on consumer GPUs. **nanoLLaVA-1.5** was announced as a compact 1B parameter vision model with significant improvements.

Canonical issue URL

AI News for 7/4/2024-7/5/2024. We checked 7 subreddits, 384 Twitters and 30 Discords (418 channels, and 3772 messages) for you. Estimated reading time saved (at 200wpm): 429 minutes. You can now tag @smol_ai for AINews discussions + try Smol Talk!

Qdrant is widely known as OpenAI's vector database of choice, and over the July 4 holiday they kicked off some big claims to replace the venerable BM25 (and even the more modern SPLADE), attempting to coin "BM42":

image.png

to solve the problem of semantic + keyword search by combining transformer attention for word importance scoring with collection-wide statistics like IDF, claiming advantages over every use case:

image.png

Only one problem... the results. Jo Bergum from Vespa (a competitor), pointed out the odd choice of Quora (a "find similar duplicate" questions dataset, not a Q&A retrieval dataset) as dataset and obviously incorrect evals if you know that dataset:

image.png

Specifically, the Quora dataset only has ~1.6 datapoints per query so their precision@10 number was obviously wrong claiming to have >4 per 10.

Nils Reimers of Cohere took BM42 and reran on better datasets for finance, biomedical, and Wikipedia domains, and sadly BM42 came up short on all accounts:

image.png

For their part, Qdrant has responded to and acknowledged the corrections, and published corrections... except still oddly running a BM25 implementation that scores worse than everyone else expects and conveniently worse than BM42.

Unfortunate for Qdrant, but the rest of us just got a lightning lesson in knowing your data, and sanity checking evals. Lastly, as always in PR and especially in AI, Extraordinary claims require extraordinary evidence.

Meta note: If you have always wanted to customize your own version of AI News, we have now previewed a janky early version of Smol Talk, which you can access here: https://smol.fly.dev


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet.

Stripe Issues and Alternatives

AI and LLM Developments

AI Art and Perception

Memes and Humor


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Progress and Implications

AI Models and Capabilities

AI Safety and Security

AI Art and Media

Robotics and Embodied AI

Miscellaneous


AI Discord Recap

A summary of Summaries of Summaries

1. LLM Performance and Optimization

2. Open Source AI Ecosystem

3. Multimodal AI and Generative Models

4. Stability AI Licensing

5. Community Tools and Platforms

6. New LLM Releases and Benchmarking Discussions:

7. Optimizing LLM Training and Inference:

8. Advancements in Open-Source AI Frameworks and Datasets:

9. Multimodal AI and Generative Models:

10. New Model Releases and Training Tips in Unsloth AI Community:


PART 1: High level Discord summaries

HuggingFace Discord


Stability.ai (Stable Diffusion) Discord


Unsloth AI (Daniel Han) Discord


Latent Space Discord


LM Studio Discord


CUDA MODE Discord


Perplexity AI Discord


LAION Discord


OpenAI Discord


Nous Research AI Discord


OpenRouter (Alex Atallah) Discord


Eleuther Discord


LangChain AI Discord


LlamaIndex Discord


Cohere Discord


OpenInterpreter Discord


Modular (Mojo 🔥) Discord


LLM Finetuning (Hamel + Dan) Discord


Interconnects (Nathan Lambert) Discord


OpenAccess AI Collective (axolotl) Discord


tinygrad (George Hotz) Discord


Torchtune Discord


AI Stack Devs (Yoko Li) Discord


DiscoResearch Discord


MLOps @Chipro Discord


Datasette - LLM (@SimonW) Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ▷ #announcements (1 messages):

  • VLM training dataset in Vietnamese
  • Highlights parser
  • See 2 sound demo
  • text2cypher model
  • Guide to Designing New Functional Proteins

HuggingFace ▷ #general (495 messages🔥🔥🔥):

  • Use of Deepeval with HuggingFace Transformers
  • Proficiency certifications in ML
  • Uploading image on HuggingFace projects using Gradio API
  • GPU recommendations for ML beginners
  • Issues with renting A100 vs. 4090 GPUs for inference

HuggingFace ▷ #today-im-learning (2 messages):

  • Building a TikTok videos dataset for harmful content classification
  • Troubleshooting LDM implementation with RGB images

HuggingFace ▷ #cool-finds (6 messages):

  • Kyutai.org's digital pirate understands English with a French accent
  • Small demo of Moshi, an audio language model
  • Graph Structure Learning (GSL) with GraphEdit and large language models
  • Claude's ease in building Deep Learning Visualizer dashboards
  • nanoLLaVA - cool VLM under 1B

HuggingFace ▷ #i-made-this (32 messages🔥):

  • Introduction of Vision-Language model for Vietnamese by Vi-VLM team
  • Vi-VLM releasing a dataset for VLM training in Vietnamese
  • Simple translation tool for converting messages to pt-br
  • CyclicFormer architecture enhancement for transformers
  • UVR5's UI completion for audio separation

HuggingFace ▷ #reading-group (7 messages):

  • triton paper reading
  • upcoming paper reading schedule
  • interest in audio-language models
  • flora paper discussion

HuggingFace ▷ #computer-vision (4 messages):

  • WHAM alternatives for human pose estimation in monocular, in-the-wild videos
  • Learning ViT and U-Net implementations
  • Using visual-semantic information to boost fine-grained image classification performance
  • Discussing zero/few shot multi-modal models at CVPR

Link mentioned: 08. PyTorch Paper Replicating - Zero to Mastery Learn PyTorch for Deep Learning: Learn important machine learning concepts hands-on by writing PyTorch code.


HuggingFace ▷ #NLP (17 messages🔥):

  • Meta-LLaMA download issues
  • API calls to models without local download
  • Inference freeze in Mistral model
  • Static KV cache documentation
  • Troubleshooting errors related to memory

Link mentioned: Release v4.38: Gemma, Depth Anything, Stable LM; Static Cache, HF Quantizer, AQLM · huggingface/transformers: New model additions 💎 Gemma 💎 Gemma is a new opensource Language Model series from Google AI that comes with a 2B and 7B variant. The release comes with the pre-trained and instruction fine-tuned v....


HuggingFace ▷ #diffusion-discussions (3 messages):

  • Running RealVisXL_V4.0_Lightning using diffusers
  • Error with yisol/IDM-VTON in Google Colab
  • Improving resume analyzer to assess project intensity

Link mentioned: SG161222/RealVisXL_V4.0_Lightning · Hugging Face: no description found


Stability.ai (Stable Diffusion) ▷ #announcements (1 messages):

  • License concerns with SD3 Medium release
  • Stability AI Community License update
  • Issues with commercial licensing in previous release
  • Improvement and support for open source community

Link mentioned: Community License — Stability AI: Our new Community License is now free for research, non-commercial, and commercial use. You only need a paid Enterprise license if your yearly revenues exceed USD$1M and you use Stability AI models in...


Stability.ai (Stable Diffusion) ▷ #general-chat (528 messages🔥🔥🔥):

  • Hyper vs turbo
  • AAM Anime Mix XL
  • Animagine XL 3.1
  • Stable Diffusion GPU usage
  • CivitAI and SD3 discussions

Unsloth AI (Daniel Han) ▷ #general (267 messages🔥🔥):

  • Gemma 2 Release and its features
  • Issues with the Gemma 2 notebooks and user feedback
  • Methods for dataset preparation and handling long-context examples
  • Performance and optimization techniques for various LLMs
  • Recent advancements and announcements in AI models and tools

Unsloth AI (Daniel Han) ▷ #announcements (1 messages):

  • Gemma 2 Release
  • Training speed and VRAM reduction
  • Context length improvements
  • 4-bit model support updates
  • Experimentation with models

Unsloth AI (Daniel Han) ▷ #off-topic (7 messages):

  • Release of Replete-AI datasets
  • Discussion on Facebook multi-token prediction
  • Fireworks.ai yi-large issues

Unsloth AI (Daniel Han) ▷ #help (121 messages🔥🔥):

  • Issues with Unsloth patch 2024.7 and checkpoints
  • Gemma 2 support in Unsloth
  • Fine-tuning models using Unsloth
  • Errors during fine-tuning and evaluation processes
  • Updating Unsloth and GGUF issues

Unsloth AI (Daniel Han) ▷ #showcase (3 messages):

  • Release of two new massive datasets by Replete-AI
  • Details and translations of Everything_Instruct_Multilingual
  • Questions about dataset deduplication and content balance

Unsloth AI (Daniel Han) ▷ #community-collaboration (10 messages🔥):

  • Pinning notebooks
  • Adding notebooks to the GitHub page
  • Correcting notebook links in the channels

Latent Space ▷ #ai-general-chat (94 messages🔥🔥):

  • AI + Blockchain funding discussions
  • Git merge tool alternatives and conflict resolutions
  • Learning AI curriculum and recommendations
  • Claude and other AI tools for coding assistance
  • Evaluations and criticisms of new search algorithms like BM42

Latent Space ▷ #ai-announcements (5 messages):

  • New podcast episode with Yi Tay of Reka
  • Discussion on the qualities of successful AI researchers
  • Comparisons of OpenAI, Google Gemini, and Reka teams
  • Technical topics covered in the podcast

Latent Space ▷ #llm-paper-club-west (34 messages🔥):

  • Issues with Discord AV
  • Migration to Zoom for better AV
  • Known compatibility issues between Discord and Linux

Link mentioned: Join our Cloud HD Video Meeting: Zoom is the leader in modern enterprise video communications, with an easy, reliable cloud platform for video and audio conferencing, chat, and webinars across mobile, desktop, and room systems. Zoom ...


Latent Space ▷ #ai-in-action-club (243 messages🔥🔥):

  • User technical difficulties and skill humor
  • Personal compliments to workshop hosts
  • Discussion on model merging tactics
  • LlamaFile vs Ollama comparison
  • Event planning and feedback

LM Studio ▷ #💬-general (157 messages🔥🔥):

  • Waiting to upgrade hardware for LM Studio
  • Comparison of Llama3 and Mistral models
  • Usage of API keys from OpenAI or Anthropic in LM Studio
  • Text embeddings and local server setup in LM Studio
  • Challenges in running large models like Llama3 70b on limited hardware

LM Studio ▷ #🤖-models-discussion-chat (130 messages🔥🔥):

  • Discussion on model behavior mismatch between different quantized versions of Gemma-2-27b
  • Using system prompts to improve coding model behaviors
  • Comparing different quantization techniques and their performance
  • Qwen2 model preset and ChatML format discussion
  • Issues and experiences with different large language models like Gemma, InternLM, and Dolphin

LM Studio ▷ #🧠-feedback (3 messages):

  • Issue with model downloads in LM on MacBook Pro M2
  • Solution for pausing/stopping downloads in LM

LM Studio ▷ #⚙-configs-discussion (5 messages):

  • Nxcode 7B JSON request
  • CodeQwen 1.5 7B ChatML compatibility
  • RTX 4060 8GB VRAM and 16 GB DDR5 RAM performance issues
  • Suggested models for mid-range GPU setups

LM Studio ▷ #🎛-hardware-discussion (61 messages🔥🔥):

  • Surface Laptop with Snapdragon Elite performance details
  • NPU and GPU utilization in Snapdragon devices
  • Comparison of CPU performance on Snapdragon and Intel devices
  • Future support for NPU in Llama.cpp
  • General discussion on hardware used with LM Studio

LM Studio ▷ #🧪-beta-releases-chat (2 messages):

  • AppImage not compatible with aarch64 CPUs
  • No ARM CPU support on Linux for LM Studio

LM Studio ▷ #amd-rocm-tech-preview (2 messages):

  • 7800XT user confirms GPU works
  • Problems loading models with GPU offload
  • Successful ROCm installation script

CUDA MODE ▷ #general (10 messages🔥):

  • Matrix multiplication in CUDA
  • Efficient remote development with pay-for-use compute
  • New blog post on executive summary process using GPTs
  • Paid CUDA/ML system certifications
  • Upcoming in-person CUDA mode event in October

CUDA MODE ▷ #triton (2 messages):

  • Triton kernels with multiple CUDA graphs create latency issues
  • SRAM contention affecting performance

CUDA MODE ▷ #torch (7 messages):

  • torch.compile not supported on Python 3.12
  • Python bytecode compatibility issues
  • TorchDynamo and Python frame evaluation API
  • TorchDynamo's role in PyTorch performance

CUDA MODE ▷ #algorithms (5 messages):

  • New method for training language models to predict multiple future tokens
  • Self speculative decoding in language models
  • Comparison between multi-token prediction and lookahead decoding baselines
  • Effectiveness of n-gram generation in multi-token prediction models

Link mentioned: Better & Faster Large Language Models via Multi-token Prediction: Large language models such as GPT and Llama are trained with a next-token prediction loss. In this work, we suggest that training language models to predict multiple future tokens at once results in h...


CUDA MODE ▷ #cool-links (1 messages):

iron_bound: https://oimo.io/works/life/


CUDA MODE ▷ #beginner (17 messages🔥):

  • Learning path for backend SDEs interested in CUDA and inference optimization
  • Challenges of finding jobs with open source contributions
  • Recommendation of CUDA Mode GitHub for beginners
  • Building a deep learning framework from scratch in C++ using CUDA
  • Using Python for CUDA kernel development vs C++

CUDA MODE ▷ #pmpp-book (4 messages):

  • Fourth edition released in 2022
  • Differences between third and fourth editions

CUDA MODE ▷ #jax (4 messages):

  • casual conversation
  • channel engagement

CUDA MODE ▷ #torchao (11 messages🔥):

  • Handling a.to method recognition and functionality
  • Removing unnecessary args in PyTorch/ao
  • Current limitations and workarounds for a.to method
  • Adding support for device and dtype handling in subclasses
  • Future functionality and testing in Torchbench models

CUDA MODE ▷ #off-topic (3 messages):

  • Thunder Sessions podcast by Lightning AI
  • Andrej Karpathy's keynote at UC Berkeley AI Hackathon 2024

CUDA MODE ▷ #llmdotc (134 messages🔥🔥):

  • CUDA MODE Discord chatbot messages
  • FP8 Gradient Issues in GPT-2 Training
  • Schedule-Free Optimizer Paper
  • GPT-2 Training Performance
  • Training Length Estimations for GPT-2

CUDA MODE ▷ #bitnet (3 messages):

  • Optimized kernels in CUDA for int2*int8 gemm
  • Release of a custom gemv for int2*int8
  • BitBLAS library for mixed-precision matrix multiplications

Link mentioned: GitHub - microsoft/BitBLAS: BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.: BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment. - microsoft/BitBLAS


Perplexity AI ▷ #general (165 messages🔥🔥):

  • Perplexity AI Repetition Issue
  • Live Internet Access Problems
  • Math Accuracy in Perplexity Pro
  • Experience with Perplexity in Stock Market
  • Subscription Plans and Model Usage

Perplexity AI ▷ #sharing (13 messages🔥):

  • Threads' Milestone
  • Ancient Aboriginal Rituals
  • Nuclear-Powered Data Centers
  • Mars Moss
  • Eating Contests

Perplexity AI ▷ #pplx-api (15 messages🔥):

  • Difference between pplx-70b-online and llama-3-sonar-large-32k-online
  • Google Dorks usage with the API
  • Temporal awareness in LLMs
  • Effectiveness of query commands in LLMs
  • Perplexity AI models and model cards

LAION ▷ #general (185 messages🔥🔥):

  • BUD-E update on new features
  • Issues with Clipdrop NSFW detection
  • Discussion on dataset availability and usage
  • Performance of various AI models and training techniques
  • Commercial licensing of Stability AI models

LAION ▷ #research (2 messages):

  • scammer alert
  • new tokenizer proposal for LLMs
  • T-FREE tokenizer paper

Link mentioned: T-FREE: Tokenizer-Free Generative LLMs via Sparse Representations for Memory-Efficient Embeddings: Tokenizers are crucial for encoding information in Large Language Models, but their development has recently stagnated, and they contain inherent weaknesses. Major limitations include computational ov...


LAION ▷ #resources (1 messages):

khazn: $50 gift card steamcommunity.com/gift/sd271azjxn2h


LAION ▷ #learning-ml (1 messages):

khazn: $50 gift card steamcommunity.com/gift/sd271azjxn2h


LAION ▷ #paper-discussion (1 messages):

khazn: $50 gift card steamcommunity.com/gift/sd271azjxn2h


OpenAI ▷ #ai-discussions (116 messages🔥🔥):

  • Moshi AI demo
  • Issues with GPT-2
  • Voice modality in OpenAI models
  • Bangla language support in chatGPT
  • API usage for AI integration

OpenAI ▷ #gpt-4-discussions (26 messages🔥):

  • Differences between free and paid ChatGPT plans
  • Handling images and PDFs in GPT knowledge base
  • Effectiveness of GPT memory
  • Accessing other GPT models within a GPT
  • External file linking and vector databases for GPT knowledge base

OpenAI ▷ #prompt-engineering (16 messages🔥):

  • Employee Recognition Program
  • Content Generation script for training courses
  • Tool to test multiple AI responses
  • Tabletop RPG prompts
  • Traffic ticket challenge guidance

Nous Research AI ▷ #research-papers (1 messages):

teknium: https://x.com/kerstingaiml/status/1809152764649574541?s=46


Nous Research AI ▷ #datasets (1 messages):

  • Replete-AI releases two massive datasets
  • Everything_Instruct and Everything_Instruct_Multilingual datasets
  • Sizes and features of new datasets
  • Influence of bagel datasets and EveryoneLLM AI models

Nous Research AI ▷ #off-topic (4 messages):

  • Upcoming Nous physical magazine contribution
  • Open-source / decentralized technology in StudioMilitary magazine

Link mentioned: Tweet from John Galt (@StudioMilitary): I'm beginning work on the first edition of our magazine. General theme is open-source / decentralized technology. Highlighting the optimistic forces in our world. If you're interested in cont...


Nous Research AI ▷ #interesting-links (5 messages):

  • Apollo project by Achyut Benz
  • flask-socketio-llm-completions GitHub repo
  • foxhop's demo chatroom app
  • LLM integration with flask-socketio

Nous Research AI ▷ #general (110 messages🔥🔥):

  • New datasets released by Replete-AI
  • Nomic AI launches GPT4ALL 3.0
  • InternLM-XComposer-2.5 model release
  • Challenges with jailbreaks for Claude 3.5 Sonnet
  • Discussion on visual latent space for LLMs

Nous Research AI ▷ #ask-about-llms (1 messages):

  • using visual-semantic information to boost image classification performance
  • zero/few shot multi-modal models discussed at CVPR
  • applying Florence 2 for supervised fine-tuning

Nous Research AI ▷ #rag-dataset (8 messages🔥):

  • crossover with pipelines, flows, and agents
  • rag dataset as 0 shot context ingestion
  • context and metadata for llm
  • HF tool processing corpus queries against hf datasets
  • keyword matching for relevance score and filtering

Nous Research AI ▷ #world-sim (10 messages🔥):

  • Users discussing lack of credits to use WorldSIM
  • Issues with using GPT-3.5 on WorldSIM
  • Prompt engineering for different models on WorldSIM
  • Positive feedback about WorldSIM
  • Buddhist world simulation on WorldSIM

OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):

  • Simple Telegram bot to interface with different AI models
  • First 1000 responses free on the bot

OpenRouter (Alex Atallah) ▷ #general (107 messages🔥🔥):

  • Quantisation of deployed LLM models in OpenRouter
  • Microsoft's API changes affecting OpenRouter
  • Infermatic's privacy policy update
  • Issues with DeepSeek Coder equations rendering
  • Mistral Codestral API pricing and performance

Eleuther ▷ #general (42 messages🔥):

  • Failed jobs on the leaderboard
  • Checksum for generative models
  • Topological Data Analysis for model fingerprinting
  • 1.58 bit LLM paper and its implementation
  • VQ-VAE immunity to posterior collapse

Eleuther ▷ #research (28 messages🔥):

  • Diffusion forcing for planning
  • Comparison with Nathan Frey's walk_jump method
  • Discussing new research strategies
  • Continual pre-training for LLMs
  • Function approximation with different homotopy classes

Eleuther ▷ #scaling-laws (5 messages):

  • efficientcube.ipynb in chinchilla repository
  • XLA capabilities in JAX
  • FLOPs estimation for JIT-ed functions in Flax
  • Critical batch size and performance degradation

Eleuther ▷ #interpretability-general (3 messages):

  • SAEs on Llama 3 8B
  • Sparse autoencoders
  • Residual stream processing

Link mentioned: EleutherAI/sae-llama-3-8b-32x · Hugging Face: no description found


Eleuther ▷ #lm-thunderdome (18 messages🔥):

  • Preprocessing Function Optimization
  • Proof-Pile Config Error
  • Metric Inconsistencies in Config
  • Long Model Names Issue
  • Evaluating Model in Parallel

Eleuther ▷ #multimodal-general (1 messages):

wendlerc: Does anyone have a good SDXL latent downscaler? I’d like to go from 128x128x4 to 64x64x4.


LangChain AI ▷ #general (75 messages🔥🔥):

  • Difficulty using LangChain
  • Preference between OpenAI or ChatOpenAI
  • PeopleGPT and Juicebox.ai functionality
  • RAG Architecture for scheduling demos
  • LangChain performance issues and improvements

LangChain AI ▷ #share-your-work (3 messages):

  • Adding demo scheduling feature to chatbot using the RAG architecture and LangChain framework
  • Blogpost on creating an E2E Image Retrieval app using Lightly SSL and FAISS
  • Beta testing for advanced research assistant and search engine with premium model access

LangChain AI ▷ #tutorials (1 messages):

dievas_: https://www.youtube.com/watch?v=yF9kGESAi3M try this one


LlamaIndex ▷ #announcements (1 messages):

  • Next webinar on RAG experimentation/evaluation with LlamaIndex and Weights and Biases
  • Announcements about the timing and focus of the upcoming webinar
  • Complex challenge of aligning LLM Judge for accurate evaluation

Link mentioned: LlamaIndex Webinar: Aligning Your LLM Judge with LlamaIndex and W&B Weave · Zoom · Luma: While creating a RAG pipeline is now straightforward, aligning your LLM Judge for accurate evaluation remains a complex challenge. In this webinar, we’ll delve…


LlamaIndex ▷ #blog (4 messages):

  • New Webinar: A Principled Approach to RAG Experimentation + Evaluation
  • Reflection as a Service
  • Becoming a Rockstar AI Engineer and Educator
  • Corrective RAG as a Service

LlamaIndex ▷ #general (71 messages🔥🔥):

  • Google Cloud Function inference pipeline with multiple model loading
  • Performance comparison of Cohere's command r+
  • Implementing conversational memory in LlamaIndex with RAG
  • Using hybrid retrievers without storing/loading from filesystem
  • Few-shot example technique for 'Poor man's RLHF'

Cohere ▷ #general (45 messages🔥):

  • Discussion about qualifications for attending the London event
  • Issue with deploying an app using Cohere's rerank API in production
  • Introduction of new members
  • Teaching AI and advanced development
  • Working on AI-Plans, a peer review platform for red teaming alignment plans

Cohere ▷ #project-sharing (17 messages🔥):

  • Featuring a tutorial on Cohere blog
  • Introducing Command R+, the new powerful model
  • Using Rhea.run to create toy apps
  • New 'Save to Project' feature in Rhea.run

OpenInterpreter ▷ #general (57 messages🔥🔥):

  • Technical question about interpreter output
  • Discussion on new MacOS Copilot, Invisibility
  • Acknowledgment and progress on Open Interpreter (OI) security
  • Open Interpreter's new debugging feature
  • Monthly House Party events

OpenInterpreter ▷ #O1 (2 messages):

  • 01 Light shipments update
  • Delays in 01 Light shipments

Modular (Mojo 🔥) ▷ #general (5 messages):

  • Discussion on casting bugs in Mojo
  • Comparison between Mojo and Python objects
  • Proposal for a Mojo Fundamentals course at EDx
  • Resources for learning Mojo

Modular (Mojo 🔥) ▷ #🔥mojo (22 messages🔥):

  • Casting file pointer to struct in Mojo
  • Calling external programs in Mojo using system or popen
  • Handling bitcast issues in Mojo with byte array manipulation
  • Pass a List as an argument to a function in Mojo
  • MLIR issue with unsigned integer casting in Mojo

Modular (Mojo 🔥) ▷ #nightly (10 messages🔥):

  • segfault issues with nightly build
  • bug report submission
  • os.path.expanduser bug
  • new nightly Mojo compiler releases
  • changelog updates

Modular (Mojo 🔥) ▷ #mojo-marathons (17 messages🔥):

  • Feedback from Modular staff on best answers
  • Interest in x86 and SVE rounds
  • PR for a better timer needing MLIR knowledge
  • Benny's solution for matrix multiplication
  • Compilation times and segfaults in test suite

Link mentioned: Matrix Multiplication: Sheet1 Contstraints,Parameters / Tuning Vectorization,Contiguous Access,Nelts, Unrollable Parallelization,Unrollable Unrolling,Contiguous Operations Tiling Square Optimized,Amorized Increase,Recursiv...


LLM Finetuning (Hamel + Dan) ▷ #general (12 messages🔥):

  • Usage of LangSmith without LangChain
  • Accusation of lack of GPU credits during AI course
  • 3rd place solution in AI Mathematical Olympiad
  • Benefits of in-context learning vs. fine-tuning

LLM Finetuning (Hamel + Dan) ▷ #🟩-modal (7 messages):

  • Discussion on monthly credits and expiration
  • Distributed finetuning issue solutions
  • Clarifying the usage and remaining balance of credits

LLM Finetuning (Hamel + Dan) ▷ #jarvis-labs (1 messages):

goktrenks: when is the expiration date for the credits? (thanks btw!)


LLM Finetuning (Hamel + Dan) ▷ #ankurgoyal_textsql_llmevals (2 messages):

  • Text2SQL use case discussion and appreciation for iterative eval dataset building

LLM Finetuning (Hamel + Dan) ▷ #workshop-2 (1 messages):

  • Applying eval framework to unstructured applications
  • Challenges of using unit tests/Level 1 evals without structured output

LLM Finetuning (Hamel + Dan) ▷ #jeremy_python_llms (2 messages):

  • Pushing models to HF_HUB for inference endpoints
  • Training models on HF_HUB as endpoints

LLM Finetuning (Hamel + Dan) ▷ #axolotl (3 messages):

  • Using type: input_output with Meta-Llama-3-8B-Instruct
  • Special tokens configuration in Axolotl
  • Training outcomes with L3 8B base vs L3 70B Instruct
  • Template usage for prompt formatting
  • Special tokens setup discrepancies between models

Link mentioned: Hamel’s Blog - Template-free axolotl: Template-free prompt construction in axolotl with the new input_output format.


LLM Finetuning (Hamel + Dan) ▷ #credits-questions (1 messages):

  • Eligibility for credits on all services
  • Enrollment date and course catch-up

LLM Finetuning (Hamel + Dan) ▷ #predibase (1 messages):

  • Expired compute credits
  • Extension request for compute credits

LLM Finetuning (Hamel + Dan) ▷ #openai (1 messages):

  • Credit grant request
  • Enrollment details

Interconnects (Nathan Lambert) ▷ #news (5 messages):

  • Unimpressed reaction to AI demo
  • Stability AI's apology and license update

Interconnects (Nathan Lambert) ▷ #other-papers (8 messages🔥):

  • BM42 vs BM25 in search engines
  • Contextual AI's focus on RAG
  • Jo Bergum's critique of Qdrant's BM42 claims

Interconnects (Nathan Lambert) ▷ #random (9 messages🔥):

  • Understanding VAEs
  • Interconnects' investment genius
  • GDP growth rate from AI for timelines
  • Anthropic Claude 3.5 Sonnet suppressing answers

Link mentioned: Tweet from Philipp Schmid (@_philschmid): I wasn't aware of that, but it looks like Anthropic Claude 3.5 Sonnet on (claude ai) is suppressing parts of his answer from the user, which are not sent to the client. You can test that with, fro...


Interconnects (Nathan Lambert) ▷ #posts (4 messages):

  • Gemini web app
  • Google AI Studio
  • Vertex AI
  • Google's AI race
  • First Amendment and weights

OpenAccess AI Collective (axolotl) ▷ #general (20 messages🔥):

  • Issues with build.nvidia API
  • Queue system for build.nvidia API
  • Script issues and resolutions
  • Pipeline using YAML examples

OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (1 messages):

  • Gemma2 update fixing issues
  • Pinned version of transformers
  • CI catching problems

OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (1 messages):

le_mess: Need more VRAM 🙂


tinygrad (George Hotz) ▷ #general (3 messages):

  • test for bug placement
  • issue reproduction
  • focused test case
  • PR management

tinygrad (George Hotz) ▷ #learn-tinygrad (12 messages🔥):

  • Contiguous Tensors in Tinygrad
  • Tinygrad Training Efficiency Concerns
  • Matrix Multiplication Blog Post
  • Using Pre-trained PyTorch Models with Tinygrad

Torchtune ▷ #general (8 messages🔥):

  • Setting evaluation parameters for Torchtune
  • Grad norm graph on wandb
  • Loss curve optimization in wandb
  • Learning rate adjustment impacts
  • Missing wandb logging metrics

Link mentioned: salman-mohammadi.): Weights & Biases, developer tools for machine learning


AI Stack Devs (Yoko Li) ▷ #ai-town-discuss (5 messages):

  • Investigating system robustness with Python and TypeScript
  • Challenges with automatic Docker installation of Convex local backend

AI Stack Devs (Yoko Li) ▷ #assets (1 messages):

  • Collection of sprite sheets
  • Aesthetics and style matching with Cloudpunk
  • Largest tilemaps on itch.io

DiscoResearch ▷ #general (1 messages):

  • Three GPTs Walk into a Bar and Write an Exec Summary blog post by dsquared70
  • Utilizing Custom GPTs for creating executive summaries
  • Processes for high-frequency, short turnaround executive summaries

Link mentioned: Three GPTs Walk into a Bar and Write an Exec Summary – D-Squared: no description found


DiscoResearch ▷ #discolm_german (2 messages):

  • Magpie model available on HuggingFace Spaces
  • Generating preference data via HuggingFace Spaces
  • Duplicated model from davanstrien/magpie
  • User feedback on Magpie model performance

Link mentioned: Magpie - a Hugging Face Space by sroecker: no description found


MLOps @Chipro ▷ #events (2 messages):

  • Claude hackathon collaboration
  • Kafka optimization webinar

Datasette - LLM (@SimonW) ▷ #llm (1 messages):

  • Potential uses for embeddings
  • New job title 'Embeddings Engineer'




{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}