Frozen AI News archive

We Solved Hallucinations

**Reddit's URL structure causes link errors in AI-generated summaries, especially with NSFW content affecting models like Claude and GPT-4.** The team fixed this glitch while still leveraging LLMs for summarizing Reddit content. **GPT-2 training costs have dramatically dropped to ~$672 using H100 GPUs and software improvements like CUDA and FlashAttention.** **FlashAttention-3 was released, achieving up to 740 TFLOPS on H100 GPUs, with FP8 nearing 1.2 PFLOPS, developed collaboratively by Meta, NVIDIA, Princeton, and Colfax.** Hopper GPUs enable major speedups with new hardware features. **Synthetic data may not improve vision tasks, as shown in recent research.** The **Avocado360 benchmark evaluates vision-language models' ability to detect avocados in images.** **Lynx, a hallucination detection model for LLMs, was introduced for real-world healthcare and fintech applications, trained by Patronus AI on Databricks Mosaic AI using Composer.**

Canonical issue URL

AI News for 7/11/2024-7/12/2024. We checked 7 subreddits, 384 Twitters and 29 Discords (463 channels, and 2566 messages) for you. Estimated reading time saved (at 200wpm): 276 minutes. You can now tag @smol_ai for AINews discussions!

Look, we've known for a while that our Reddit summaries are ridden with... erm... links that don't go where they claim to go. You keep reminding us! (Thanks!)

image.png

The reason that this happens specifically to our Reddit summaries much much more than our Discord or Twitter recaps is because of Reddit's URL structure.

Here is a typical Reddit URL:

https://www.reddit.com/r/LocalLLaMA/comments/1cxnrov/disappointing_if_true_meta_plans_to_not_open_the/

The slug at the end (disappointing_if_true_meta_plans_to_not_open_the) is just an attempt to make a human readable slug out of the title, AND the subreddit at the start (r/LocalLLaMA) is also just for human readability. In practice, all of it is ignored in favor of the real slug, that 7 character alphanumeric set (1cxnrov). Here, we'll prove it:

https://www.reddit.com/r/SmolAI/comments/1cxnrov/ainews_is_the_best/

Despite having changed the subreddit and the human slug, Reddit sends you to the same post as before based on the real slug.

So Reddit URLs, much more than most URLs, are hyper, hyper sensitive to small mistakes in attention, even if all we are asking the LLM to do is copy from a source docment with the reference link neatly spelled out.

And... both Claude and GPT4 are trained on an awful lot of NSFW Reddit URLs (in multiple languages!). Put these two facts together and you can see what we've been dealing with.

So.. we went ahead and fixed the glitch, while still using LLMs to format, select, and summarize across a full corpus of Reddit submissions and comments. Tweet @Smol_AI if you have guesses on how we do it.

It's been another content light day, so please enjoy our conversation with Clementine Fourrier on LLM Evals (our coverage in May and the future of the Open LLM Leaderboard:

https://www.youtube.com/watch?v=E-UhbYc8m24


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

Compute and Hardware Improvements

LLM Evaluation and Benchmarking

LLM Applications and Frameworks

Memes and Humor


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

Theme 1. WizardLM 3 and LLM Optimization Techniques

Theme 2. Advanced AI-Generated Visual Content

Theme 3. AI Progress Tracking and Benchmarking

Theme 4. AI Content Regulation and Copyright Issues


AI Discord Recap

A summary of Summaries of Summaries

1. LLM Advancements and Training Techniques

2. Open Source AI Advancements

3. Community Collaboration and Knowledge Sharing

4. Hardware Benchmarking and Adoption


PART 1: High level Discord summaries

HuggingFace Discord


Stability.ai (Stable Diffusion) Discord


CUDA MODE Discord


Modular (Mojo šŸ”„) Discord


Unsloth AI (Daniel Han) Discord


Nous Research AI Discord


Perplexity AI Discord


Eleuther Discord


Latent Space Discord


LangChain AI Discord


LM Studio Discord


OpenAI Discord


Cohere Discord


Interconnects (Nathan Lambert) Discord


tinygrad (George Hotz) Discord


OpenAccess AI Collective (axolotl) Discord


OpenRouter (Alex Atallah) Discord


LAION Discord


LlamaIndex Discord


OpenInterpreter Discord


LLM Finetuning (Hamel + Dan) Discord


Mozilla AI Discord


DiscoResearch Discord


MLOps @Chipro Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI Stack Devs (Yoko Li) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Torchtune Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ā–· #announcements (1 messages):

  • qdurllm demo
  • Advanced RAG workshop
  • Intel HF model repository
  • Self-reviewing coding assistant
  • Training chatbot with LlamaIndex

Links mentioned:


HuggingFace ā–· #general (613 messagesšŸ”„šŸ”„šŸ”„):

  • GPU Models and Issues
  • Cloud and Free Resources
  • Training Techniques
  • HF Integrations
  • Jokes and Community Engagement

Links mentioned:


HuggingFace ā–· #today-im-learning (6 messages):

  • Embedding models using mouse movements
  • Transfer learning in triplet loss
  • Classification objectives in contrastive learning
  • Sampling rates and batch sizes
  • Knowledge graphs implementation

Link mentioned: Support Vector Machine SVM ( Machine Learning pt 3 ): In this video i try to explain SVMs from the basic and I try to make it easy and simple , if ya wanna know about the SK-Learn click this : https://youtu.be/...


HuggingFace ā–· #cool-finds (2 messages):

  • Supervised fine-tuning in TRL
  • Ripple_net library for search engines

Links mentioned:


HuggingFace ā–· #i-made-this (11 messagesšŸ”„):

  • mypo dataset
  • Indonesian Hate Speech dataset
  • ripple_net library
  • RAG app for PDFs
  • Support Vector Machine SVM video

Links mentioned:


HuggingFace ā–· #reading-group (3 messages):

  • Paper Plans
  • Transformer Performance
  • New LLM Paradigm

HuggingFace ā–· #core-announcements (1 messages):

  • AuraFlow model
  • LoRA support
  • Offloading at the modeling level
  • State-of-the-art results on GenEval
  • Community feedback

Links mentioned:


HuggingFace ā–· #computer-vision (3 messages):

  • RT-DETR Object Detection
  • Hiera Vision Transformer

Links mentioned:


HuggingFace ā–· #NLP (4 messages):

  • LLM Finetuning Toolkit
  • Phi-3 models discussion
  • Multimodal image RAG

Link mentioned: GitHub - georgian-io/LLM-Finetuning-Toolkit: Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.: Toolkit for fine-tuning, ablating and unit-testing open-source LLMs. - georgian-io/LLM-Finetuning-Toolkit


HuggingFace ā–· #diffusion-discussions (1 messages):

  • Architecture Explanation
  • Implementation from Scratch

Stability.ai (Stable Diffusion) ā–· #general-chat (341 messagesšŸ”„šŸ”„):

  • Reactor installation for ComfyUI
  • Deforum Stable Diffusion techniques
  • AIl troubleshooting delays
  • Model merging and performance
  • Effective image upscalers

Links mentioned:


CUDA MODE ā–· #general (13 messagesšŸ”„):

  • FA3 collaboration
  • H100 deployment
  • Warpgroup Pingponging
  • Support for Ampere
  • Discord functionality

CUDA MODE ā–· #torch (7 messages):

  • ResNet18 on A100 vs A40
  • torch.compile max-autotune issue
  • Floating point errors

CUDA MODE ā–· #cool-links (3 messages):

  • Q-Galore paper
  • Llama3 405B release
  • LoQT paper

Links mentioned:


CUDA MODE ā–· #beginner (18 messagesšŸ”„):

  • GPU access
  • Google Colab
  • Coreweave and Lambda Labs
  • Google Cloud GPU
  • Nsight Compute

CUDA MODE ā–· #pmpp-book (6 messages):

  • Matrix-Matrix Multiplication in CUDA
  • Thread Assignment in Matmul Kernels
  • Data Access Patterns in CUDA

CUDA MODE ā–· #torchao (3 messages):

  • Tensor Subclass Support
  • Bitnet Work
  • FSDP and Distributed Training
  • Developer Guide

CUDA MODE ā–· #llmdotc (176 messagesšŸ”„šŸ”„):

  • FA3 vs cuDNN vs ThunderKittens
  • Unified Memory impact
  • Fine-tuning large models
  • FP8 optimizations
  • LLM.C community initiatives

Links mentioned:


CUDA MODE ā–· #youtube-watch-party (1 messages):

vkaul11: Hi


Modular (Mojo šŸ”„) ā–· #general (189 messagesšŸ”„šŸ”„):

  • Primeagen Video Discussion
  • REPL Behavior in Mojo
  • Mojo Community Meeting
  • GIL Removal in Python 3.13
  • Comparative Network Speed in Different Languages

Links mentioned:


Modular (Mojo šŸ”„) ā–· #polls (2 messages):

  • MAX framework
  • new website
  • NVIDIA GPU performance
  • PyTorch & ONNX optimization
  • Mojo programming language

Link mentioned: Modular: Own your endpoint. Control your AI.: The Modular Accelerated Xecution (MAX) platform is the worlds only platform to unlock performance, programmability, and portability for your AI workloads.


Modular (Mojo šŸ”„) ā–· #mojo (17 messagesšŸ”„):

  • rust-lang/mdbook for Mojo documentation
  • Mojo playground capabilities
  • Mojo standard library documentation
  • Mojo-LSP support
  • Mojo in production environments

Links mentioned:


Modular (Mojo šŸ”„) ā–· #max (5 messages):

  • MAX Model execution
  • Mojo support for VariadicLists
  • PythonObjects as inputs
  • Numpy Arrays in Modularity
  • Data type issues in model execution

Modular (Mojo šŸ”„) ā–· #max-gpu (7 messages):

  • Nightly changelog
  • Custom GPU kernels
  • Max vs Mojo

Modular (Mojo šŸ”„) ā–· #nightly (2 messages):

  • New Mojo compiler release
  • EqualityComparable issue

Unsloth AI (Daniel Han) ā–· #general (148 messagesšŸ”„šŸ”„):

  • Hermes 2.5 performance
  • Mistral extension
  • Model Merging Strategies
  • Open Empathic Project
  • Gemini API Updates

Links mentioned:


Unsloth AI (Daniel Han) ā–· #off-topic (23 messagesšŸ”„):

  • Open Diloco
  • Distributed GPU Workloads
  • CodeGeeX4-ALL-9B
  • Prompt Engineering
  • TF-ID Models

Links mentioned:


Unsloth AI (Daniel Han) ā–· #help (28 messagesšŸ”„):

  • Training Local Models
  • Continued Pretraining
  • Training Data Recommendations
  • Model Parameter Discrepancies
  • Resource for RAG Systems

Links mentioned:


Unsloth AI (Daniel Han) ā–· #showcase (1 messages):

lh0x00: <@280027697328029696> , have you had experience with Spanish yet?


Unsloth AI (Daniel Han) ā–· #community-collaboration (5 messages):

  • Evaluating Programming Models
  • Tetris as a Benchmark
  • Coding Models and Dataset Overlap

Unsloth AI (Daniel Han) ā–· #research (12 messagesšŸ”„):

  • Ada and WGMMA/TMA/FP8 compatibility
  • Decoder as Embedding Model
  • Latent Array in training
  • Meta's LLaMA 3 model release

Link mentioned: nvi (flad): no description found


Nous Research AI ā–· #off-topic (7 messages):

  • OpenAI Model Capabilities
  • Anthropic's AI Safety Levels
  • OpenAI Strategy
  • Community Opinions on OpenAI

Link mentioned: Tweet from AI Notkilleveryoneism Memes āøļø (@AISafetyMemes): OpenAI just told employees at an all-hands meeting they’re ā€œon the cuspā€ of models capable of ā€œproblem-solving tasks as well as a human with a doctorate-level education.ā€ (Re-read that: DOCTOR. LEVEL...


Nous Research AI ā–· #interesting-links (2 messages):

  • GPT-2 reproduction in llm.c
  • Safetensors.cpp

Links mentioned:


Nous Research AI ā–· #general (132 messagesšŸ”„šŸ”„):

  • Multi-threaded asynchronous FSM in Rust
  • Issues with Hermes 2 AI Assistant
  • VRAM Requirements for LLaMA 3
  • Fine-tuning LLMs without answers
  • Improving LLMs' reasoning with prompting

Link mentioned: gist:b8257a67933d891a9f3bc19822b4305a: GitHub Gist: instantly share code, notes, and snippets.


Nous Research AI ā–· #rag-dataset (3 messages):

  • Surya new models
  • Marker speedup
  • Model merging

Links mentioned:


Nous Research AI ā–· #world-sim (4 messages):

  • Terminal of truths
  • Learning to learn
  • Embodiment of models

Perplexity AI ā–· #general (132 messagesšŸ”„šŸ”„):

  • Perplexity Labs and Usage
  • Claude 3.5 vs Claude 3 Opus
  • Perplexity Outage and Issues
  • Coding Concerns with Perplexity
  • Subscription Models and AI Preferences

Links mentioned:

Write the entire...: Certainly! I'll modify the existing code to include local data storage using SQLite. This will allow the application to store and retrieve candle data...Perplexity: Perplexity is a free AI-powered answer engine that provides accurate, trusted, and real-time answers to any question.Perplexity: Perplexity is a free AI-powered answer engine that provides accurate, trusted, and real-time answers to any question.Perplexity Pro: Perplexity Pro is the most powerful way to search the internet with unlimited Pro Search, upgraded AI models, unlimited file upload, image generation, and API credits.


Perplexity AI ā–· #sharing (5 messages):

  • AI in Diabetes Management
  • Perplexity AI Discord Community
  • New Developments: Smart Ring, Vending Machines, Redbox Shutdown
  • Rule of Thirds in Photography
  • Health, Strength, and Power Tips

Links mentioned:


Perplexity AI ā–· #pplx-api (10 messagesšŸ”„):

  • Error 524 in API response
  • Switching models causing issues
  • Backend deployment with Discord and Perplexity API
  • API errors with VPN usage
  • Cloudflare causing issues

Eleuther ā–· #general (116 messagesšŸ”„šŸ”„):

  • GPT-4Chan and TruthfulQA
  • Utility of Benchalmarks
  • Jsonnet's Role in Configuration
  • London AI Meetups
  • Importance of Staying Updated with Research

Links mentioned:


Eleuther ā–· #research (15 messagesšŸ”„):

  • Memory bandwidth increases for GPT-2 training
  • Quantization techniques for LLMs
  • Breakdown of SOTA LLMs on simple problems
  • Temporal distances in stochastic MDPs
  • Causal reasoning from passive data

Links mentioned:


Eleuther ā–· #scaling-laws (1 messages):

  • Neocortex Neuron Count
  • Body Mass : Brain Mass
  • Intelligence Metrics

Eleuther ā–· #lm-thunderdome (1 messages):

  • lm-eval Python API
  • Implementing a server for model evaluation
  • Converting custom models

Latent Space ā–· #ai-general-chat (37 messagesšŸ”„):

  • FlashAttention-3 release
  • OpenAI's revenue rumors
  • OpenAI AGI level framework
  • Decentralized AI training
  • Compound AI systems funding

Links mentioned:


Latent Space ā–· #ai-announcements (1 messages):

swyxio: new podcast drop! https://x.com/swyx/status/1811898574416019562


Latent Space ā–· #ai-in-action-club (86 messagesšŸ”„šŸ”„):

  • 3E acronym
  • Logprob Evaluation
  • Langgraph State Management
  • PDF to Markdown tools
  • RAG Architectures

Links mentioned:


LangChain AI ā–· #general (68 messagesšŸ”„šŸ”„):

  • Chroma Vector Store
  • OpenAI Embedding Function
  • FAISS vs Chroma
  • LangChain Agents and Tools
  • Using OpenAI Vector Store as Retriever

Links mentioned:


LangChain AI ā–· #share-your-work (1 messages):

  • Structured Data Synthesis
  • Indexify
  • Towards AI Publication

Link mentioned: Tweet from Prashant Dixit (@Prashant_Dixit0): Structured data Extraction from Unstructured pipelines Used Indexify by @tensorlake for this example. Indexify is a data framework created to build ingestion and extraction pipelines for unstructured...


LM Studio ā–· #šŸ’¬-general (25 messagesšŸ”„):

  • Dell Inspiron 3847 upgrades and limitations
  • NPU support in x elite
  • FlashAttention for LLMs
  • Debugging GPU issues in Linux
  • Shifting to Linux from Windows

Links mentioned:


LM Studio ā–· #šŸ¤–-models-discussion-chat (8 messagesšŸ”„):

  • Salesforce Einstein
  • Local Model Benchmarks

Link mentioned: Dubesor LLM Benchmark table: no description found


LM Studio ā–· #šŸŽ›-hardware-discussion (19 messagesšŸ”„):

  • 3090 vs 4090 for AI
  • NVIDIA 5090 rumors
  • Multi-GPU setups for AI
  • V100 compute nodes
  • Performance of ARM computers with LLMs

Link mentioned: Reddit - Dive into anything: no description found


LM Studio ā–· #amd-rocm-tech-preview (6 messages):

  • OpenCL backend issues
  • Cuda vs ROCM
  • Vulkan support

LM Studio ā–· #šŸ› -dev-chat (10 messagesšŸ”„):

  • Setting up RAG in React with LM Studio
  • Negative experiences in Discord dev channels
  • Discussion about Rust vs C++
  • LM Studio SDK for integration

OpenAI ā–· #ai-discussions (19 messagesšŸ”„):

  • Decentralized AI Computation
  • Automated Personalized SEO
  • OpenAI's Tier System
  • Claude vs ChatGPT on Document Reading
  • GPT-4o and Sora Update

Links mentioned:


OpenAI ā–· #gpt-4-discussions (25 messagesšŸ”„):

  • ChatGPT-5 release speculation
  • Optimizing ChatGPT configurations
  • ChatGPT-4o performance
  • New features expected in ChatGPT-5
  • DALL-E image generation issues

OpenAI ā–· #prompt-engineering (2 messages):

  • Chatbot with RAG
  • Contradiction in instructions

OpenAI ā–· #api-discussions (2 messages):

  • RAG chatbot development
  • Prompt contradictions

Cohere ā–· #general (35 messagesšŸ”„):

  • Command R Plus model use cases
  • AI news automation with Cohere
  • Cohere toolkit open source
  • Creating unique emojis
  • OpenArena on GitHub

Links mentioned:


Cohere ā–· #project-sharing (3 messages):

  • Embedding Model Cost Reduction
  • Project-Sharing Etiquette

Interconnects (Nathan Lambert) ā–· #news (23 messagesšŸ”„):

  • Llama 3 release
  • OpenAI's new project Strawberry
  • Self-hosting large models
  • API vs self-hosting costs
  • Sensitive data handling with large models

Links mentioned:


Interconnects (Nathan Lambert) ā–· #ml-questions (4 messages):

  • Soft-target distillation
  • Mistral-7B instruct-finetuning
  • Language model instruction tuning processes
  • AgentInstruct paper
  • KnowledgePile and AutoMathText datasets

Interconnects (Nathan Lambert) ā–· #random (7 messages):

  • GPT-4 Pricing
  • OpenAI's AGI Progress
  • Self-driving Similarities
  • New GPT-4 Skills
  • OpenAI Revenue Speculation

Links mentioned:


tinygrad (George Hotz) ā–· #general (2 messages):

  • Indexing Kernel in tinygrad
  • PyTorch 2024 H2 Roadmaps

Links mentioned:


tinygrad (George Hotz) ā–· #learn-tinygrad (26 messagesšŸ”„):

  • Custom Weight and Bias in Network
  • Implementing Gradient Descent from Scratch
  • Performance Issues with Manual Gradient Descent
  • Tensor Operations and Realization
  • Indexing Tensors and Kernel Performance

Links mentioned:


OpenAccess AI Collective (axolotl) ā–· #general (15 messagesšŸ”„):

  • H100 performance
  • Attention masking in reward models
  • OpenRouter API usage
  • Flash attention versions
  • OpenArena open-source project

OpenAccess AI Collective (axolotl) ā–· #axolotl-dev (5 messages):

  • GaLore and Q-Galore
  • Dataset Shuffling

Link mentioned: Paper page - Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients: no description found


OpenAccess AI Collective (axolotl) ā–· #general-help (1 messages):

  • LoRA finetuning
  • Layer selection
  • Few-shot learning challenges

OpenAccess AI Collective (axolotl) ā–· #datasets (1 messages):

  • General-purpose multi-turn chat dataset
  • Dataset recommendations

OpenAccess AI Collective (axolotl) ā–· #axolotl-help-bot (1 messages):

wasamikirua: how can i push model to the hub after lora merge ?


OpenRouter (Alex Atallah) ā–· #general (15 messagesšŸ”„):

  • Training Data Concerns
  • Integrations (Beta)
  • Prompting Image Models
  • 405b Model Update
  • Specialized Models

Links mentioned:


LAION ā–· #general (2 messages):

  • Clip retrieval
  • Dataset access

LAION ā–· #research (10 messagesšŸ”„):

  • MLP Architecture Efficiency
  • Memory Usage in Model Training
  • Nematron 340B Code Examples
  • AuraFlow Model Announcement
  • Alice in Wonderland (AIW) Problem

Links mentioned:


LlamaIndex ā–· #blog (4 messages):

  • Agentic RAG Cookbooks
  • Cypher Snippet for Entity Deduplication
  • Knowledge Graph Creation Challenges
  • LlamaCloud Data Pipeline Management

Link mentioned: blogs/llm/llama_index_neo4j_custom_retriever.ipynb at master Ā· tomasonjo/blogs: Jupyter notebooks that support my graph data science blog posts at https://bratanic-tomaz.medium.com/ - tomasonjo/blogs


LlamaIndex ā–· #general (6 messages):

  • Function calling on Gemini models
  • Error with Gemin-1.5-flash-latest model
  • Updating vertexai integration package
  • Indexing large code library
  • Reviewing spec documents with RAG

Link mentioned: Enable Function calling and agent runner for Vertex AI by wadave Ā· Pull Request #14088 Ā· run-llama/llama_index: Description Changed are highlighed: Enabled function calling for Vertex AI llama-index-integrations/llms/llama-index-llms-vertex/llama_index/llms/vertex/base.py Added tool/function roles for Gemin...


OpenInterpreter ā–· #general (6 messages):

  • Agent Invocation Issue
  • LiteLLM Error
  • Phi-3 Fast Function Calls
  • Open Interpreter GUI Integration

Links mentioned:


OpenInterpreter ā–· #ai-content (1 messages):

notnaton: https://youtu.be/SoFepHI6sQ0?si=2Y1zkghH2XyaN9_k


LLM Finetuning (Hamel + Dan) ā–· #general (1 messages):

  • Self-hosted ML Telemetry
  • Langfuse
  • WandB
  • OpenLLMTelemetry

LLM Finetuning (Hamel + Dan) ā–· #asia-tz (2 messages):

  • API key for OpenAI
  • Chatbot project tutorial

LLM Finetuning (Hamel + Dan) ā–· #fireworks (1 messages):

  • Account Credits
  • User Query
  • Account ID
  • Credit Balance Check

LLM Finetuning (Hamel + Dan) ā–· #openai (1 messages):

slac.eth6408: Do we know the date of OpenAI credit expiration?


Mozilla AI ā–· #announcements (1 messages):

  • Llamafile goes to Washington
  • Builders Accelerator
  • Upcoming Events
  • Open Source AI Definition

Link mentioned: The Open Source AI Definition – draft v. 0.0.8: version 0.0.8 Leave comments for this text Note: This document is made of three parts: A preamble, stating the intentions of this document; the Definition of Open Source AI itself; and a checklist …


Mozilla AI ā–· #llamafile (1 messages):

  • llama.cpp matmul/matvec
  • ggml-quants.c file
  • integer dotproducts
  • float activations

DiscoResearch ā–· #disco_judge (2 messages):

  • LLM Arena
  • WizardLM Paper
  • OpenArena GitHub Repository

Links mentioned:


MLOps @Chipro ā–· #general-ml (1 messages):

  • Product coverage
  • Research coverage
  • Recommendation systems
  • Information Retrieval
  • Retrieval-Augmented Generation

{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}