Frozen AI News archive

Problems with MMLU-Pro

**MMLU-Pro** is gaining attention as the successor to MMLU on the **Open LLM Leaderboard V2** by **HuggingFace**, despite community concerns about evaluation discrepancies and prompt sensitivity affecting model performance, notably a **10-point improvement** in **Llama-3-8b-q8** with simple prompt tweaks. **Meta's MobileLLM** research explores running sub-billion parameter LLMs on smartphones using shared weights and deeper architectures. **Salesforce's APIGen** introduces an automated dataset generation system for function-calling tasks outperforming larger models. **Runway Gen-3 Alpha** launches an AI video generator for paid users creating realistic 10-second clips. **Nomic AI's GPT4All 3.0** offers an open-source desktop app supporting thousands of local models. AI assistants with multimodal capabilities and affordable access to multiple LLMs like ChatGPT, Claude, Llama, and Gemini are emerging. **Meta 3D Gen** advances text-to-3D asset generation, while Argil AI enables deepfake video creation from text threads. Research on transformer grokking and reasoning highlights advances in robust reasoning capabilities.

Canonical issue URL

AI News for 7/5/2024-7/8/2024. We checked 7 subreddits, 384 Twitters and 29 Discords (462 channels, and 4661 messages) for you. Estimated reading time saved (at 200wpm): 534 minutes. You can now tag @smol_ai for AINews discussions!

There's been a lot of excitement for MMLU-Pro replacing the saturated MMLU, and, ahead of Dan Hendrycks making his own update, HuggingFace has already anointed MMLU-Pro the successor in the Open LLM Leaderboard V2 (more in an upcoming podcast with Clementine). It's got a lot of improvements over MMLU...

image.png

but... the good folks at /r/LocalLlama have been digging into it and finding issues, first with math heaviness, but today more damningly some alarming discrepancies in how models are evaluated by the MMLU-Pro team across sampling params, system prompts, and answer extraction regex:

image.png

For their part, the MMLU-Pro team acknowledge the discrepancies (both between models and between the published paper and what the code actually does) but claim that their samples have minimal impact, but the community is correctly pointing out that the extra attention and customization paid to the closed models disadvantage open models.

Experience does tell us that current models are still highly sensitive to prompt engineering, and simple tweaks of the system prompt improved Llama-3-8b-q8's performance by 10 points (!!??!).

image.png

Disappointing but fixable, and maintaining giant benchmarks are always a messy task, yet one would hope that these simple sources of variance would have been controlled better given the high importance we are increasingly placing on them.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

AI Developments

AI Agents and Assistants

AI Art and Video

AI Research and Techniques

Robotics Developments

AI Culture and Society

Memes and Humor


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

Technology Advancements

Model Releases and Benchmarks

Discussions and Opinions

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries

1. Advancements in Model Architectures and Training

2. Innovations in AI Efficiency and Deployment


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord


HuggingFace Discord


Perplexity AI Discord


LM Studio Discord


OpenAI Discord


Unsloth AI (Daniel Han) Discord


Latent Space Discord


CUDA MODE Discord


Nous Research AI Discord


Modular (Mojo 🔥) Discord


Cohere Discord


Eleuther Discord


LAION Discord


LangChain AI Discord


OpenInterpreter Discord


OpenRouter (Alex Atallah) Discord


LlamaIndex Discord


tinygrad (George Hotz) Discord


OpenAccess AI Collective (axolotl) Discord


LLM Finetuning (Hamel + Dan) Discord


Interconnects (Nathan Lambert) Discord


Alignment Lab AI Discord


LLM Perf Enthusiasts AI Discord


AI Stack Devs (Yoko Li) Discord


MLOps @Chipro Discord


Mozilla AI Discord


DiscoResearch Discord


The Torchtune Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Stability.ai (Stable Diffusion) ▷ #general-chat (804 messages🔥🔥🔥):

  • Model Licensing
  • Performance and Troubleshooting
  • Generation Techniques and Tools
  • Community and Ethical Concerns
  • Image Upscaling Techniques

Links mentioned:


HuggingFace ▷ #general (605 messages🔥🔥🔥):

  • Hermes 2
  • GPTs Agents
  • OpenAI's sidebars
  • Fundraising for AI projects
  • Inference API issues

Links mentioned:


HuggingFace ▷ #today-im-learning (4 messages):

  • Boid AI
  • LLM/GenAI Glossary
  • GPA Predictor with Scikit-Learn
  • Generative Text Project

Link mentioned: Tweet from Prashant Dixit (@Prashant_Dixit0): ✨Open-sourcing comprehensive LLM Glossary✨ Explore, Learn, and Add terms about #LLMs and #GenAI. Let's make AI easy for everyone. 🚨Adding new terms on regular basis Don't forget to give st...


HuggingFace ▷ #cool-finds (16 messages🔥):

  • Claude Artifacts
  • PersonaHub Dataset
  • Pseudonymization Techniques
  • Admin Requests

Links mentioned:


HuggingFace ▷ #i-made-this (24 messages🔥):

  • in10search Tabs Sidepanel AI
  • ZeroGPU HuggingFace Space
  • qdurllm
  • AI on-call developer: merlinn
  • DarkWebSight

Links mentioned:


HuggingFace ▷ #computer-vision (22 messages🔥):

  • Torchmetrics for Object Detection
  • RT-DETR Model Release
  • CogVLM2 for Vision-Language Models
  • Zero-shot Object Detection Models
  • MaskFormer and Instance Segmentation

Links mentioned:


HuggingFace ▷ #NLP (7 messages):

  • Label Error in NLP Dataset
  • Extending deepseek-ai model context length
  • Byte Pair Encoding Implementation in C
  • Comprehensive LLM/GenAI Glossary

Links mentioned:


HuggingFace ▷ #diffusion-discussions (1 messages):

  • Artifacting in sd-vae
  • Common issues in sd-vae reconstruction

HuggingFace ▷ #gradio-announcements (1 messages):

  • Enhanced Documentation Search on Gradio
  • Navigation of Gradio Documentation Pages

Link mentioned: Gradio: Build & Share Delightful Machine Learning Apps


Perplexity AI ▷ #general (502 messages🔥🔥🔥):

  • Issues with Perplexity
  • Pro Search and Limitations
  • Subscription Alternatives
  • Image Generation
  • Technical Problems and Bugs

Links mentioned:


Perplexity AI ▷ #sharing (15 messages🔥):

  • Minecraft Underground Survival
  • Average Cost Research
  • Relational Table Considerations
  • Current Redemption Programs
  • Next iPad Mini Release

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (9 messages🔥):

  • Online model performance
  • API request processing
  • API vs Perplexity search results
  • Beta access delay
  • Multi-step search in API

LM Studio ▷ #💬-general (249 messages🔥🔥):

  • Hermes 2.5
  • Mistral struggles
  • Model Merging
  • Open Empathic
  • IPEX-LLM integration

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (163 messages🔥🔥):

  • Experiences with Different Model Versions
  • Model Performance Issues
  • Model Quantization Discussions
  • Fine-tuning and Customization
  • Categorizing Text Prompts

Links mentioned:


LM Studio ▷ #🧠-feedback (4 messages):

  • x64bit installer for LM Studio
  • Features of LM Studio
  • Community feedback on LM Studio
  • Vision-enabled models
  • Tool calling and model capabilities

LM Studio ▷ #📝-prompts-discussion-chat (1 messages):

  • RAG applications
  • Optimal placement of retrieved context
  • System message vs final user message

LM Studio ▷ #⚙-configs-discussion (3 messages):

  • internllm2_5 config
  • models for understanding PDFs
  • using LMStudio with Shell GPT

Link mentioned: GitHub - TheR1D/shell_gpt: A command-line productivity tool powered by AI large language models like GPT-4, will help you accomplish your tasks faster and more efficiently.: A command-line productivity tool powered by AI large language models like GPT-4, will help you accomplish your tasks faster and more efficiently. - TheR1D/shell_gpt


LM Studio ▷ #🎛-hardware-discussion (44 messages🔥):

  • Snapdragon Elite X Machines
  • RAM upgrades and costs
  • Unified Memory in Windows and Mac
  • External GPUs
  • Feasibility of using Quad Xeon Servers for AI

LM Studio ▷ #🧪-beta-releases-chat (2 messages):

  • Suspicious Activity in Chat
  • Discord Update Delays

LM Studio ▷ #autogen (1 messages):

  • Cost Warning Suppression
  • LM-Studio Configuration
  • Messaging Bug

LM Studio ▷ #amd-rocm-tech-preview (2 messages):

  • LM Studio
  • Generation Speed
  • Fedora 40 Kinoite
  • 7900XTX

LM Studio ▷ #🛠-dev-chat (3 messages):

  • Removing CPU requirement for app
  • Forcing the model to load into RAM
  • GPU offload configuration

OpenAI ▷ #ai-discussions (325 messages🔥🔥):

  • Hermes 2
  • Mistral struggles
  • Model Merging
  • Open Empathic
  • Cloudflare blocking AI bots

Links mentioned:


OpenAI ▷ #gpt-4-discussions (13 messages🔥):

  • GPT-4o vs GPT-4
  • Verification issues
  • Custom GPTs + Zapier integration

OpenAI ▷ #prompt-engineering (3 messages):

  • Content Creation Tips
  • Increasing Engagement
  • Platform Optimization
  • Content Calendar Structure
  • Tracking Metrics for Success

OpenAI ▷ #api-discussions (3 messages):

  • Content creation tips
  • Audience engagement strategies
  • Platform optimization advice
  • Content calendar structure
  • Key metrics for content success

Unsloth AI (Daniel Han) ▷ #general (167 messages🔥🔥):

  • Qwen Model underrating
  • Martin Shkreli presence
  • SLM finetuning practice
  • Unsloth Studio Beta UI
  • AMD vs NVIDIA for LLM training

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (18 messages🔥):

  • Kaggle disk space limit
  • Anthropic's steering method
  • Model pruning
  • AI research community
  • LivePortrait

Link mentioned: Gate: Gate Platform


Unsloth AI (Daniel Han) ▷ #help (120 messages🔥🔥):

  • Training Phi-3 with Alpaca dataset
  • Inference speed and efficiency of Llama-3 vs Phi 3.5 mini
  • Issues with GGUF conversion post training
  • DPO with Gemma 2 27B
  • RAG approach with fine tuned models

Links mentioned:


Unsloth AI (Daniel Han) ▷ #research (13 messages🔥):

  • Asking for help in forums
  • Need for a job channel

Link mentioned: Don't ask to ask, just ask: no description found


Latent Space ▷ #ai-general-chat (26 messages🔥):

  • LLM Coding Efficiency
  • Bug Fix Documentation
  • Inspect AI Framework
  • Dario Amodei's Insights
  • Schedule-Free Optimizers

Links mentioned:


Latent Space ▷ #ai-announcements (5 messages):

  • HN post for podcast
  • Fortnite's new game mode
  • Communication problems at work
  • Upvotes and engagement on HN

Link mentioned: New Links | Hacker News: no description found


Latent Space ▷ #ai-in-action-club (243 messages🔥🔥):

  • AI in Action
  • AI Engineer World Fair
  • LlamaFile vs. Ollama
  • Model Merging
  • Wearables and Privacy

Links mentioned:


CUDA MODE ▷ #general (10 messages🔥):

  • CUDA Certification vs GitHub Repos
  • NVIDIA Deep Learning Institute
  • Peak FLOPS Comparison
  • Educational Expense Strategies

Links mentioned:


CUDA MODE ▷ #torch (10 messages🔥):

  • torch.compile the missing manual
  • PyTorch tensors and type erasure
  • Flexibility vs templates in graph creation
  • PyTorch profiler and FLOP estimation
  • FlopCounterMode vs with_flops

Link mentioned: torch.compile, the missing manual: torch.compile, the missing manual You are here because you want to use torch.compile to make your PyTorch model run faster. torch.compile is a complex and relatively new piece of software, and so you ...


CUDA MODE ▷ #jobs (1 messages):

  • Compiler enthusiasts job opening
  • Thunder compiler optimization project

Links mentioned:


CUDA MODE ▷ #beginner (20 messages🔥):

  • CUDA beginner project
  • Using Python vs. C++ for CUDA
  • Starting with CUDA
  • Tinygrad and Teenygrad
  • SpMM with 2:4 sparsity pattern

CUDA MODE ▷ #torchao (19 messages🔥):

  • 2:4 sparsity with int8 quantization
  • Hackable pure Python low-bit optimizers
  • Non-contiguous gradients issue
  • FP8 Adam optimization
  • Regression tests on CI machines

Links mentioned:


CUDA MODE ▷ #off-topic (13 messages🔥):

  • AliExpress Anniversary Promo
  • Creative Pixel Art Tool
  • Summer Vacation for Startup Founders
  • Techsupportgore Subreddit
  • Potential Online Scams

Links mentioned:


CUDA MODE ▷ #irl-meetup (1 messages):

fancytrevor: curious if anyone has sf meetup recommendations


CUDA MODE ▷ #llmdotc (179 messages🔥🔥):

  • muP Experiments
  • FP8 Precision
  • CUDA Checkpointing
  • Inference Optimizations
  • LLM101n Course Plan

Links mentioned:


Nous Research AI ▷ #research-papers (3 messages):

  • Critiques in Preference Learning
  • Test-Time-Training Layers

Links mentioned:


Nous Research AI ▷ #off-topic (2 messages):

  • Nous Magazine
  • Cryptoland
  • YouTube video
  • Fantasy Division

Link mentioned: Whatever Happened to Cryptoland?: there is no way anyone could've seen this coming⚔️ You're gunna wanna check this out: https://fantasydivision.online/References: https://docs.google.com/docu...


Nous Research AI ▷ #interesting-links (2 messages):

  • Dataline by RamiAwar
  • LLM Reasoning Capabilities

Links mentioned:


Nous Research AI ▷ #general (211 messages🔥🔥):

  • GPT4 Benchmark Scores
  • Temperature Effects
  • In-Context Learning Examples
  • Prompt Caching Costs
  • BitNet Training

Links mentioned:


Nous Research AI ▷ #ask-about-llms (2 messages):

  • Deterministic Reports with LLMs
  • Integration of Traditional Programming

Nous Research AI ▷ #rag-dataset (4 messages):

  • RAG and Hallucinations
  • Wikipedia-Style Citations
  • RAG and Hugging Face Agents

Links mentioned:


Nous Research AI ▷ #world-sim (5 messages):

  • WorldSIM simulation success
  • Next era of simulation

Modular (Mojo 🔥) ▷ #general (68 messages🔥🔥):

  • Updating WSL to WSL2 for Mojo Installation
  • Dependency Hell in Python
  • Mojo Rounding Function Bugs
  • Getting 'Mojician' Badge
  • Mojo's Int64 vs Float64 Behavior

Links mentioned:


Modular (Mojo 🔥) ▷ #🔥mojo (10 messages🔥):

  • __del__ method in Mojo
  • Mojo 3D Graphics Examples
  • Common libc functions in Mojo
  • Cross compilation with Mojo
  • Using partition method in Mojo

Link mentioned: lightbug_http/external/libc.mojo at main · saviorand/lightbug_http: Simple and fast HTTP framework for Mojo! 🔥. Contribute to saviorand/lightbug_http development by creating an account on GitHub.


Modular (Mojo 🔥) ▷ #nightly (9 messages🔥):

  • Mojo compiler nightly releases
  • Changelog updates and PRs
  • C pointer semantics removal
  • Expression color changes
  • Tuple unpacking request

Modular (Mojo 🔥) ▷ #mojo-marathons (77 messages🔥🔥):

  • Optimization of Matmul Algorithm
  • SIMD and Cache Performance
  • Compile Time Issues with Mojo
  • Autotuning for Performance

Links mentioned:


Cohere ▷ #general (67 messages🔥🔥):

  • AI-Plans platform
  • Using the rerank API
  • Cohere community introductions
  • Meta Learning by Radek Osmulski
  • Dark theme for Coral Chat Interface

Links mentioned:


Cohere ▷ #project-sharing (78 messages🔥🔥):

  • Rhea Platform
  • AI Creation and Interaction
  • Organizational Accounts
  • User Experience Feedback
  • Coding Club Projects with Children

Links mentioned:


Eleuther ▷ #announcements (1 messages):

  • Top-k Sparse Autoencoders
  • Llama 3 8B
  • Automated Pipeline for SAE Features
  • Training SAEs for 70B Model

Eleuther ▷ #general (33 messages🔥):

  • PDF markup tools
  • Training variable resolution ViT using IJEPA
  • Evaluating LlaVa LLM
  • Randomness of Copilot
  • Determinism in LLM completions

Links mentioned:


Eleuther ▷ #research (67 messages🔥🔥):

  • T-FREE Tokenizer
  • Research on Model Expansion Efficiency
  • The BitNet Transformer
  • Gradient Conflicts in Diffusion Models
  • Quantization in Inference

Links mentioned:


Eleuther ▷ #scaling-laws (3 messages):

  • Attention as Hypernetwork
  • Empirical results of Attention as Hypernetwork

Link mentioned: Attention as a Hypernetwork: Transformers can under some circumstances generalize to novel problem instances whose constituent parts might have been encountered during training but whose compositions have not. What mechanisms und...


Eleuther ▷ #interpretability-general (2 messages):

  • Mech Interp Reading List v2
  • Opinionated List of Favourite Papers
  • Mechanistic Interpretability
  • Reading List
  • Literature Review

Link mentioned: An Extremely Opinionated Annotated List of My Favourite Mechanistic Interpretability Papers v2 — AI Alignment Forum: This post represents my personal hot takes, not the opinions of my team or employer. This is a massively updated version of a similar list I made two…


Eleuther ▷ #lm-thunderdome (32 messages🔥):

  • LlaVa LLM evaluation
  • Feature requests in lm-evaluation-harness
  • AISI's Inspect vs. lm-eval harness
  • Long-context evaluation benchmarks

Links mentioned:


Eleuther ▷ #multimodal-general (1 messages):

wendlerc: Does anyone have a good SDXL latent downscaler? I’d like to go from 128x128x4 to 64x64x4.


Eleuther ▷ #gpt-neox-dev (1 messages):

  • Docker container usage
  • GPT-Neox deployment
  • Kubernetes for large-scale jobs
  • Docker Compose vs. Kubernetes

LAION ▷ #general (41 messages🔥):

  • JPEG XL Image Codec
  • Kolors GitHub Repository
  • Noise Scheduling in Machine Learning
  • Meta VLM Ads
  • IJEPA Training with Variable Resolution ViT

Links mentioned:


LAION ▷ #research (28 messages🔥):

  • VALL-E 2
  • Terminator model discussion
  • New caption model - CapPa
  • VisualKeras tool

Links mentioned:


LangChain AI ▷ #general (52 messages🔥):

  • Handling CSV files in LangChain
  • LangChain utility functions
  • LangGraph setup issues
  • Running LLMs locally
  • Async configuration in LangChain

Links mentioned:


LangChain AI ▷ #langserve (1 messages):

  • LangServe Deployment Issues
  • LangGraph Cloud Announcement

LangChain AI ▷ #share-your-work (10 messages🔥):

  • doesVideoContain
  • qdurllm
  • OranScribe
  • LLM glossary
  • advanced research assistant

Links mentioned:


LangChain AI ▷ #tutorials (2 messages):

  • LangGraph state
  • Langchain + Graph RAG + GPT-4o

Links mentioned:


OpenInterpreter ▷ #general (32 messages🔥):

  • Skills library with RAG
  • Security prioritization by OI team
  • GraphRAG
  • 4th of July house party
  • Langchain in RAG system

Links mentioned:


OpenInterpreter ▷ #O1 (10 messages🔥):

  • Shipment Timeline
  • O1 Talking Capability
  • Text Display Options
  • Google I/O Demo Glasses
  • Linux Module Error

OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

  • New model browser UI
  • Noromaid Mixtral deprecation

Link mentioned: Tweet from OpenRouter (@OpenRouterAI): Announcing a brand-new model marketplace UI ✨ Explore 180 active language models processing 74 billion tokens/week 👇


OpenRouter (Alex Atallah) ▷ #app-showcase (6 messages):

  • Viinyx AI launch
  • Text to image API services

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (27 messages🔥):

  • Crypto payments
  • Perplexity models
  • Generative video future
  • OpenRouter provider options
  • Model pricing competition

LlamaIndex ▷ #blog (6 messages):

  • Agentic RAG for Stock Trading
  • Toolkits for RAG dataset generation
  • Agents as Microservices
  • Multi-Document Financial Analyst Agent
  • RAG Retrieval Evaluations

LlamaIndex ▷ #general (21 messages🔥):

  • AI application mentorship
  • Claude 3 models in Bedrock
  • Knowledge graphs from GitHub code
  • Structured data queries with LlamaIndex
  • ReAct agent observations

Links mentioned:


tinygrad (George Hotz) ▷ #general (7 messages):

  • team red's drivers
  • Instinct cards
  • custom grads API
  • tinygrad functions
  • Monday team meeting

tinygrad (George Hotz) ▷ #learn-tinygrad (20 messages🔥):

  • requires_grad behavior
  • multi-GPU training docs
  • tensor comparison method
  • Adam optimizer issue
  • new methods for Tinygrad tensors

OpenAccess AI Collective (axolotl) ▷ #general (8 messages🔥):

  • Model Merging
  • MInference
  • RAM Issue
  • Offload Config

Link mentioned: GitHub - microsoft/MInference: To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which reduces inference latency by up to 10x for pre-filling on an A100 while maintaining accuracy.: To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which reduces inference latency by up to 10x for pre-filling on an A100 while maintaining accu...


OpenAccess AI Collective (axolotl) ▷ #community-showcase (1 messages):

  • Yi-1.5-9B-Chat training
  • Hermes-2.5 integration
  • Benchmark results
  • Future plans for extended context length

Link mentioned: juvi21/Hermes-2.5-Yi-1.5-9B-Chat · Hugging Face: no description found


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (5 messages):

  • chat_template
  • mistral finetuning in axolotl

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


LLM Finetuning (Hamel + Dan) ▷ #general (8 messages🔥):

  • MLOps implementation
  • Distributed VLLM inference
  • FP8 quantization issues
  • Chat template challenges

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #replicate (1 messages):

  • Replicate billing setup issues

LLM Finetuning (Hamel + Dan) ▷ #jeremy_python_llms (1 messages):

  • Transformers & Torch
  • Integrating with OpenAI/Anthropic models

LLM Finetuning (Hamel + Dan) ▷ #credits-questions (1 messages):

  • Credit Claims Closed
  • Credit Eligibility

LLM Finetuning (Hamel + Dan) ▷ #predibase (1 messages):

4.8.15.16.23.42_: the first 25 credits are available for all but only for 1 month 🙂


Interconnects (Nathan Lambert) ▷ #news (2 messages):

  • Interconnects Bot Feedback

Interconnects (Nathan Lambert) ▷ #other-papers (8 messages🔥):

  • RAG discussions
  • Enterprises and RAG
  • RAG use cases
  • early AI boom
  • retrieval and cost efficiency

Alignment Lab AI ▷ #general-chat (6 messages):

  • Buzz excitement
  • FPGA meeting
  • Calendly scheduling

Link mentioned: meeting - Auto Meta: no description found


LLM Perf Enthusiasts AI ▷ #general (1 messages):

jeffreyw128: wow flash 1.5 is actually so good


AI Stack Devs (Yoko Li) ▷ #assets (1 messages):

  • Google image searches for sprites
  • Purchased assets for tilesets

MLOps @Chipro ▷ #events (1 messages):

jonononono: Anyone going to europython? Doing a talk on vectorization 👀


Mozilla AI ▷ #llamafile (1 messages):

  • Gemma 2 9B
  • Small Language Models (SLMs)
  • Serverless AI inference

Link mentioned: Serverless AI Inference with Gemma 2 using Mozilla's llamafile on AWS Lambda: Google's Gemma 2 9B is a recently released open-source language model that has garnered significant attention in our community. This lightweight model, is part of the Gemma family of models devel...


DiscoResearch ▷ #discolm_german (1 messages):

  • Experiment with base models
  • Hermes-2-Theta-Llama-3-70B
  • Llama3-DiscoLeo-Instruct-70B


{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}