Frozen AI News archive

super quiet day

**AI21 Labs** released **Jamba 1.5**, a scaled-up State Space Model optimized for long context windows with **94B parameters** and up to **2.5X faster inference**, outperforming models like **Llama 3.1 70B** on benchmarks. The **Phi-3.5** model was praised for its safety and performance, while **Dracarys**, a new **70B open-source coding model** announced by **Bindu Reddy**, claims superior benchmarks over Llama 3.1 70B. Discussions on **California's SB 1047** AI safety legislation involve **Stanford** and **Anthropic**, highlighting a balance between precaution and industry growth. Innovations include **uv virtual environments** for rapid setup, **LangChain's LangSmith** resource tags for project management, and multi-agent systems in **Qdrant** enhancing data workflows. Community events like the **RAG workshop** by **AWS**, **LangChain**, and **Elastic** continue to support AI learning and collaboration. Memes remain a popular way to engage with AI industry culture.

Canonical issue URL

AI News for 8/21/2024-8/22/2024. We checked 7 subreddits, 384 Twitters and 30 Discords (214 channels, and 2393 messages) for you. Estimated reading time saved (at 200wpm): 283 minutes. You can now tag @smol_ai for AINews discussions!

There are a LOT of whispers flying around about the coming AI releases this fall, but nothing publicly citable, sorry.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

AI Models and their Evaluations

AI Safety and Legislation

AI Tools and Innovations

Conferences & Meetups

Humor and Memes


This provides a comprehensive summary of key discussions from the AI High Signal Twitter list with a detailed focus on model performance, safety, tools, innovations, community events, and humor. Each category draws from multiple sources to ensure the narrative remains well-grounded and informative.


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Microsoft's Phi-3.5 Models: Capabilities and Controversies

Theme 2. AI for Creative Writing and Roleplay

All AI Reddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

AI Image Generation and Training

AI and Software Development

Prosthetics and Biotechnology


AI Discord Recap

A summary of Summaries of Summaries by GPT4O-Aug (gpt-4o-2024-08-06)

1. LLM Model Releases and Features

2. Performance and Optimization Techniques

3. Data Handling and Preprocessing

4. Community and Collaboration Initiatives

5. AI in Industry Applications


PART 1: High level Discord summaries

LM Studio Discord


HuggingFace Discord


Unsloth AI (Daniel Han) Discord


aider (Paul Gauthier) Discord


Stability.ai (Stable Diffusion) Discord


CUDA MODE Discord


OpenRouter (Alex Atallah) Discord


Nous Research AI Discord


OpenAccess AI Collective (axolotl) Discord


LlamaIndex Discord


Perplexity AI Discord


Modular (Mojo 🔥) Discord


Cohere Discord


OpenAI Discord


Eleuther Discord


MLOps @Chipro Discord


Interconnects (Nathan Lambert) Discord


LangChain AI Discord


Latent Space Discord


OpenInterpreter Discord


AI21 Labs (Jamba) Discord


tinygrad (George Hotz) Discord


Torchtune Discord


LAION Discord


Gorilla LLM (Berkeley Function Calling) Discord


DSPy Discord


Mozilla AI Discord


DiscoResearch Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Finetuning (Hamel + Dan) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

LM Studio ▷ #announcements (1 messages):

  • LM Studio 0.3.0
  • LM Studio Features
  • LM Studio Community
  • LM Studio UI
  • LM Studio Models

Links mentioned:


LM Studio ▷ #general (412 messages🔥🔥🔥):

  • LM Studio 0.3.0
  • Gemma 2
  • LLaMa 3.1
  • LM Studio performance
  • LM Studio UI

Links mentioned:


LM Studio ▷ #hardware-discussion (72 messages🔥🔥):

  • SLI/NVLink for LLMs
  • GPU Memory for Large Models
  • Model Size vs Speed
  • Waiting for New Hardware
  • GPU Recommendations

Link mentioned: Funny Very GIF - Funny Very Sloth - Discover & Share GIFs: Click to view the GIF


HuggingFace ▷ #announcements (1 messages):

  • Offensive Security Reconnaissance
  • Deep Learning Course
  • Unity ML Agents
  • Garfield dataset
  • Tensor parallelism

Links mentioned:


HuggingFace ▷ #general (246 messages🔥🔥):

  • Hugging Face prepaid credit system
  • AI21 Labs Jamba 1.5
  • GPU and VRAM
  • HackAI Challenge
  • Prepaid cards

Links mentioned:

#art": 5 likes, 0 comments - noaroggendorff on August 21, 2024: "glimmer #art". Mr Krabs Money GIF - Mr Krabs Money Spongebob - Discover & Share GIFs: Click to view the GIFReddit - Dive into anything: no description foundLotr Lord Of The Rings GIF - LOTR Lord Of The Rings Theoden - Discover & Share GIFs: Click to view the GIFCat Cat Meme GIF - Cat Cat meme Cat staring - Discover & Share GIFs: Click to view the GIFDrama Queen GIF - Drama Queen Dramaticing - Discover & Share GIFs: Click to view the GIFIamproudofyou My Hero GIF - Iamproudofyou My Hero - Discover & Share GIFs: Click to view the GIFHackAI - Dell and NVIDIA Challenge: Code, Create, Conquer - Build groundbreaking Generative AI projects using NVIDIA AI WorkbenchFree personality test | 16Personalities: no description foundPersonality Types | 16Personalities: no description foundPersonality Types | 16Personalities: no description foundno title found: no description found


HuggingFace ▷ #today-im-learning (7 messages):

  • Neuralink Paper Selection
  • Coding with Neuralink
  • Reading Research Papers

HuggingFace ▷ #cool-finds (2 messages):

  • 3DGS
  • ShapeSplat dataset
  • Gaussian Splats
  • Self-Supervised Pretraining
  • Point Cloud Representation

Link mentioned: ShapeSplat: A Large-scale Dataset of Gaussian Splats and Their Self-Supervised Pretraining: no description found


HuggingFace ▷ #i-made-this (10 messages🔥):

  • Fullbound
  • Patch Tripper
  • Offensive Security Reconnaissance
  • On-device Transcription
  • Inferless

Links mentioned:


HuggingFace ▷ #reading-group (1 messages):

  • Language Alignment Techniques
  • DPO Paper
  • Direct Preference Optimization

HuggingFace ▷ #computer-vision (1 messages):

  • Swin Transformers as Mask R-CNN Backbone

HuggingFace ▷ #NLP (2 messages):

  • Multilingual NLP Research

Link mentioned: Call for Reviewers - 4th Multilingual Representation Learning (MRL) Workshop, EMNLP 2024 : This form is for anyone who is interested in being a reviewer for our workshop at EMNLP 2024. Reviewers are invited to indicate their interest and their application will be assessed in relation to th...


HuggingFace ▷ #diffusion-discussions (9 messages🔥):

  • Quanto Qint2 Quantization
  • NeuroSama API
  • CFG++ Support in Diffusers
  • Fine-tuning Diffusion Models with LoRA
  • Importing 3D Models into Images

Link mentioned: Shows how to run Flux schnell under 17GBs without bells and whistles. It additionally shows how to serialize the quantized checkpoint and load it back.: Shows how to run Flux schnell under 17GBs without bells and whistles. It additionally shows how to serialize the quantized checkpoint and load it back. - inference_with_torchao_serialized.py


Unsloth AI (Daniel Han) ▷ #general (153 messages🔥🔥):

  • Gemma-2-27B-it fine-tuning
  • Mistral Nemo 12b
  • Unsloth Pro
  • Unsloth training
  • Unsloth multi-gpu

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (15 messages🔥):

  • Mistral-NeMo-Minitron-8B-Base
  • Data Preprocessing
  • Tokenization

Link mentioned: nvidia/Mistral-NeMo-Minitron-8B-Base · Hugging Face: no description found


Unsloth AI (Daniel Han) ▷ #help (79 messages🔥🔥):

  • Mistral Fine-tuning
  • Unsloth Installation
  • Ollama Installation
  • Inference Issues
  • Stop Tokens

Links mentioned:

  PyTorch

: no description foundSigma Handshak Handshake GIF - Sigma Handshak Handshake Khar - Discover & Share GIFs: Click to view the GIFtext_classification_scripts/unsloth_classification.ipynb at main · timothelaborie/text_classification_scripts: Scripts for text classification with llama and bert - timothelaborie/text_classification_scriptsUnsloth AI: Our mission is to make LLMs for everyone 🦥. Unsloth AI has 7 repositories available. Follow their code on GitHub.Conda installation detailed instructions · Issue #73 · unslothai/unsloth: I'm trying to follow the instructions for installing unsloth in a conda environment, the problem is that the conda gets stuck when running the install lines. I've tried running it twice, both ...


Unsloth AI (Daniel Han) ▷ #showcase (3 messages):

  • The Living AI Dataset
  • Empathy and Love in AI
  • Speech to Text and Text to Speech AI
  • Heroism in the Modern World
  • Hope in a Culture

Links mentioned:


Unsloth AI (Daniel Han) ▷ #research (1 messages):

mahiatlinux: https://huggingface.co/papers/2408.03314


aider (Paul Gauthier) ▷ #general (125 messages🔥🔥):

  • Aider Shell Commands
  • Playwright Installation
  • CodeCompanion
  • Aider Token Usage
  • OpenRouter

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (92 messages🔥🔥):

  • Aider Installation Issues
  • Ollama DeepCoder Freezing
  • Aider and Git
  • Using Aider With Sonnet
  • Optimizing Token Usage

Links mentioned:


aider (Paul Gauthier) ▷ #links (5 messages):

  • Vercel v0 chat
  • Aider
  • Cursor

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (173 messages🔥🔥):

  • ComfyUI
  • Stable Diffusion
  • Flux
  • AI image sorters
  • Hydrus

Links mentioned:


CUDA MODE ▷ #torch (43 messages🔥):

  • Triton INT8 performance
  • Flash Attention FP8 support
  • FP8 quantization for attention
  • FP8 quantization in general
  • Row-wise quantization for weights

Links mentioned:


CUDA MODE ▷ #announcements (1 messages):

  • PyTorch
  • CUDA
  • LLM Training
  • GPU
  • PyTorch Extensions

Links mentioned:


CUDA MODE ▷ #cool-links (2 messages):

  • 2:4 Sparsity
  • Tetris Clone for PSX

Link mentioned: GitHub - jbreckmckye/notris: Tetris clone for PlayStation 1 (PSX): Tetris clone for PlayStation 1 (PSX). Contribute to jbreckmckye/notris development by creating an account on GitHub.


CUDA MODE ▷ #jobs (3 messages):

  • HRT internships
  • HRT trading
  • Algo Dev internships
  • SWE internships
  • HRT market making

CUDA MODE ▷ #beginner (2 messages):

  • CUDA introduction
  • CUDA resources

CUDA MODE ▷ #youtube-recordings (1 messages):

budr0001: If you click on the link in my reply it will take you to the lecture 16 post.


CUDA MODE ▷ #torchao (40 messages🔥):

  • INT8 Mixed Precision Training
  • FP8 Adam
  • Character AI Training
  • 4-bit Adam Optimizer
  • TensorCore INT8 Support

Links mentioned:


CUDA MODE ▷ #sequence-parallel (1 messages):

ericauld: Also tree-based, from June: https://arxiv.org/abs/2401.10774


CUDA MODE ▷ #off-topic (8 messages🔥):

  • Remote Work Contracts
  • Office Transition

CUDA MODE ▷ #irl-meetup (3 messages):

  • Triton Conf
  • GPU Enjoyers
  • Triton Language

Link mentioned: GitHub - triton-lang/triton: Development repository for the Triton language and compiler: Development repository for the Triton language and compiler - triton-lang/triton


CUDA MODE ▷ #hqq-mobius (1 messages):

  • Model Distillation
  • GPU limitations
  • Logit Compression
  • Sparsification
  • Quantization

CUDA MODE ▷ #llmdotc (12 messages🔥):

  • H100 L2 Side Hashing
  • GPU Performance and Power Efficiency
  • FP8 Training Stability
  • Neuralink Person
  • Llama Model Training

Link mentioned: Tweet from xr-5 🐀 (@xariusrke): 1/n FP8 training is hard - loss divergence and instability often lead to the conclusion that it’s not possible. But we’ve found a recipe to train a 1B LLaMA model to match the convergence of bfloat16 ...


CUDA MODE ▷ #rocm (24 messages🔥):

  • RDNA4 vs RDNA3
  • AMD GPU Performance
  • FA3 Implementation
  • 7900 Series vs 3090
  • Triton and FA Fork

Link mentioned: Examining AMD’s RDNA 4 Changes in LLVM: As 2024 continues on, because time never stops, AMD has been working on their upcoming RDNA 4 architecture. Part of this involves supporting open source projects like LLVM. If done right, merging t…


CUDA MODE ▷ #cudamode-irl (1 messages):

kitrak_rev: Everyone got mail? Guess i was rejected in the list then 😦


OpenRouter (Alex Atallah) ▷ #announcements (1 messages):

  • OpenRouter tool parameters

OpenRouter (Alex Atallah) ▷ #app-showcase (3 messages):

  • BenchmarkAggregator
  • LLM Evaluation Framework
  • Oz's Projects

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (95 messages🔥🔥):

  • Llama 3.1 Tools
  • OpenRouter MoE
  • OpenRouter context limits
  • OpenAI fine-tuning
  • Cursor Composer

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

  • Nous Merch Store Launch

Link mentioned: Nous Research: Nous Research


Nous Research AI ▷ #general (64 messages🔥🔥):

  • NousResearch Hermes 3
  • OpenAI Compute Grants
  • AI21 Jamba
  • Live2D
  • The_Living_AI_Dataset

Links mentioned:


Nous Research AI ▷ #ask-about-llms (19 messages🔥):

  • AI Agents
  • Langchain
  • Building your own Agent
  • Discord Bot for impersonating friends
  • Learning from scratch

Nous Research AI ▷ #interesting-links (12 messages🔥):

  • CUDA utils
  • PDF cleaning
  • VLMs for Content Extraction
  • ColPali
  • ColBERT

Links mentioned:


Nous Research AI ▷ #reasoning-tasks (1 messages):

gwyntel: this IS the smoke spot, we got wedding cake and northern lights on deck!


OpenAccess AI Collective (axolotl) ▷ #general (22 messages🔥):

  • Mistral Fine-Tuning
  • Jamba 1.5
  • Memory Consumption
  • Model Selection
  • Phi3.5

Link mentioned: ai21labs/AI21-Jamba-1.5-Mini · Hugging Face: no description found


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (71 messages🔥🔥):

  • Phi 3.5 Mini
  • Flash Attention
  • Exploding Gradients
  • Chat Template Issues
  • Model Merging

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (1 messages):

  • accelerate fp8

LlamaIndex ▷ #blog (4 messages):

  • LlamaCloud
  • LlamaIndex 0.11
  • RAG Pipeline
  • BoxReaderTextExtraction
  • Workflows

LlamaIndex ▷ #general (84 messages🔥🔥):

  • Ollama memory usage
  • LlamaIndex Property graph
  • SchemaLLMPathExtractor
  • Query Generation with LlamaIndex
  • PandasAI for CSV data

Links mentioned:


LlamaIndex ▷ #ai-discussion (2 messages):

  • LlamaIndex
  • RAG
  • multi-strategy workflow
  • graph store
  • knowledge graph

Link mentioned: Multi-Strategy Workflow with Reflection in LlamaIndex: An In-depth Exploration: Ankush k Singal


Perplexity AI ▷ #general (67 messages🔥🔥):

  • Perplexity API
  • Mistral Large 2
  • Perplexity's Sources
  • Perplexity Image Generation
  • Perplexity Subscription Plans

Link mentioned: What is Collections?: Explore Perplexity's blog for articles, announcements, product updates, and tips to optimize your experience. Stay informed and make the most of Perplexity.


Perplexity AI ▷ #sharing (11 messages🔥):

  • Lore's Emotional Instability
  • Microplastics in Brains
  • Retirement Preparation
  • Jonathan Ive's San Francisco Investment
  • Nightcore Music Effects

Links mentioned:


Perplexity AI ▷ #pplx-api (6 messages):

  • Perplexity API
  • Domain Filtering
  • Citations

Link mentioned: Chat Completions: Generates a model's response for the given chat conversation.


Modular (Mojo 🔥) ▷ #general (12 messages🔥):

  • Github Desktop
  • Git send-email
  • Git am
  • Modular Community Manager
  • Modular Calendar

Link mentioned: Modular Community Chat ☕ - Caroline Frasca: no description found


Modular (Mojo 🔥) ▷ #mojo (33 messages🔥):

  • math.isclose open source
  • Mojo github search issues
  • Mojo Docs Search
  • Mojo async/sync performance
  • Mojo stdlib and import paths

Links mentioned:


Modular (Mojo 🔥) ▷ #max (8 messages🔥):

  • Mojo/MAX installation issues
  • Modular CLI
  • Debian installation
  • Mac installation
  • venv activation

Links mentioned:


Cohere ▷ #discussions (21 messages🔥):

  • RAG Development & Evaluation Strategies
  • Cohere x w&b Webinar
  • Cohere Developer Office Hours
  • C4AI Program
  • Chunking XLSX Files

Link mentioned: Cohere For AI (C4AI): Cohere For AI is a non-profit research lab that seeks to solve complex machine learning problems. We support fundamental research that explores the unknown, and are focused on creating more points of ...


Cohere ▷ #questions (18 messages🔥):

  • LLM Agent Task Determination
  • Cohere API 403 Error
  • VPN and API Access
  • Chunking XLSX Files
  • C4ai

Cohere ▷ #api-discussions (7 messages):

  • 403 Forbidden
  • Cohere API from R
  • Cohere API using curl
  • Cohere Command R+ 128k context
  • OpenAI structured outputs

Link mentioned: Chat Non-streaming — Cohere: Generates a text response to a user message. To learn how to use the Chat API with Streaming and RAG follow our Text Generation guides .


Cohere ▷ #projects (6 messages):

  • Jozu Hub
  • Cohere Model Support
  • ModelKit

Link mentioned: no title found: no description found


OpenAI ▷ #ai-discussions (30 messages🔥):

  • Expected Value of an Item
  • AI Assistance with Math Problems
  • Ideogram 2.0 Review
  • SwarmUI & Flux
  • French Community

OpenAI ▷ #gpt-4-discussions (13 messages🔥):

  • GPT Training Data
  • Life Coach App
  • Custom GPTs
  • GPT4o vs GPT4
  • GPT Formatting

OpenAI ▷ #prompt-engineering (1 messages):

madame_architect: I'm not sure I know exactly what you're asking for. Could you say more?


OpenAI ▷ #api-discussions (1 messages):

madame_architect: I'm not sure I know exactly what you're asking for. Could you say more?


Eleuther ▷ #general (10 messages🔥):

  • Open Source AI Models
  • Eleuther.ai
  • DPO Fine-tuning
  • Instruction Prompt Templates
  • Multi-turn Chat Data Prep

Links mentioned:


Eleuther ▷ #research (27 messages🔥):

  • Model Evaluation
  • Model Performance Degradation
  • Benchmark Track Rebuttals
  • Model Merging
  • Model Distillation

Link mentioned: ShortCircuit: AlphaZero-Driven Circuit Design: Chip design relies heavily on generating Boolean circuits, such as AND-Inverter Graphs (AIGs), from functional descriptions like truth tables. While recent advances in deep learning have aimed to acce...


Eleuther ▷ #scaling-laws (1 messages):

catboy_slim_: it is a specific dataset that is so far as i can tell still publicly available


Eleuther ▷ #lm-thunderdome (7 messages):

  • HellaSwag evaluation
  • lm-evaluation-harness
  • generate_until
  • filtering pipeline
  • log likelihood

Links mentioned:


MLOps @Chipro ▷ #general-ml (37 messages🔥):

  • LightGBM
  • LSTM
  • Commodity Price Forecasts
  • Kaggle Competitions
  • Production Data

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (31 messages🔥):

  • AI burnout
  • Frontier Labs
  • Greg Brockman's Work Hours
  • Twitter anxiety

Interconnects (Nathan Lambert) ▷ #nlp (6 messages):

  • Lilian Weng Blog
  • Diffusion Models
  • Generative Models
  • Distillation
  • Score-based generative modeling

Link mentioned: What are Diffusion Models?: [Updated on 2021-09-19: Highly recommend this blog post on score-based generative modeling by Yang Song (author of several key papers in the references)]. [Updated on 2022-08-27: Added classifier-free...


LangChain AI ▷ #general (33 messages🔥):

  • LLM for NL to SQL
  • Prebuilt Queries for SQL
  • LLM Accuracy and SQL
  • RAG for SQL
  • ChromaDB RAG for CSV

LangChain AI ▷ #share-your-work (2 messages):

  • Flags for AI
  • 4149 AI
  • Proactive Guidance in Slack
  • AI for Research

Links mentioned:


Latent Space ▷ #ai-general-chat (29 messages🔥):

  • Ideogram 2.0
  • Mistral-NeMo-Minitron-8B
  • Sovereign AI
  • v0 updates
  • AI development trends

Links mentioned:


Latent Space ▷ #ai-announcements (3 messages):

  • GPT-4o fine-tuning
  • Genie coding agent
  • SWE-Bench
  • RAG
  • In-context learning

Link mentioned: Tweet from swyx.ai (@swyx): 🆕 @latentspacepod: Is finetuning GPT4o worth it? w/ @AlistairPullen of @cosine_sh Betteridge's law says no: with 59 different flavors of RAG, and >2million token context + prompt caching, it...


OpenInterpreter ▷ #general (16 messages🔥):

  • Open Interpreter usability
  • Open Interpreter searching issues
  • Open Interpreter model issues
  • Open Interpreter feature requests
  • Open Interpreter model suggestions

AI21 Labs (Jamba) ▷ #announcements (1 messages):

  • Jamba 1.5
  • Jamba 1.5 Mini
  • Jamba 1.5 Large
  • SSM-Transformer Architecture
  • Long Context Handling

Links mentioned:


AI21 Labs (Jamba) ▷ #jamba (4 messages):

  • Jamba-1.5 Fine Tuning
  • Jamba-1.5 Large

Link mentioned: Jamba 1.5 Release: Jamba 1.5 Release: A Hybrid SSM-Transformer Model for Agentic AI\n\nAI21 has announced the release of Jamba 1.5, a new version of its hybrid SSM-Transformer model that combines the strengths of b...


AI21 Labs (Jamba) ▷ #general-chat (2 messages):

  • API Rate Limits
  • OpenAI API Rate Limits

tinygrad (George Hotz) ▷ #general (4 messages):

  • Code Review
  • mypyc Compilation
  • Tinygrad

Torchtune ▷ #general (4 messages):

  • Torchtune and T5
  • Weight Mapping

Links mentioned:


LAION ▷ #general (3 messages):

  • LinkedIn Survey
  • Infinite Generative YouTube

Link mentioned: Survey | Professional Network Platforms: The most powerful, simple and trusted way to gather experience data. Start your journey to experience management and try a free account today.


LAION ▷ #research (1 messages):

  • Multilingual Representation Learning Workshop at EMNLP 2024
  • Reviewer Sign Up
  • Workshop Topics

Link mentioned: Call for Reviewers - 4th Multilingual Representation Learning (MRL) Workshop, EMNLP 2024 : This form is for anyone who is interested in being a reviewer for our workshop at EMNLP 2024. Reviewers are invited to indicate their interest and their application will be assessed in relation to th...


Gorilla LLM (Berkeley Function Calling) ▷ #leaderboard (2 messages):

  • Website Leaderboard
  • Huggingface Leaderboard
  • Model Evaluation
  • Equal Importance

Gorilla LLM (Berkeley Function Calling) ▷ #discussion (1 messages):

  • Evaluating fine-tuned model on BFCL locally

DSPy ▷ #general (1 messages):

amanshrestha: anyway we can use prompt caching , antropic api?


Mozilla AI ▷ #announcements (1 messages):

  • Open Source AI
  • OSI Definition

DiscoResearch ▷ #mixtral_implementation (1 messages):

  • German Instruction Tuning Data


{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}