Frozen AI News archive

not much happened today

**Meta AI** released **Llama 3.2** models including **1B, 3B text-only** and **11B, 90B vision** variants with **128K token context length** and adapter layers for image-text integration. These models outperform competitors like **Gemma 2** and **Phi 3.5-mini**, and are supported on major platforms including **AWS, Azure, and Google Cloud**. **OpenAI CTO Mira Murati** announced her departure. **Allen AI** released **Molmo**, an open-source multimodal model family outperforming proprietary systems. **Google** improved **Gemini 1.5** with Flash and Pro models. **Meta** showcased **Project Orion AR glasses** and hinted at a **Quest 3S** priced at $300. Discussions covered new benchmarks for multimodal models, model optimization, and AI safety and alignment.

Canonical issue URL

AI News for 9/25/2024-9/26/2024. We checked 7 subreddits, 433 Twitters and 31 Discords (224 channels, and 3282 messages) for you. Estimated reading time saved (at 200wpm): 342 minutes. You can now tag @smol_ai for AINews discussions!

Many are still processing the surprise management turnover at OpenAI. Sama and gdb both posted statements. It seems the Anthropic rumors were postponed, but in the meantime the new blueberry model rumor mill is just getting started.


Since it's a quiet day, you could help out AINews by checking out the RAG++ course from Weights and Biases! We featured it yesterday but forgot to include the text link. Sorry!

Swyx: Something we also missed in our initial scan yesterday was chapters 6 and 7 on response synthesis and optimmization. Chapter 6 in particular is exactly what we had to do to build AINews - everything you see below is AI generated thanks to these techniques.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

Meta Releases Llama 3.2 Models

Other AI News

AI Research and Development


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Open Source Vision-Language Models Challenging Proprietary Giants

Theme 2. Llama 3.2: Meta's Multimodal Leap in Open Source AI

Theme 3. Qwen 2.5: Alibaba's Breakthrough in Open Source LLMs

Theme 4. EU AI Regulations Impact on Model Availability and Development

Theme 5. Challenges in Scaling and Reliability of Large Language Models

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

AI Model Advancements and Releases

AI Research and Techniques

AI Industry Developments

AI Applications and Demonstrations

AI Ethics and Regulation

Hardware and Infrastructure


AI Discord Recap

A summary of Summaries of Summaries by O1-mini

Theme 1. Llama 3.2 Model Releases and Performance

Theme 2. AI Model Fine-Tuning and Optimization

Theme 3. Hardware and GPU Discussions for AI

Theme 4. AI Policies and Corporate Shifts

Theme 5. Community Tools and Integrations


PART 1: High level Discord summaries

HuggingFace Discord


Unsloth AI (Daniel Han) Discord


LM Studio Discord


aider (Paul Gauthier) Discord


Nous Research AI Discord


GPU MODE Discord


Stability.ai (Stable Diffusion) Discord


Eleuther Discord


Perplexity AI Discord


OpenAI Discord


Interconnects (Nathan Lambert) Discord


OpenRouter (Alex Atallah) Discord


OpenAccess AI Collective (axolotl) Discord


Latent Space Discord


LlamaIndex Discord


DSPy Discord


Torchtune Discord


Modular (Mojo 🔥) Discord


LangChain AI Discord


Cohere Discord


tinygrad (George Hotz) Discord


LAION Discord


LLM Agents (Berkeley MOOC) Discord


OpenInterpreter Discord


MLOps @Chipro Discord


Alignment Lab AI Discord


Mozilla AI Discord


Gorilla LLM (Berkeley Function Calling) Discord


The LLM Finetuning (Hamel + Dan) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ▷ #announcements (1 messages):

  • Tau Innovations
  • Gemini Object Detection
  • LLama 3.2 Review
  • Reasoning Models for Software Engineering
  • Custom Arabic Semantic Search Model

Link mentioned: Is LLama 3.2 any good ?: Let's try to see if Llama 3.2 is actually any good


HuggingFace ▷ #general (491 messages🔥🔥🔥):

  • OpenAI developments
  • Hugging Face models
  • Machine learning and neuroscience
  • Llama 3.2 updates
  • AI deployment solutions

Links mentioned:


HuggingFace ▷ #today-im-learning (5 messages):

  • AGENTIC RL SWARM
  • Legal Reasoning Models
  • RAG Tool Utilization
  • RL Models in Legal Field
  • CUDA and FP8 Training

HuggingFace ▷ #cool-finds (3 messages):

  • Llama 3.2 Release
  • OpenAI’s New ChatGPT Capabilities
  • New Research Paper

Links mentioned:


HuggingFace ▷ #i-made-this (142 messages🔥🔥):

  • AI Model Comparison Tool
  • Ollama Integration
  • Hugging Face Models
  • Community Feedback
  • Future Enhancements

Links mentioned:


HuggingFace ▷ #computer-vision (5 messages):

  • Multimodal Models
  • Contrastive Learning
  • Visual Learning
  • 3D/4D Understanding

HuggingFace ▷ #NLP (1 messages):

  • Word Injection in Embeddings
  • Parameter Freezing Techniques

Link mentioned: Old dog, new tricks: Word Injection in the text embedding models: In real life, we rarely experience a mind-blowing moment that makes us rethink everything we know. Present experiences are rather built on top of past ones, and gathering knowledge is a gradual proces...


HuggingFace ▷ #diffusion-discussions (2 messages):

  • Colab Free Tier Performance
  • Default Training Logic in Diffusers
  • MSE Loss Computation in Diffusion Models

Link mentioned: Train a diffusion model: no description found


Unsloth AI (Daniel Han) ▷ #general (486 messages🔥🔥🔥):

  • Llama 3.2 Fine-Tuning
  • NVIDIA GPU Specifications
  • LLM Training Techniques
  • Marketing Data Datasets
  • OLLaM and LM Studio Issues

Links mentioned:


Unsloth AI (Daniel Han) ▷ #announcements (1 messages):

  • Llama 3.2 Release
  • New Notebooks
  • Model Uploads

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (7 messages):

  • OpenAI's Corporate Shift
  • Investor Concerns at OpenAI
  • Llama 3.2 on GPU
  • Petscan API Usage
  • Management Vesting Lock-ins

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (110 messages🔥🔥):

  • Fine-Tuning Llama 3.1 & 3.2
  • Unsloth Library Issues
  • Training Configuration
  • Qwen Model Errors
  • Alpaca Template for Fine-Tuning

Links mentioned:


Unsloth AI (Daniel Han) ▷ #research (15 messages🔥):

  • Fine-tuning Unsloth with KTO
  • Flexora for Low-Rank Adaptation
  • INT-FlashAttention GitHub Resource
  • Data Packing in LLM Pre-training
  • Using DeepSpeed for Pre-training GPT-2

Links mentioned:


LM Studio ▷ #general (175 messages🔥🔥):

  • Llama 3.2 Performance
  • Llama 3.2 Vision Models
  • Model Compatibility and Usage
  • Embedding Issues in LM Studio
  • New Models and Updates

Links mentioned:


LM Studio ▷ #hardware-discussion (165 messages🔥🔥):

  • LLM Recommendations for Hardware
  • Performance on Different GPUs
  • VRAM Limitations
  • Integration of Multiple GPUs
  • Upcoming GPU Releases

Links mentioned:


aider (Paul Gauthier) ▷ #general (230 messages🔥🔥):

  • Aider Updates
  • Model Roles in Aider
  • User Experience Improvements
  • Aider Modes
  • Configuration Options

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (21 messages🔥):

  • Augmenting model inputs
  • Issues with Sonnet performance
  • Using Aider with GUI frameworks
  • Dependency update tools
  • File creation issues in Aider

Links mentioned:


aider (Paul Gauthier) ▷ #links (1 messages):

fry69_61685: https://simonwillison.net/2024/Sep/25/o1-preview-llm/


Nous Research AI ▷ #announcements (1 messages):

  • Hermes 3
  • Llama-3.1 8B

Link mentioned: HuggingChat: Making the community's best AI chat models available to everyone.


Nous Research AI ▷ #general (224 messages🔥🔥):

  • Nous Hermes 3
  • Llama 3.2 models
  • GPU advancements
  • Generative AI concerns
  • AI event in Singapore

Links mentioned:


Nous Research AI ▷ #ask-about-llms (17 messages🔥):

  • Llama 3.2 Vision Encoder
  • Inference Requirements for Llama 3.2
  • Llama 3.2 Model Availability
  • DisTRo Optimizer
  • Judgement and Reward Modelling in Hermes 3

Nous Research AI ▷ #research-papers (1 messages):

trre: https://arxiv.org/abs/2409.16897


Nous Research AI ▷ #interesting-links (2 messages):

  • Wordware App Updates
  • Opus Insight Enhancements
  • ThinkLab Features
  • O1-Preview Limitations

Links mentioned:


Nous Research AI ▷ #research-papers (1 messages):

trre: https://arxiv.org/abs/2409.16897


GPU MODE ▷ #general (5 messages):

  • Scam Link Identification
  • Textbook Discussion

GPU MODE ▷ #triton (6 messages):

  • Triton Conference 2024
  • Proton kernel optimization
  • Triton compiled wheel for Windows

Links mentioned:


GPU MODE ▷ #torch (9 messages🔥):

  • PyTorch Memory Profiling
  • TorchDispatchMode
  • Autograd Optimization
  • PyTorch Profiler
  • Memory Visualization

Links mentioned:


GPU MODE ▷ #cool-links (2 messages):

  • Llama 3.2 Models
  • Llama 3.1 Impact
  • Edge and Mobile AI
  • Availability Issues

Link mentioned: no title found: no description found


GPU MODE ▷ #beginner (8 messages🔥):

  • CUDA project setup
  • clangd configuration issues
  • RTX 3070 compute capabilities
  • Neovim vs. other IDEs
  • Channel organization

GPU MODE ▷ #pmpp-book (1 messages):

  • Reading Images in C
  • Chapter 3 Overview

GPU MODE ▷ #torchao (49 messages🔥):

  • torchao compatibility
  • FP8 on CogVideoX
  • Windows support for ML
  • Training issues with NVIDIA
  • Community discussions on development

Links mentioned:


GPU MODE ▷ #off-topic (10 messages🔥):

  • Easy to Use GPU Solutions
  • Prime Intellect on Lambda
  • IRL Hackathon Planning
  • Edge LLM Challenge
  • Team Collaboration for Challenges

Links mentioned:


GPU MODE ▷ #irl-meetup (1 messages):

  • Guatemala Meetups
  • Regional Connections

GPU MODE ▷ #triton-puzzles (2 messages):

  • Numba
  • Convolution and Deep Learning
  • Attention and Transformers

GPU MODE ▷ #llmdotc (25 messages🔥):

  • RoPE forward integration
  • Fused RMSNorm implementation
  • SwiGLU performance
  • Forward pass matching
  • REPKV backward kernel review

Links mentioned:


GPU MODE ▷ #rocm (3 messages):

  • GroupedGemm examples
  • Architecture development

GPU MODE ▷ #arm (5 messages):

  • iPhone 16 Pro
  • bfloat16 enablement
  • M4 SME exploration
  • Apple GPU microarchitecture

Links mentioned:


GPU MODE ▷ #metal (1 messages):

  • Block Scheduling
  • Randomization in Scheduling

GPU MODE ▷ #self-promotion (3 messages):

  • GPU Performance Optimization
  • CUDA Virtual Connect
  • Llama 3.2 with CuPy

Links mentioned:


GPU MODE ▷ #diffusion (11 messages🔥):

  • Creating a separate channel
  • Flux models benchmarks
  • DiffusionKit usage
  • Block diagram for Flux
  • Reddit diagram sharing

Link mentioned: GitHub - argmaxinc/DiffusionKit: On-device Inference of Diffusion Models for Apple Silicon: On-device Inference of Diffusion Models for Apple Silicon - argmaxinc/DiffusionKit


Stability.ai (Stable Diffusion) ▷ #general-chat (135 messages🔥🔥):

  • AI Art Discussions
  • Stable Diffusion User Queries
  • Model Testing and Issues
  • Voice Conversion with RVC
  • User Experience with Lora Models

Link mentioned: bingbangboom/flux_dreamscape · Hugging Face: no description found


Eleuther ▷ #general (2 messages):

  • Advertising Policy
  • Self-taught ML Engineer Introduction

Eleuther ▷ #research (67 messages🔥🔥):

  • MLM transformer with high masking
  • Qwen2 benchmarks concerns
  • FP6 vs BF16 performance
  • Hyperbolic Vision Transformer
  • Verbatim memorization in LLMs

Links mentioned:


Eleuther ▷ #scaling-laws (8 messages🔥):

  • Scaling laws dataset
  • Broken neural scaling laws
  • Chinchilla scaling laws
  • Google research on scaling laws

Link mentioned: google-research/revisiting_neural_scaling_laws/README.md at master · google-research/google-research: Google Research. Contribute to google-research/google-research development by creating an account on GitHub.


Eleuther ▷ #interpretability-general (1 messages):

  • Filler tokens in LLMs
  • Parallelization of tasks
  • Supervision in natural language context

Eleuther ▷ #lm-thunderdome (22 messages🔥):

  • Exact Match Metric Adjustment
  • Debugging Token Generation
  • Improving Documentation for Utils
  • Task Execution Default Behavior
  • Token Counting Script Enhancement

Links mentioned:


Eleuther ▷ #multimodal-general (1 messages):

sky.moo: How can I run vision llms locally?


Eleuther ▷ #gpt-neox-dev (4 messages):

  • FA3 support
  • Training on H100s
  • Testing assistance
  • Grant opportunities from Modal

Perplexity AI ▷ #general (87 messages🔥🔥):

  • Perplexity AI context retention issues
  • Llama 3.2 release
  • Cloudflare verification frustrations
  • Feature request submissions
  • AI model capabilities in lyric generation

Link mentioned: Tweet from Phi Hoang (@apostraphi): haystacks don't stand a chance


Perplexity AI ▷ #sharing (8 messages🔥):

  • Mira Murati departs OpenAI
  • AI vs. reCAPTCHA
  • Energy consumption of AI
  • Garbage In, Garbage Out
  • Bike lanes in cities

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (4 messages):

  • Perplexity in Zapier
  • System and User Content
  • External Models Availability

OpenAI ▷ #ai-discussions (76 messages🔥🔥):

  • Meta AI Access Issues
  • Llama 3.2 Licensing Concerns
  • AI IDE Recommendations
  • Challenges with AI Search Tools
  • Reinforcement Learning Resources

Link mentioned: Llama can now see and run on your device - welcome Llama 3.2: no description found


OpenAI ▷ #gpt-4-discussions (20 messages🔥):

  • Advanced Voice Mode Limitations
  • File Upload Issues in o1
  • Performance Comparison of GPT Models
  • Concerns About Policy Violations
  • Anticipated Features for Future Releases

OpenAI ▷ #prompt-engineering (1 messages):

blckreaper: not if you use o1 mini it will tell you you suck and to do better


OpenAI ▷ #api-discussions (1 messages):

blckreaper: not if you use o1 mini it will tell you you suck and to do better


Interconnects (Nathan Lambert) ▷ #news (20 messages🔥):

  • 3.2 license comparison
  • OpenAI leadership changes
  • Molmo pointing feature excitement
  • Barret Zoph departure
  • Opus project updates

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (44 messages🔥):

  • OpenAI leadership changes
  • Profit interest structures
  • California oversight on non-profits
  • Impact of Anthropic on OpenAI
  • Tax implications for employees

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (21 messages🔥):

  • LLaMA Stack
  • FTC AI Crackdown
  • NeurIPS Submission Rejections
  • Rewardbench Benchmark
  • C++ Knowledge in Academia

Links mentioned:


Interconnects (Nathan Lambert) ▷ #memes (6 messages):

  • Molmo model skepticism
  • Impact of social media
  • LLaMA vs Molmo performance

Link mentioned: Molmo - A new model that outperforms Llama 3.2, available in the E.U: Posted in r/LocalLLaMA by u/franklbt • 104 points and 51 comments


Interconnects (Nathan Lambert) ▷ #posts (6 messages):

  • Meeting duration
  • One-on-ones
  • Project leadership

OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

  • Vision Llama Release
  • Gemini Tokenization Changes
  • Gemini Pricing Adjustments
  • OpenRouter Endpoints

Link mentioned: Tweet from OpenRouter (@OpenRouterAI): Five new endpoints for today's Llama 3.2 release are now live Including a free vision Llama! 🎁🦙


OpenRouter (Alex Atallah) ▷ #general (87 messages🔥🔥):

  • OpenRouter Updates
  • Llama Model Restrictions
  • Chat Completion Models
  • Gemini Performance Changes
  • Math Quiz Rewriting Models

Links mentioned:


OpenRouter (Alex Atallah) ▷ #beta-feedback (2 messages):

  • Bring Your Own Key Beta Test

OpenAccess AI Collective (axolotl) ▷ #general (65 messages🔥🔥):

  • License Compliance Issues
  • OpenAI's CTO Resignation
  • Mistral Discussions
  • Open Source Advancements
  • Continued Pretraining Example

Link mentioned: Tweet from Mira Murati (@miramurati): I shared the following note with the OpenAI team today.


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (4 messages):

  • LoRA+ Issue
  • Multimodal/VLM Assistance

OpenAccess AI Collective (axolotl) ▷ #runpod-help (3 messages):

  • Llama 3.2 Inference
  • GPU Requirements
  • Runpod H100 GPUs
  • Quantization Effects

OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (5 messages):

  • Tokenizer Padding Token
  • Adding Special Tokens
  • Model Resizing for Tokenization

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


Latent Space ▷ #ai-general-chat (73 messages🔥🔥):

  • Mira Murati's Departure
  • Meta's Orion AR Glasses
  • Google's AlphaChip
  • AI Product Creation Platform Arcade
  • GitHub Copilot Browser Integration

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

swyxio: new meetup in <@&1284244976024424630> led by <@656968717883670570> ! https://lu.ma/i8ulstlw


LlamaIndex ▷ #blog (2 messages):

  • Engineering roles hiring
  • NVIDIA competition
  • LLM applications
  • Prizes and rewards
  • Developer resources

Link mentioned: NVIDIA and LlamaIndex Developer Contest: Stand a chance to win cash prizes, a GeForce RTX GPU, and more.


LlamaIndex ▷ #general (47 messages🔥):

  • ReAct Agent Messages
  • VectorStoreIndex Confusion
  • LlamaTrace Project Issues
  • Score Evaluations
  • KnowledgeGraph RAG vs QueryFusion

Links mentioned:


DSPy ▷ #show-and-tell (3 messages):

  • Langtrace DSPy integration
  • DSPy project setup
  • Experiment tracking with Langtrace
  • Code samples for DSPy classification and summarization

Link mentioned: DSPy - Langtrace AI Docs: no description found


DSPy ▷ #general (38 messages🔥):

  • Integration of conversation history in models
  • STORM paper and GitHub resources
  • Creating agents with DSPy
  • Dspy.LM and Azure OpenAI APIs
  • Challenges with TypedChainOfThought

Links mentioned:


DSPy ▷ #examples (3 messages):

  • Number of Classes in Models
  • Subtle Distinctions in Class Signatures

Torchtune ▷ #general (35 messages🔥):

  • Yamashi's Green Card Dilemma
  • Llama 3.2 Access Issues
  • Torchtune Dataset Error
  • MetaAI Login Restrictions
  • Visual Question Answering Datasets

Links mentioned:


Torchtune ▷ #dev (7 messages):

  • NeMo ASR optimizations
  • INT-FlashAttention for INT8
  • Quantization benefits
  • Triton kernel hosting

Links mentioned:


Modular (Mojo 🔥) ▷ #general (2 messages):

  • MOToMGP pass manager error
  • Mojo/MAX branded backgrounds

Modular (Mojo 🔥) ▷ #announcements (1 messages):

  • Verification Process
  • Onboarding Questions
  • Channel Posting Restrictions

Modular (Mojo 🔥) ▷ #mojo (19 messages🔥):

  • Mojo compilation
  • Rust interoperability
  • Error handling improvements

Link mentioned: learn-mojo/packages/mojo-ffi/mojo-call-rust at main · better-mojo/learn-mojo: learn mojo. Contribute to better-mojo/learn-mojo development by creating an account on GitHub.


Modular (Mojo 🔥) ▷ #max (3 messages):

  • MAX API usage
  • User feedback on MAX API

LangChain AI ▷ #general (21 messages🔥):

  • LLM Response Behavior
  • Source Document Retrieval
  • Debugging Tools Usage

Cohere ▷ #discussions (14 messages🔥):

  • Generative AI Bubble
  • PhD student engagement
  • Instructional header formatting
  • Token speculation
  • Community welcoming atmosphere

Link mentioned: The Subprime AI Crisis: None of what I write in this newsletter is about sowing doubt or "hating," but a sober evaluation of where we are today and where we may end up on the current path. I believe that the artifi...


Cohere ▷ #questions (7 messages):

  • Rerank fine-tuning
  • Avg Hard Negatives per Query
  • Model performance and data quality

tinygrad (George Hotz) ▷ #general (5 messages):

  • Arbitrary View Mergeability Proof
  • Pairwise vs Global Mergeability
  • View Merging Optimization
  • View Merging and Symbolic Reduction

Link mentioned: Tinygrad-view-merging-proof/Proof for when a new view must be appended.pdf at 8672f35c1147798c8e9a78bfab28b9ff79bf45e6 · pschilliOrange/Tinygrad-view-merging-proof: Contribute to pschilliOrange/Tinygrad-view-merging-proof development by creating an account on GitHub.


tinygrad (George Hotz) ▷ #learn-tinygrad (12 messages🔥):

  • Tinygrad performance issues
  • Metal compatibility errors
  • Comparison between Tinygrad and PyTorch
  • Tinygrad customization benefits

Link mentioned: GitHub - mesozoic-egg/tinygrad-notes: Tutorials on tinygrad: Tutorials on tinygrad. Contribute to mesozoic-egg/tinygrad-notes development by creating an account on GitHub.


LAION ▷ #general (9 messages🔥):

  • LLaMA 3.2 Vision
  • OpenAI's Function Calling API
  • Voice Cloning Technology

Links mentioned:


LAION ▷ #research (6 messages):

  • MaskBit
  • MonoFormer
  • HuggingFace papers
  • VQGAN enhancements
  • Multimodality transformer models

LLM Agents (Berkeley MOOC) ▷ #mooc-questions (8 messages🔥):

  • Quiz 3 Questions
  • RAG Model Capabilities
  • Social Alignment in LLM Agents
  • CCMP_AI Project
  • Course Registration Confirmation

LLM Agents (Berkeley MOOC) ▷ #mooc-readings-discussion (1 messages):

  • Healthcare multi-agent systems
  • AgentClinic Benchmark

Link mentioned: AgentClinic: a multimodal agent benchmark: Paper Reading: Evaluation of AI in simulated clinical environments


OpenInterpreter ▷ #general (6 messages):

  • Llama 3.2 experiments
  • Open Interpreter issues
  • Tech Week SF meet-up

OpenInterpreter ▷ #ai-content (3 messages):

  • Llama 3.2 Testing
  • NERD Task Challenges

Link mentioned: Llama-3.2 (1B, 3B, 11B, 90B) : The WORST New LLMs EVER!? (Fully Tested & Beats "Nothing"): Join this channel to get access to perks:https://www.youtube.com/@AICodeKing/joinIn this video, I'll be fully testing all the new Llama-3.2 Vision & Small La...


MLOps @Chipro ▷ #general-ml (3 messages):

  • Alternatives to partition_pdf
  • Channel etiquette

Alignment Lab AI ▷ #general (1 messages):

rusch: bruh how is this not promotion


Mozilla AI ▷ #announcements (1 messages):

  • Mozilla AI
  • Locally-run AI models
  • Large Language Models (LLMs)
  • Continue tool
  • Nature article

Gorilla LLM (Berkeley Function Calling) ▷ #discussion (1 messages):

  • Function Calling Evaluation
  • Custom Evaluation Dataset
  • LLM Integration
  • Berkeley Function-Calling Leaderboard

Link mentioned: Berkeley Function Calling Leaderboard: no description found




{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}