Frozen AI News archive

a quiet weekend

**OpenAI** released the new **o1** model, leveraging reinforcement learning and chain-of-thought prompting to excel in reasoning benchmarks, achieving an IQ-like score of **120**. **Google DeepMind** introduced **DataGemma** to reduce hallucinations by connecting LLMs with real-world data, and unveiled **ALOHA** and **DemoStart** for robot dexterity using diffusion methods. **Adobe** previewed its **Firefly AI Video Model** with text-to-video and generative extend features. **Mistral** launched the multimodal **Pixtral 12B** model, and **Tencent** presented the **GameGen-O** open-world video game generation model. Several research papers from **Stanford**, **OpenAI**, **Microsoft**, **Mila**, and **Notre Dame** focus on advanced reasoning, self-verification, and reflection tuning techniques. Experts like **Terence Tao** and **George Hotz** have shared mixed but optimistic views on o1's capabilities. Seed funding rounds include **Supermaven** ($12M) and **11x** ($24M).

Canonical issue URL

AI News for 9/13/2024-9/16/2024. We checked 7 subreddits, 433 Twitters and 30 Discords (220 channels, and 6976 messages) for you. Estimated reading time saved (at 200wpm): 757 minutes. You can now tag @smol_ai for AINews discussions!

Everyone spent the weekend exploring o1 and opinions are quite mixed so far:

image.png

Astrophysics PhDs and George Hotz and Terence Tao like it, and someone manually graded it with an IQ of 120 on a custom IQ quiz.

In other news:

One has to wonder just how good the upcoming Gemini 2 will have to be to compare with o1...


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

AI Model Developments and Industry Updates

AI Research and Papers

AI Capabilities and Benchmarks

Industry Perspectives and Debates


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Llama 3.1 405B: Open-Source Rival to GPT-4

Theme 2. O1 Model's Advanced Reasoning Capabilities

Theme 3. Comparing Online LLM Providers and Services

Theme 4. Advancements in Local LLM Tools and Applications

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

AI Model Advancements and Capabilities

AI Research and Infrastructure

AI Model Releases and Comparisons

Societal and Economic Impacts of AI

AI Applications and Tools


AI Discord Recap

A summary of Summaries of Summaries by O1-preview

Theme 1: OpenAI's O1 Models Ignite Debate in AI Community

Theme 2: AI Coding Tools Transform Development Workflows

Theme 3: Fine-Tuning and Training Models Remain Complex

Theme 4: Creative Applications of AI Gain Traction

Theme 5: Security and Ethical Concerns Surrounding AI Technologies


PART 1: High level Discord summaries

aider (Paul Gauthier) Discord


Unsloth AI (Daniel Han) Discord


OpenRouter (Alex Atallah) Discord


Perplexity AI Discord


HuggingFace Discord


Nous Research AI Discord


OpenAI Discord


CUDA MODE Discord


LM Studio Discord


Latent Space Discord


Interconnects (Nathan Lambert) Discord


Cohere Discord


Stability.ai (Stable Diffusion) Discord


Modular (Mojo 🔥) Discord


Eleuther Discord


LlamaIndex Discord


OpenInterpreter Discord


OpenAccess AI Collective (axolotl) Discord


LangChain AI Discord


DSPy Discord


tinygrad (George Hotz) Discord


Torchtune Discord


LAION Discord


MLOps @Chipro Discord


Gorilla LLM (Berkeley Function Calling) Discord


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

aider (Paul Gauthier) ▷ #general (679 messages🔥🔥🔥):

  • O1 and Claude Models
  • AI and Job Impact
  • Using Aider with LLMs
  • AI Coding Tools
  • Prompt Engineering

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (132 messages🔥🔥):

  • Aider scripting
  • Clipboard usage with Aider
  • Aider configuration options
  • Model interaction and response
  • File handling commands in Aider

Links mentioned:


aider (Paul Gauthier) ▷ #links (12 messages🔥):

  • Zed on Linux
  • OpenAI Benchmark Insights
  • Game Gen - O Tool
  • Library of New Tools
  • The Big Prompt Library

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (751 messages🔥🔥🔥):

  • Gemma2 Training
  • Mistral Models
  • LLama 3.1 Performance
  • Qwen 2.5 Release
  • Using Gradient Accumulation

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (72 messages🔥🔥):

  • Job Hunting
  • PhD Discussions
  • Recruitment Process
  • LeetCode Interviews
  • Industry Shifts

Unsloth AI (Daniel Han) ▷ #help (133 messages🔥🔥):

  • Unsloth Pro Pricing
  • Multi-GPU Training
  • API Token for Hugging Face
  • Training Loss Profile
  • Model Fine-tuning Issues

Links mentioned:


Unsloth AI (Daniel Han) ▷ #research (37 messages🔥):

  • Fine-tuning voice models
  • Papers with Code reliability
  • Scaling laws in Llora rank
  • Direct Preference Optimization (DPO)
  • Testing DPO loss types

Link mentioned: DPO Trainer: no description found


OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):

fn5io: as posted on hacker news today: https://github.com/bklieger-groq/g1


OpenRouter (Alex Atallah) ▷ #general (775 messages🔥🔥🔥):

  • OpenRouter model context limits
  • User feedback on model performance
  • Prompt engineering techniques
  • Howarding provider efficiency
  • Prompt caching functionality

Links mentioned:


OpenRouter (Alex Atallah) ▷ #beta-feedback (1 messages):

  • Hyperbolic Key Integrations
  • DeepSeek Ignored Provider
  • JSON Configuration Issues

Perplexity AI ▷ #general (615 messages🔥🔥🔥):

  • Issues with Perplexity AI
  • Comparison of AI Models
  • Pricing and Promotions for AI Services
  • Model Performance and Features
  • User Experiences with Perplexity and Competitors

Links mentioned:


Perplexity AI ▷ #sharing (37 messages🔥):

  • Aerospike Engine
  • Google Search Differences
  • Korean Emotion Video Dataset
  • Microstrategy's Billion Dollar Investment
  • Minecraft Moderation Issues

Perplexity AI ▷ #pplx-api (18 messages🔥):

  • API Errors
  • Model Availability
  • Domain Specific Filters
  • Timeout Issues
  • Citation Discrepancies

Link mentioned: Supported Models - Perplexity: no description found


HuggingFace ▷ #general (571 messages🔥🔥🔥):

  • Fine-tuning LLMs
  • Hugging Face Inference API
  • GPU Resource Management
  • Quantization Techniques
  • SQL Integration with Datasets

Links mentioned:


HuggingFace ▷ #today-im-learning (7 messages):

  • Web Scraping Hugging Face Documentation
  • Training AI for Non-Technical Users
  • Challenges in Scraping
  • Community Engagement in AI Learning

HuggingFace ▷ #cool-finds (4 messages):

  • Medical AI Research
  • Inference API Documentation Improvements
  • Vchitect 2.0 Launch
  • Chai-1 Foundation Model
  • New Medical LLMs

Links mentioned:


HuggingFace ▷ #i-made-this (23 messages🔥):

  • Flux Image Generation
  • OCR Demo Popularity
  • Room Cleaner App
  • Interactive World & Character Generative AI
  • AI Multi-Agent System in Java

Links mentioned:


HuggingFace ▷ #reading-group (3 messages):

  • The Keys to the White House prediction system
  • Influence of character on presidential picks
  • Bias in predictions

Link mentioned: The Keys to the White House - Wikipedia: no description found


HuggingFace ▷ #NLP (7 messages):

  • Tokenizer Training
  • Llama.cpp Public Domain Models
  • Pretrained VLMs
  • Finetuning Nomic Embeddings
  • Open Source LLMs with PyTorch

Link mentioned: Training and Finetuning Embedding Models with Sentence Transformers v3: no description found


HuggingFace ▷ #diffusion-discussions (4 messages):

  • Tokenizer Training Techniques
  • Nitro Giveaway Announcement

Nous Research AI ▷ #general (332 messages🔥🔥):

  • StealC malware
  • GameGen-O
  • Drag-based image editing advancements
  • LLM reflections
  • Suno slop in music

Links mentioned:


Nous Research AI ▷ #ask-about-llms (118 messages🔥🔥):

  • Precision Annealing
  • Loss Monitoring in Training
  • Fine-tuning Challenges
  • Evaluation Metrics
  • Model Comparisons

Links mentioned:


Nous Research AI ▷ #research-papers (20 messages🔥):

  • Model Routing in Small Devices
  • Multi-Language Model Training
  • Characteristics of Enacted Language in LLMs
  • Scaling LLM Inference
  • Transformers vs. LLMs

Links mentioned:


Nous Research AI ▷ #interesting-links (17 messages🔥):

  • Kairos AI Religion
  • NotebookLM Features
  • Podcast Feedback
  • Overmind Prompt
  • Synthetic Dataset Generation

Links mentioned:


Nous Research AI ▷ #research-papers (20 messages🔥):

  • Model Inference Scaling
  • Enacted Language Limitations
  • Multi-Language Datasets
  • Transformers vs. LLMs
  • Performance Limits of Transformers

Links mentioned:


OpenAI ▷ #ai-discussions (82 messages🔥🔥):

  • O1 AI capabilities
  • Chat classification using OpenAI
  • Custom GPTs vs. OpenAI API
  • Recent API issues
  • Functionality changes in GPT-4o

Links mentioned:


OpenAI ▷ #gpt-4-discussions (100 messages🔥🔥):

  • O1-Preview Access
  • Fine-tuning with GPT Models
  • Custom GPT Functionality
  • Model Performance Comparisons
  • User Experience Feedback

OpenAI ▷ #prompt-engineering (119 messages🔥🔥):

  • Discord Bot for Nation RP Server
  • ChatGPT Response Consistency
  • Game with Probability Mechanics
  • Extracting Text from PDFs
  • Image and Message Sharing in Discord

OpenAI ▷ #api-discussions (119 messages🔥🔥):

  • ChatGPT and game mechanics
  • Prompt engineering for consistent responses
  • Nation RP Discord server development
  • Managing unit types and battle queries
  • Exploring models' responses and misinformation

CUDA MODE ▷ #general (29 messages🔥):

  • CUDA-MODE Hackathon
  • Metal Discussion Group
  • Upsample Bicubic2D Shader
  • Quantization and Optimizing NNs

Links mentioned:


CUDA MODE ▷ #triton (5 messages):

  • Kernel launch overhead
  • CUDA Graphs performance
  • Profiling CUDA execution
  • Reducing Triton kernel launch time

Link mentioned: High kernel launch overhead · Issue #2637 · triton-lang/triton: Hey team, I'm suffering high triton kernel launch overhead. Here's my nsys capture: The kernel executes around 80us on GPU, however, it takes 220us to launch, which causes the performance degr...


CUDA MODE ▷ #torch (3 messages):

  • Multistreaming with torch.compile
  • NVCC flags for Torch extensions

Link mentioned: CUDA_NO_HALF2_OPERATORS for CUDA 9.2: We are using these flags to use the internal PyTorch half operations instead of the one from the CUDA libraries. This dates quite a while back, so I might miss some things but If I remember it correc...


CUDA MODE ▷ #algorithms (3 messages):

  • INT8 mixed-precision training
  • torchao 0.5 release
  • speedup on consumer GPUs
  • dynamic int8 matmul function

Links mentioned:


CUDA MODE ▷ #cool-links (5 messages):

  • GameGen-O
  • GPU Optimization in ML/DL
  • GPU Superclusters
  • Larry Ellison's Appeal for GPUs

Links mentioned:


CUDA MODE ▷ #beginner (2 messages):

  • Custom CUDA Kernel Training
  • Neural Network Basics
  • Learning Resources for CUDA

CUDA MODE ▷ #pmpp-book (3 messages):

  • Programming Massively Parallel Applications Solutions

CUDA MODE ▷ #torchao (40 messages🔥):

  • Quantization Strategies
  • Triton vs CuBLAS Performance
  • Kernel Performance Benchmarks
  • FP8 and INT8 Comparison
  • Contributing to TorchAO

Links mentioned:


CUDA MODE ▷ #off-topic (59 messages🔥🔥):

  • Open-Sora project
  • DLSS inferencing
  • ESRGAN and SD for upscaling
  • Temporal consistency issues in video upscaling
  • Image/video generation compute limitations

Links mentioned:


CUDA MODE ▷ #irl-meetup (1 messages):

ssp3ll: I am in Toronto as well


CUDA MODE ▷ #triton-puzzles (2 messages):

  • Triton Puzzles
  • Gradio app issues

CUDA MODE ▷ #triton-viz (2 messages):

  • Gradio version issue

CUDA MODE ▷ #llmdotc (138 messages🔥🔥):

  • Llama 3 support
  • RMSNorm implementation
  • GQA implementation
  • CUDA mode preparations
  • cuDNN and non-cuDNN path considerations

Links mentioned:


CUDA MODE ▷ #lecture-qa (1 messages):

  • NCCL Test
  • MPI Command Issues

CUDA MODE ▷ #bitnet (22 messages🔥):

  • BitNet training state
  • Ternary model distillation
  • BitNet efficiency concerns
  • Custom hardware for quantization
  • Packing ternary weights

Links mentioned:


CUDA MODE ▷ #cudamode-irl (10 messages🔥):

  • LLM Internals Visualization
  • Hack Ideas Forum
  • Custom Kernels Document
  • Project Thumbs Up
  • Access Issues on Forum

Link mentioned: Tweet from Brandon (@brandon_xyzw): Data from an "attention head" in an LLM when you update the prompt in realtime


CUDA MODE ▷ #liger-kernel (29 messages🔥):

  • Release of Liger-Kernel v0.3.0
  • Conferences in Europe
  • Sparse Mixture of Experts Implementation
  • Triton LayerNorm Issues
  • Building from Source on Ubuntu

Links mentioned:


LM Studio ▷ #general (223 messages🔥🔥):

  • GPU Acceleration Issues
  • Model Compatibility and Performance
  • Using LM Studio on Linux
  • Long Prompt Support in Models
  • Exploring AI Tools and Resources

Links mentioned:


LM Studio ▷ #hardware-discussion (101 messages🔥🔥):

  • Strix Halo APU
  • NVIDIA RTX 4090
  • Power Supply Units Comparison
  • System RAM for Larger Models
  • Water Cooling Solutions

Links mentioned:


Latent Space ▷ #ai-general-chat (116 messages🔥🔥):

  • OpenAI's o1 models
  • Cursor AI's coding assistant
  • AI Scaling Laws
  • Anthropic evals
  • Funding rounds for AI startups

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

  • OpenAI API
  • Structured Outputs
  • ChatGPT Latest
  • Voice Mode
  • O1 Meetup

Link mentioned: Tweet from Latent.Space (@latentspacepod): 🆕 pod: From API to AGI: Structured Outputs, OpenAI API platform and O1 Q&A Our @openai weekend special! https://latent.space/p/openai-api-and-o1 - 1hr convo with @michpokrass on Structured Outputs...


Latent Space ▷ #ai-in-action-club (133 messages🔥🔥):

  • Cursor Scaling Issues
  • HTEC AI Report
  • Neovim Resources
  • Vim Challenges
  • AI Programming Content

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (41 messages🔥):

  • OpenAI o1 Benchmark Results
  • Humanity's Last Exam
  • Lobbying and AI Policy
  • Dan Hendrycks Controversy
  • AI and Compute Budgets

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (5 messages):

  • Self Reflection Papers
  • Consistency in Reasoning and Outputs

Interconnects (Nathan Lambert) ▷ #ml-drama (7 messages):

  • Matt from IT
  • Meme content
  • Planetcoo Barbarons theory
  • Reinforcement Learning observations

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (53 messages🔥):

  • OpenAI o1 models
  • Nando's transition to Microsoft AI
  • Gemini Flash 1.5 performance
  • Model reasoning mechanics
  • Token economics

Links mentioned:


Interconnects (Nathan Lambert) ▷ #memes (1 messages):

xeophon.: https://x.com/burny_tech/status/1834741998898536774?s=46 <:berk:750111476483752166>


Interconnects (Nathan Lambert) ▷ #rl (7 messages):

  • Reverse Curriculum Learning
  • LLMs applications
  • Niche applications in RL

Interconnects (Nathan Lambert) ▷ #reads (6 messages):

  • Ben Thompson's accuracy
  • Information ingestion stack
  • David Perell's influence
  • YouTube video on Ben Thompson

Link mentioned: How Ben Thompson Built a Writing Empire: What if writing a newsletter could pay your rent? Well, it can. And today, you’re going to learn how.Ben Thompson makes millions of dollars a year with his w...


Interconnects (Nathan Lambert) ▷ #posts (92 messages🔥🔥):

  • OAI's ChatGPT and API Data
  • OLMoE Naming Concerns
  • LLM Model Evolution
  • Poe Subscription Discussion
  • Initialism Rankings

Cohere ▷ #discussions (87 messages🔥🔥):

  • Fei-Fei Li's Reasoning Method
  • Command-R-Plus-08-2024 Performance Issues
  • Linguistic Model Effectiveness
  • Censorship in LLMs
  • Nick Frosst and Good Kid

Links mentioned:


Cohere ▷ #announcements (1 messages):

  • Cohere Developer Office Hours
  • Command model family refresh
  • RAG capability improvements
  • Safety Modes feature
  • Updated pricing for Command models

Links mentioned:


Cohere ▷ #questions (96 messages🔥🔥):

  • Pretrained VLM usage
  • Updates on Command-r-08-2024
  • Local model deployment queries
  • Similarity search and embeddings
  • Finetuning chat models with flexible token usage

Links mentioned:


Cohere ▷ #api-discussions (13 messages🔥):

  • Production key creation issue
  • Sagemaker client billing error
  • Banking card issues with Cohere

Cohere ▷ #projects (2 messages):

  • Job Posting Concerns
  • Relevance to Cohere

Stability.ai (Stable Diffusion) ▷ #general-chat (181 messages🔥🔥):

  • FLUX Models Usage
  • Character Design with Stable Diffusion
  • Image Editing and Inpainting
  • ControlNet and LoRA Training
  • Tech Support for Stable Diffusion

Links mentioned:


Modular (Mojo 🔥) ▷ #announcements (2 messages):

  • User verification process
  • Onboarding questions
  • Server changes discussion channel
  • Latency issues with verification bot

Modular (Mojo 🔥) ▷ #mojo (91 messages🔥🔥):

  • Mojo and Python interoperability
  • Count Leading Zeros
  • Creating Slices of String Literals
  • Zero-Copy Data Interop
  • LLVM Intrinsics at Comptime

Links mentioned:


Eleuther ▷ #general (30 messages🔥):

  • Mixed Precision Training
  • CoreWeave Valuation
  • AI's Impact on Society
  • Foundation Models in Biotech
  • LM Evaluation Harness and Azure

Links mentioned:


Eleuther ▷ #research (23 messages🔥):

  • RWKV team advancements
  • Overfitting models on small datasets
  • Sequential Monte Carlo steering
  • Autoregressiveness in long documents
  • OpenAI's reasoning system o1

Links mentioned:


Eleuther ▷ #scaling-laws (7 messages):

  • Looped computation with KV cache
  • Complex algorithms on fixed VMs
  • Scaling laws literature suggestions
  • World model accuracy with large models

Eleuther ▷ #interpretability-general (10 messages🔥):

  • Goodfire.ai fundraising
  • SAE latents terminology
  • Robustness of LLM layers
  • Credit conventions in academic papers

Links mentioned:


Eleuther ▷ #lm-thunderdome (1 messages):

  • lm-evaluation-harness repo
  • Azure OpenAI integration

Eleuther ▷ #gpt-neox-dev (6 messages):

  • Hugging Face's Pile Deduped Dataset
  • Launching Multi-Node with GPT-NeoX on Polaris
  • Job Submission Issues on Polaris

LlamaIndex ▷ #blog (5 messages):

  • LlamaParse Excel Capabilities
  • TypeScript Workflows in LlamaIndex
  • Unit Testing in LLM Applications
  • Vectara-Agentic Library
  • Code Generation Agent

LlamaIndex ▷ #general (51 messages🔥):

  • LlamaIndex and ChromaDB Integration
  • LanceDB Query Issues
  • SitemapReader HTTP Errors
  • ReactAgents Streaming Responses
  • Job Opportunities in Crypto Projects

Links mentioned:


LlamaIndex ▷ #ai-discussion (4 messages):

  • Local LLM cost optimization
  • Confidentiality with Local LLM
  • Framework comparisons

OpenInterpreter ▷ #general (11 messages🔥):

  • Point of Diminishing Returns for GPUs
  • Non-streaming Response in Command Line Mode
  • Getting Started with Open Interpreter
  • Using APIs and Requests in Programming

OpenInterpreter ▷ #O1 (17 messages🔥):

  • Livekit Setup Errors
  • ChatGPT O1 Model Release
  • Comparative Model Functionality

Link mentioned: Trying To Sneak Out Of The House GIF - Arnold Schwarzenegger Sneaking Out Camouflage - Discover & Share GIFs: Click to view the GIF


OpenInterpreter ▷ #ai-content (2 messages):

  • MoA LLM Library
  • Mixture-of-Agents for Coding

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (16 messages🔥):

  • O1 model discussion
  • Attention optimization
  • Reflection tasks and agents
  • Special tokens for pretraining

Link mentioned: FlexAttention: The Flexibility of PyTorch with the Performance of FlashAttention:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (11 messages🔥):

  • Tokenization Error Bug
  • Phi 3.5 Sentence Classifier
  • vLLM and Adapter Issues
  • Fused Attention Configuration

Links mentioned:


LangChain AI ▷ #general (22 messages🔥):

  • GenAI/RAG/CV Consultation Services
  • Impact of OpenAI on Society
  • LangGraph Cloud Pricing
  • Streaming LLM Output Issues
  • Managing Chat History in LangChain

Link mentioned: JSON parser | 🦜️🔗 LangChain: This output parser allows users to specify an arbitrary JSON schema and query LLMs for outputs that conform to that schema.


LangChain AI ▷ #share-your-work (1 messages):

  • Reading PDFs with Tables
  • Comparing Table Data
  • Sample Implementations for PDFs
  • Time Consumption in Data Processing

DSPy ▷ #show-and-tell (1 messages):

batmanosama: Nhttps://www.interconnects.ai/p/reverse-engineering-openai-o1


DSPy ▷ #general (19 messages🔥):

  • RAG Query Structure
  • DSPy LM Release
  • Visual LLM Models in DSPy
  • GitHub Contributions

Links mentioned:


tinygrad (George Hotz) ▷ #general (12 messages🔥):

  • Runtime Type Checking in Tinygrad
  • Tinygrad Test Failures on AMD
  • Tinygrad Update Issues
  • GitHub Pull Request Discussions

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (4 messages):

  • Tinygrad Ecosystem Libraries
  • VRAM Spike Analysis
  • Tensor Modification Errors

Link mentioned: Issues · tinygrad/tinygrad: You like pytorch? You like micrograd? You love tinygrad! ❤️ - Issues · tinygrad/tinygrad


Torchtune ▷ #general (8 messages🔥):

  • full_finetune_distributed recipe
  • Checkpoints management
  • Learning rate scheduling
  • CUDA vs CPU operations

Link mentioned: torchtune/recipes/lora_finetune_distributed.py at 4fbe7b2d4956b3790c51d7a255c0040cf5c38fad · pytorch/torchtune: A Native-PyTorch Library for LLM Fine-tuning. Contribute to pytorch/torchtune development by creating an account on GitHub.


Torchtune ▷ #dev (6 messages):

  • Online Packing Support
  • Merge Conflicts Resolved
  • CI GPU Test Failures
  • Cache Position Updates
  • urllib Version Incompatibility

LAION ▷ #general (5 messages):

  • Generative AI in Art
  • Diffusion Illusions
  • Internship Opportunities
  • Optical Illusions
  • Image Dataset Generation

Links mentioned:


LAION ▷ #resources (1 messages):

blanchon.jl: That super great !


MLOps @Chipro ▷ #general-ml (4 messages):

  • Pretrained VLMs
  • Anomaly Detection Techniques

Gorilla LLM (Berkeley Function Calling) ▷ #leaderboard (1 messages):

  • Model function calling bug

Gorilla LLM (Berkeley Function Calling) ▷ #discussion (2 messages):

  • Function Call Errors
  • AST Decoding Issues





{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}