Frozen AI News archive

o1: OpenAI''s new general reasoning models

**OpenAI** has released the **o1** model family, including **o1-preview** and **o1-mini**, focusing on test-time reasoning with extended output token limits over 30k tokens. The models show strong performance, ranking in the 89th percentile on competitive programming, excelling in USA Math Olympiad qualifiers, and surpassing PhD-level accuracy on physics, biology, and chemistry benchmarks. Notably, **o1-mini** performs impressively despite its smaller size compared to **gpt-4o**. The release highlights new scaling laws for test-time compute that scale loglinearly. Additionally, **Nvidia** is reportedly losing AI chip market share to startups, with a shift in developer preference from CUDA to **llama** models for web development, though Nvidia remains dominant in training. This news reflects significant advances in reasoning-focused models and shifts in AI hardware competition.

Canonical issue URL

AI News for 9/11/2024-9/12/2024. We checked 7 subreddits, 433 Twitters and 30 Discords (216 channels, and 4377 messages) for you. Estimated reading time saved (at 200wpm): 416 minutes. You can now tag @smol_ai for AINews discussions!

o1, aka Strawberry, aka Q*, is finally out! There are two models we can use today: o1-preview (the bigger one priced at $15 in / $60 out) and o1-mini (the STEM-reasoning focused distillation priced at $3 in/$12 out) - and the main o1 model is still in training. This caused a little bit of confusion.

image.png

There are a raft of relevant links, so don’t miss:

Inline with the many, many leaks leading up to today, the core story is longer “test-time inference” aka longer step by step responses - in the ChatGPT app this shows up as a new “thinking” step that you can click to expand for reasoning traces, even though, controversially, they are hidden from you (interesting conflict of interest…):

image.png

Under the hood, o1 is trained for adding new reasoning tokens - which you pay for, and OpenAI has accordingly extended the output token limit to >30k tokens (incidentally this is also why a number of API parameters from the other models like temperature and role and tool calling and streaming, but especially max_tokens is no longer supported).

image.png

The evals are exceptional. OpenAI o1:

image.png

You are used to new models showing flattering charts, but there is one of note that you don’t see in many model announcements, that is probably the most important chart of all. Dr Jim Fan gets it right: we now have scaling laws for test time compute, and it looks like they scale loglinearly.

image.png

We unfortunately may never know the drivers of the reasoning improvements, but Jason Wei shared some hints:

image.png

Usually the big model gets all the accolades, but notably many are calling out the performance of o1-mini for its size (smaller than gpt 4o), so do not miss that.

image.png


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

Nvidia and AI Chip Competition

New AI Models and Releases

AI Research and Developments

AI Tools and Applications

AI Industry and Market Trends

AI Ethics and Societal Impact

Memes and Humor


AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Reviving and Improving Classic LLM Architectures

Theme 2. New Open-Source Speech Models Pushing Boundaries

Theme 3. Benchmarking and Cost Analysis of LLM Deployments

Theme 4. Developers Embrace AI Coding Assistants, Unlike Artists with AI Art

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

AI Model Releases and Improvements

AI Research and Breakthroughs

AI Industry News

AI Ethics and Societal Impact


AI Discord Recap

A summary of Summaries of Summaries. We leave the 4o vs o1 comparisons for your benefit

GPT4O (gpt-4o-2024-05-13)

1. OpenAI O1 Model Launch and Reactions

2. AI Model Performance and Benchmarking

3. Challenges in AI Training and Inference

4. Innovations in AI Tools and Frameworks

5. AI Applications in Various Domains

GPT4O-Aug (gpt-4o-2024-08-06)

1. OpenAI O1 Model Launch and Performance

2. DeepSeek V2.5 Features and User Feedback

3. AI Regulation and Community Reactions

4. Innovations in LLM Training and Optimization

5. Challenges in AI Model Deployment and Integration

o1-mini (o1-mini-2024-09-12)

o1-preview (o1-preview-2024-09-12)

OpenAI's o1 Model Launches to Mixed Reviews

Reflection 70B Model Faces Backlash Over Bold Claims

AI Developers Embrace New Tools for Enhanced Workflows

Skepticism Surrounds New AI Models Despite Promised Improvements

California's AI Regulation Bill SB 1047 Raises Industry Concerns


PART 1: High level Discord summaries

OpenRouter (Alex Atallah) Discord


aider (Paul Gauthier) Discord


OpenAI Discord


Nous Research AI Discord


HuggingFace Discord


Modular (Mojo 🔥) Discord


Perplexity AI Discord


Interconnects (Nathan Lambert) Discord


Unsloth AI (Daniel Han) Discord


Latent Space Discord


CUDA MODE Discord


Stability.ai (Stable Diffusion) Discord


LM Studio Discord


OpenInterpreter Discord


Cohere Discord


Eleuther Discord


LlamaIndex Discord


DSPy Discord


Torchtune Discord


OpenAccess AI Collective (axolotl) Discord


LAION Discord


LangChain AI Discord


tinygrad (George Hotz) Discord


LLM Finetuning (Hamel + Dan) Discord


Gorilla LLM (Berkeley Function Calling) Discord


MLOps @Chipro Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

OpenRouter (Alex Atallah) ▷ #announcements (1 messages):

  • DeepSeek V2.5
  • DeepSeek merger
  • Reflection endpoint
  • Data privacy

Link mentioned: Tweet from OpenRouter (@OpenRouterAI): DeepSeek 2.5 now has a full-precision provider, @hyperbolic_labs! It also doesn't log prompts, for data-conscious users. You can configure your data preferences in /settings/privacy


OpenRouter (Alex Atallah) ▷ #general (778 messages🔥🔥🔥):

  • OpenAI O1 Model
  • DeepSeek Endpoint Performance
  • Model Pricing and Costs
  • User Experiences with LLMs
  • Limitations of O1 and GPT-4o

Links mentioned:


aider (Paul Gauthier) ▷ #general (658 messages🔥🔥🔥):

  • OpenAI o1 Pricing and Performance
  • Aider vs. Cursor
  • User Experiences with o1
  • System Prompt Modifications in Aider
  • Jailbreak Attempts with o1

Links mentioned:


aider (Paul Gauthier) ▷ #questions-and-tips (62 messages🔥🔥):

  • Aider scripting
  • DeepSeek integration
  • Git ignore issues
  • Environment variables in Windows
  • Cached token usage

Links mentioned:


aider (Paul Gauthier) ▷ #links (2 messages):

  • PLANSEARCH algorithm
  • LLM code generation
  • Diversity in LLM outputs

Links mentioned:


OpenAI ▷ #annnouncements (1 messages):

  • OpenAI o1 models
  • AI reasoning capabilities
  • ChatGPT updates
  • API access for developers

OpenAI ▷ #ai-discussions (416 messages🔥🔥🔥):

  • GPT-4o vs GPT o1
  • Advanced Voice Mode
  • Performance of AI Models in Math
  • Use of Chain of Thought
  • Limitations of File Upload Features

Links mentioned:


OpenAI ▷ #gpt-4-discussions (27 messages🔥):

  • ChatGPT memory issues
  • OpenAI o1 model reactions
  • Custom GPT publishing problems
  • Windows app status
  • API for customizable GPTs

OpenAI ▷ #prompt-engineering (6 messages):

  • Library Location
  • Prompt Performance

OpenAI ▷ #api-discussions (6 messages):

  • Location of Library
  • Prompt-Labs
  • Server Search Challenges
  • Physics Integration
  • Response Time

Nous Research AI ▷ #general (296 messages🔥🔥):

  • Nous Research AI Discord Chatbot Performance
  • OpenAI o1 Preview Discussion
  • MCTS and its Relation to LLMs
  • Consensus Prompting Technique
  • Evaluation of Hermes 3 and o1 Models

Links mentioned:


Nous Research AI ▷ #ask-about-llms (48 messages🔥):

  • Small LLM Recommendations
  • Using AI for Marketing
  • Mistral Model Discussion
  • Cohere Models' Performance
  • Safety Alignment in AI

Links mentioned:


Nous Research AI ▷ #research-papers (4 messages):

  • Hallucinations in Large Language Models
  • Dolphin Architecture for Energy Efficiency
  • Contextual Processing in Language Models

Links mentioned:


Nous Research AI ▷ #interesting-links (2 messages):

  • AI models reasoning
  • Sci Scope newsletter

Link mentioned: Sci Scope: An AI generated newspaper on AI research


Nous Research AI ▷ #research-papers (4 messages):

  • Hallucinations in Language Models
  • Dolphin Architecture for Long Contexts
  • Projector in Language Models
  • Efficiency in Language Processing

Links mentioned:


Nous Research AI ▷ #reasoning-tasks (2 messages):

  • Survey Paper Implementation
  • Alternatives to LLMs for Deductive Reasoning
  • General Reasoning Engines
  • Syllogisms and Logic Application

Links mentioned:


HuggingFace ▷ #announcements (1 messages):

  • CLI Tools Revolution
  • Reflection 70B Model
  • Persian Dataset
  • Arena Learning Performance
  • Fine-Tuning LLMs

HuggingFace ▷ #general (324 messages🔥🔥):

  • AI Regulation Discussions
  • ML Community Sentiment
  • Free Speech vs. Hate Speech
  • Debate Reactions
  • Fine-tuning Models with Qlora

Links mentioned:


HuggingFace ▷ #today-im-learning (1 messages):

lunarflu: awesome, let us know how it goes! <:hugging_rocket:968127385864134656>


HuggingFace ▷ #cool-finds (2 messages):

  • Exa AI search engine
  • Meta Discovery framework in gaming

Links mentioned:


HuggingFace ▷ #i-made-this (5 messages):

  • HOPE Agent
  • DebateThing.com
  • Synthetic Data Generation Packages
  • Hugging Face Community Mapping
  • Batch Size in Sentence Transformers

Links mentioned:


HuggingFace ▷ #reading-group (1 messages):

  • Team Building in ML
  • Collaborative Projects in AI

HuggingFace ▷ #NLP (9 messages🔥):

  • NSFW text detection datasets
  • Transformer architecture and embeddings
  • Mean pooling in embeddings
  • Self-supervised training techniques
  • Building models from scratch

Modular (Mojo 🔥) ▷ #general (274 messages🔥🔥):

  • Mojo Error Handling
  • String Conversion in Mojo

Links mentioned:


Modular (Mojo 🔥) ▷ #mojo (24 messages🔥):

  • ExplicitlyCopyable trait
  • MojoDojo open source community
  • Recommendation systems in Mojo
  • Chaining in Rust-like syntax
  • Accessing errno in Mojo

Modular (Mojo 🔥) ▷ #max (3 messages):

  • Embedding in MAX
  • Vector databases
  • Semantic search
  • Natural Language Processing
  • MAX Engine

Link mentioned: Modular: Semantic Search with MAX Engine: In the field of natural language processing (NLP), semantic search focuses on understanding the context and intent behind queries, going beyond mere keyword matching to provide more relevant and conte...


Perplexity AI ▷ #general (253 messages🔥🔥):

  • OpenAI O1 Preview
  • AI Music Trends
  • Suno Discord Insights
  • User Experiences with APIs
  • Uncovr.ai Development

Links mentioned:


Perplexity AI ▷ #sharing (16 messages🔥):

  • Air Raid Sirens History
  • AI Developments in Gaming
  • IDE Recommendations
  • Dark Shipping Trends
  • Supercomputer Innovations

Perplexity AI ▷ #pplx-api (3 messages):

  • Enterprise Pro License Issues
  • Free Credit Balance Concerns
  • Comparison of Perplexity Pro and Sonar Online

Interconnects (Nathan Lambert) ▷ #news (213 messages🔥🔥):

  • OpenAI o1 Model Release
  • California AI Regulation SB 1047
  • AI Model Benchmarking
  • Chain of Thought in AI
  • AI Process and Reasoning

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (6 messages):

  • Matt's API Fix
  • Recent Release Excitement
  • Benchmark Speculation

Link mentioned: Tweet from Xeophon (@TheXeophon): @terryyuezhuo lol what if matts benchmarks were done with o1


Interconnects (Nathan Lambert) ▷ #memes (2 messages):

  • xeophon
  • logan

Interconnects (Nathan Lambert) ▷ #posts (6 messages):

  • RLHF in Private Models
  • Enterprise Offerings by Scale AI
  • Model Behavior Shifting
  • Training Reliability
  • Domain Expertise Challenges

Unsloth AI (Daniel Han) ▷ #general (179 messages🔥🔥):

  • Reflection Model Controversy
  • Unsloth and GPU Compatibility
  • KTO Model Alignment
  • Solar Pro Preview Model
  • LLM Distillation Challenges

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (12 messages🔥):

  • Discord recommendations for AI
  • Beta testing Android app
  • Using LLMs with a single file

Link mentioned: GitHub - Mozilla-Ocho/llamafile: Distribute and run LLMs with a single file.: Distribute and run LLMs with a single file. Contribute to Mozilla-Ocho/llamafile development by creating an account on GitHub.


Unsloth AI (Daniel Han) ▷ #help (29 messages🔥):

  • Finetuning Gemma with Alpaca
  • Using Unsloth for Llama 3.1
  • Serving Quantized Models with vLLM
  • Dataset Formatting Issues
  • Learning Rate Schedulers for Qlora

Links mentioned:


Unsloth AI (Daniel Han) ▷ #research (5 messages):

  • Sci Scope newsletter
  • Unsloth AI performance
  • Fast Forward optimization strategy
  • OpenAI O1 preview

Links mentioned:


Latent Space ▷ #ai-general-chat (133 messages🔥🔥):

  • OpenAI o1 model launch
  • Devin AI coding agent
  • Inference time scaling
  • Reasoning capabilities
  • AI industry reactions

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

swyxio: o1 meetup today in <@&979487661490311320> https://x.com/swyx/status/1834300361102029286


CUDA MODE ▷ #general (5 messages):

  • PTQ and Static Quantization
  • Dynamic Quantization Discussion
  • CUDA MODE Notes
  • De-Quantization Error

Link mentioned: Christian Mills - CUDA MODE Lecture Notes: My notes from the CUDA MODE reading group lectures run by Andreas Kopf and Mark Saroufim.


CUDA MODE ▷ #triton (6 messages):

  • Utilizing Tensor Cores
  • Support for uint4
  • Block Pointer Usage

Links mentioned:


CUDA MODE ▷ #torch (1 messages):

  • Torch Inductor Internals
  • PyTorch Native Compiler

Link mentioned: TorchInductor: a PyTorch-native Compiler with Define-by-Run IR and Symbolic Shapes: The PyTorch team has been building TorchDynamo, which helps to solve the graph capture problem of PyTorch with dynamic Python bytecode transformation. To actually make PyTorch faster, TorchDynamo must...


CUDA MODE ▷ #cool-links (5 messages):

  • HF transformers LLMs
  • OpenAI O1 Preview
  • Learning to Reason with LLMs
  • Iterative Improvements with LLMs

Link mentioned: Tweet from mobicham (@mobicham): Simple trick to make HF transformers LLMs work properly with torch.compile + static cache without implementing a custom generator!


CUDA MODE ▷ #jobs (1 messages):

  • Aurora Innovation
  • Hiring for Engineers
  • Commercial Launch Plans
  • Recent Funding Round
  • Driverless Truck Corridor

Links mentioned:


CUDA MODE ▷ #beginner (7 messages):

  • Rust CUDA Crate
  • NCCL Allreduce Documentation
  • Cublas GEMM Function Issues

Link mentioned: GitHub - Rust-GPU/Rust-CUDA: Ecosystem of libraries and tools for writing and executing fast GPU code fully in Rust.: Ecosystem of libraries and tools for writing and executing fast GPU code fully in Rust. - Rust-GPU/Rust-CUDA


CUDA MODE ▷ #pmpp-book (1 messages):

  • 1 Radius Blur Implementation
  • GitHub Contributions

Link mentioned: pmpp/src/convolution.cu at main · jfischoff/pmpp: Exercises for Programming Massively Parrallel Processors - jfischoff/pmpp


CUDA MODE ▷ #torchao (5 messages):

  • TorchAO limitations
  • AO document availability
  • CUDA matmul operations
  • Quantization techniques

Links mentioned:


CUDA MODE ▷ #off-topic (49 messages🔥):

  • O1 and its limitations
  • Comparison of AI models
  • Hidden Chain of Thought (CoT)
  • OpenAI's decision-making
  • Research landscape in AI

Links mentioned:


CUDA MODE ▷ #hqq-mobius (1 messages):

  • torch.compile support
  • transformers model.generate
  • HFGenerator deprecation

CUDA MODE ▷ #llmdotc (2 messages):

  • Mixed Input Scale Factor
  • GEMM Performance in Hopper

CUDA MODE ▷ #cudamode-irl (15 messages🔥):

  • CUDA Hackathon Resources
  • Cooley-Tukey FFT Implementation
  • GH200 Compute Credits
  • Custom Kernels and Track Proposals

Links mentioned:


CUDA MODE ▷ #liger-kernel (4 messages):

  • GEMM FP8 Implementation
  • SplitK GEMM Documentation
  • Test Case Scaling Challenges

Link mentioned: gemm fp8 e4m3 by AndreSlavescu · Pull Request #185 · linkedin/Liger-Kernel: Summary Implemented FP8 gemm with E4M3 representation for FP8. Issue #65 Testing Done tested square matrices of varying sizes (64, 256, 512, 1024, 2048) + non-square matrices of varying sizes ...


Stability.ai (Stable Diffusion) ▷ #general-chat (93 messages🔥🔥):

  • Reflection LLM
  • AI Photo Generation Services
  • Flux Model Performance
  • Lora Training Issues
  • Dynamic Thresholding in Models

Links mentioned:

. lllyasviel has 48 repositories available. Follow their code on GitHub.GitHub - bmaltais/kohya_ss at sd3-flux.1: Contribute to bmaltais/kohya_ss development by creating an account on GitHub.GitHub - kohya-ss/sd-scripts at sd3: Contribute to kohya-ss/sd-scripts development by creating an account on GitHub.GitHub - jhc13/taggui: Tag manager and captioner for image datasets: Tag manager and captioner for image datasets. Contribute to jhc13/taggui development by creating an account on GitHub.Home | COSMIC365: no description found


LM Studio ▷ #general (57 messages🔥🔥):

  • LM Studio RAG Implementation
  • Ternary Models Performance
  • Stable Diffusion Integration
  • Community Meet-up in London
  • OpenAI O1 Access

Links mentioned:


LM Studio ▷ #hardware-discussion (35 messages🔥):

  • 4090D GPUs
  • Power requirements for GPUs
  • Mixing GPU vendors
  • Model precision settings
  • ROCm support for AMD GPUs

Links mentioned:


OpenInterpreter ▷ #general (25 messages🔥):

  • Rich Python Library
  • Open Interpreter Skills
  • Prompt Caching for Claude
  • Exploring Hardware Possibilities with OI
  • Token Usage in Open Interpreter

Link mentioned: GitHub - Textualize/rich: Rich is a Python library for rich text and beautiful formatting in the terminal.: Rich is a Python library for rich text and beautiful formatting in the terminal. - Textualize/rich


OpenInterpreter ▷ #O1 (36 messages🔥):

  • Setting up LiveKit Server
  • OpenAI O1 launch
  • iPhone app installation
  • Python certificate update
  • Windows installation challenges

Links mentioned:


OpenInterpreter ▷ #ai-content (5 messages):

  • OpenAI o1 Models
  • Access to o1
  • Cursor and o1-mini
  • Quality of outputs

Link mentioned: Tweet from OpenAI (@OpenAI): We're releasing a preview of OpenAI o1—a new series of AI models designed to spend more time thinking before they respond. These models can reason through complex tasks and solve harder problems ...


Cohere ▷ #discussions (46 messages🔥):

  • Parakeet Project
  • GPU Hardware Survey
  • Upcoming Developer Office Hours
  • Finetuning Llama2 for American Bar Exam

Links mentioned:


Cohere ▷ #questions (1 messages):

  • Best practices for guardrails in RAG applications
  • Tools for evaluating RAG performance

Cohere ▷ #api-discussions (1 messages):

  • Polymorphic objects
  • Structured JSON in Cohere
  • Cohere schema limitations

Cohere ▷ #projects (3 messages):

  • Personal AI Executive Assistant
  • Scheduling Automation
  • AI General Purpose Agent
  • Rate Limit Concerns
  • Open Source Codebase

Link mentioned: GitHub - mcruz90/ai-need-help: Contribute to mcruz90/ai-need-help development by creating an account on GitHub.


Eleuther ▷ #general (12 messages🔥):

  • Contributing to ongoing projects
  • PhD and postdoc transition
  • Error with sqlitedict in virtual environment

Eleuther ▷ #research (18 messages🔥):

  • RWKV-7
  • Pixtral 12B results
  • Transfusion PyTorch implementation
  • OpenAI's o1 model series
  • Extreme Chain of Thought (CoT)

Links mentioned:


Eleuther ▷ #scaling-laws (5 messages):

  • CoT and scaling laws
  • Impact of context length on compute
  • Multiple branches in CoT
  • Tree search dynamics
  • Log-log plot observations

Eleuther ▷ #gpt-neox-dev (9 messages🔥):

  • Multinode training challenges
  • DDP performance with slow links
  • HSDP for inter-node communication
  • Optimal global batch size
  • Token estimation methods

LlamaIndex ▷ #blog (5 messages):

  • AI Scheduler Workshop
  • Multi-document RAG System for Automotive
  • OpenAI o1 and o1-mini Models Integration
  • LlamaIndex TypeScript Release
  • LlamaIndex Hackathon

Links mentioned:


LlamaIndex ▷ #general (28 messages🔥):

  • Llama3.1 function calling
  • Advanced RAG with chat_engine
  • Streaming ReAct agent activity
  • LlamaParse usage on private server
  • Filtering by specific document in Qdrant

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):

  • LangChain
  • Llama_Index
  • HayStack
  • Retrieval Augmented Generation (RAG)
  • LLM development

Link mentioned: Do we need LangChain, LlamaIndex and Haystack, and are AI Agents dead?: LangChain, LLamaIndex and Haystack are python frameworks for LLM development, especially Retrieval Augmented Generation (RAG) applications.


DSPy ▷ #papers (2 messages):

  • Output Evaluation Techniques
  • Hallucination Reduction
  • Website Citation Concerns

DSPy ▷ #general (22 messages🔥):

  • DSPy Chatbot Customization
  • OpenAI O1 Pricing
  • Requests Per Minute for O1
  • DSPy and O1 Integration

Torchtune ▷ #general (5 messages):

  • Mixed Precision Training
  • FP16 vs BF16
  • Quantization Issues
  • Tokenizer API Standardization

Torchtune ▷ #dev (17 messages🔥):

  • FlexAttention integration
  • PackedDataset performance
  • INT8 mixed-precision training
  • FSDP integration
  • QAT vs INT8 training

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (4 messages):

  • New OpenAI model
  • Llama Index
  • Reflection 70B

OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (6 messages):

  • Duplicating Code for Chat Formats
  • Internal Data Format for Datasets
  • Chat Template Feature
  • Multimodal Support in Chat Templates
  • Llama 3.1 Performance

OpenAccess AI Collective (axolotl) ▷ #general-help (3 messages):

  • DPO Custom Formats
  • Llama 3.1 Tool Call Issues

Link mentioned: Issues · axolotl-ai-cloud/axolotl: Go ahead and axolotl questions. Contribute to axolotl-ai-cloud/axolotl development by creating an account on GitHub.


LAION ▷ #general (8 messages🔥):

  • Flux by RenderNet
  • Changes in SD team
  • API/web-only model from SD
  • Open source status of SD

Link mentioned: Tweet from rendernet (@rendernet_ai): 🚀 Flux from @bfl_ml is now on RenderNet! Create hyper-realistic, character-consistent images with just one reference image — no LORAs required. Ready to bring your characters to life? Try it with j...


LAION ▷ #research (1 messages):

helium__: https://openai.com/index/introducing-openai-o1-preview/


LAION ▷ #resources (1 messages):

  • Sci Scope Highlights
  • AI Research Newsletter
  • ArXiv Paper Summaries
  • Navigation of AI Literature

Link mentioned: Sci Scope: An AI generated newspaper on AI research


LangChain AI ▷ #general (2 messages):

  • ``

LangChain AI ▷ #share-your-work (1 messages):

  • HTML chunking package
  • Web automation tools
  • HTML parsing techniques

Link mentioned: GitHub - KLGR123/html_chunking: A Python implementation for token-aware HTML chunking that preserves structure and attributes, with optional cleaning and attribute length control.: A Python implementation for token-aware HTML chunking that preserves structure and attributes, with optional cleaning and attribute length control. - KLGR123/html_chunking


tinygrad (George Hotz) ▷ #general (2 messages):

  • George Hotz's Engagement Policy
  • Terms of Service for ML Developers

tinygrad (George Hotz) ▷ #learn-tinygrad (1 messages):

xprose7820: http://www.catb.org/~esr/faqs/smart-questions.html


LLM Finetuning (Hamel + Dan) ▷ #general (2 messages):

  • LLM Applications
  • LLM Observability
  • LLM Monitoring
  • LLM Integrations
  • Fine-tuning for Translations

Link mentioned: Literal AI - RAG LLM observability and evaluation platform: Literal AI is the RAG LLM evaluation and observability platform built for Developers and Product Owners.


Gorilla LLM (Berkeley Function Calling) ▷ #leaderboard (2 messages):

  • Test Results Accuracy
  • Prompt Splicing for Qwen2-7B-Chat

MLOps @Chipro ▷ #general-ml (1 messages):

  • Predictive Maintenance
  • Unsupervised Learning in Maintenance
  • Embedded Systems Monitoring




{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}