Frozen AI News archive

Microsoft AgentInstruct + Orca 3

**Microsoft Research** released **AgentInstruct**, the third paper in its **Orca** series, introducing a generative teaching pipeline that produces **25.8 million** synthetic instructions to fine-tune **mistral-7b**, achieving significant performance gains: +40% AGIEval, +19% MMLU, +54% GSM8K, +38% BBH, +45% AlpacaEval, and a 31.34% reduction in hallucinations. This synthetic data approach follows the success of **FineWeb** and **Apple's Rephrasing research** in improving dataset quality. Additionally, **Tencent** claims to have generated **1 billion** diverse personas for synthetic data. On AI Twitter, notable discussions included a shooting incident at a Trump rally and recent ML research highlights such as **FlashAttention-3**, **RankRAG**, and **Mixture of A Million Experts**.

Canonical issue URL

AI News for 7/12/2024-7/15/2024. We checked 7 subreddits, 384 Twitters and 29 Discords (465 channels, and 4913 messages) for you. Estimated reading time saved (at 200wpm): 505 minutes. You can now tag @smol_ai for AINews discussions!

The runaway success of FineWeb this year (our coverage here, tech report here) combined with Apple's Rephrasing research has basically served as existence proofs that there can be at least an order of magnitude improvement in dataset quality for pre- and post-training. With content shops either lawyering up or partnering up, research has turned to improving synthetic dataset generation to extend the runway on the tokens we have already compressed or scraped.

Microsoft Research has made the latest splash with AgentInstruct: Toward Generative Teaching with Agentic Flows, (not to be confused with AgentInstruct of Crispino et al 2023) the third in its Orca series of papers:

The core concept is that raw documents is transformed by multiple agents playing different roles to provide diversity (for 17 listed capabilities), which are then used by yet more agents to generate and refine instructions in a "Content Transformation Flow".

image.png

Out of this pipeline comes 22 million instructions aimed at teaching those 17 skills, which when combined with the 3.8m instructions from prior Orca papers makes "Orca 2.5" - the 25.8m instruction synthetic dataset that the authors use to finetune Mistral 7b to produce the results they report:

This is just the latest entry in this genre of synthetic data research, most recently with Tencent claiming 1 billion diverse personas on their related work.

image.png

This seems both obvious that it will work yet also terribly expensive and inefficient compared to FineWeb, but whatever works!


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

Shooting Incident at Trump Rally

AI and ML Research and Developments

Coding, APIs and Developer Tools

Humor, Memes and Off-Topic Discussions


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. We recently improved the anti-hallucination measures but are still tuning the filtering, clustiner, and summary quality.

Theme 1. AI Research Publication Lag in Fast-Paced Development

Theme 2. AI's Impact on Employment: TurboTax Layoffs

Theme 3. AI Integration in Creative Workflows: ComfyUI GLSL Node


AI Discord Recap

A summary of Summaries of Summaries

1. Pushing the Boundaries of LLMs

2. Hardware Innovations Powering AI

3. Open Source Driving AI Innovation


PART 1: High level Discord summaries

HuggingFace Discord


Unsloth AI (Daniel Han) Discord


Stability.ai (Stable Diffusion) Discord


LM Studio Discord


OpenAI Discord


Modular (Mojo 🔥) Discord


Perplexity AI Discord


Nous Research AI Discord


CUDA MODE Discord


Cohere Discord


Eleuther Discord


tinygrad (George Hotz) Discord


Latent Space Discord


OpenAccess AI Collective (axolotl) Discord


Interconnects (Nathan Lambert) Discord


LangChain AI Discord


LlamaIndex Discord


OpenInterpreter Discord


LLM Finetuning (Hamel + Dan) Discord


LAION Discord


DiscoResearch Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI Stack Devs (Yoko Li) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Torchtune Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ▷ #general (989 messages🔥🔥🔥):

  • HF Inference API
  • GPT integration
  • Model performance issues
  • Leaderboard Upvotes
  • Llama2 Chat model setup

Links mentioned:


HuggingFace ▷ #today-im-learning (3 messages):

  • Intro to PANDAS
  • Graph Machine Learning
  • K-Nearest Neighbor

Links mentioned:


HuggingFace ▷ #cool-finds (7 messages):

  • Ripple_Net library
  • FlashAttention-3 beta release
  • Model inference deployment
  • Learning calculus

Links mentioned:


HuggingFace ▷ #i-made-this (17 messages🔥):

  • NPM module supports Hugging Face Inference
  • Llama3 8B distributed on heterogeneous home cluster
  • Initial training of DPO models by user
  • Quantizing Hugging Face models on Intel GPUs
  • Continuous batching with OpenAI API

Links mentioned:


HuggingFace ▷ #reading-group (3 messages):

  • Improvement in Transformer Performance with Epochs
  • New LLM Paradigm
  • Discussion on Paper or Observation
  • Ongoing Project

HuggingFace ▷ #computer-vision (2 messages):

  • EfficientNetB7 and Swin transformer
  • OpenPose installation issues

HuggingFace ▷ #NLP (13 messages🔥):

  • LLM-Finetuning-Toolkit
  • phi-3 models on vCPU
  • RAG for multimodal image
  • Argostranslate training guide
  • Semantic search engine for emails

Link mentioned: GitHub - georgian-io/LLM-Finetuning-Toolkit: Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.: Toolkit for fine-tuning, ablating and unit-testing open-source LLMs. - georgian-io/LLM-Finetuning-Toolkit


HuggingFace ▷ #diffusion-discussions (2 messages):

  • Transformer architecture explanation
  • Training Hybrid Model on Huggingface
  • EfficientNetB7 and Swin Transformer
  • Colab for computation

Unsloth AI (Daniel Han) ▷ #general (502 messages🔥🔥🔥):

  • Llama 3 Release
  • Gemini API
  • Model Finetuning Issues
  • Training Data Formats
  • Training Checkpoints and Strategies

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (35 messages🔥):

  • MovieChat GitHub repository
  • Generating prompts with model feedback
  • Anthropic's column models
  • LLMs judging artforms
  • Issues with Firework models and troubleshooting

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (87 messages🔥🔥):

  • Instruct vs base models
  • Synthetic data generation
  • Loading gguf files with llamacpp
  • SQLDatabaseChain performance issues
  • Training and evaluation in Unsloth

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (13 messages🔥):

  • Ghost 8B Beta model
  • Training datasets
  • Dataset concerns
  • Model performance
  • Open-source data

Link mentioned: Ghost 8B Beta: A large language model was developed with goals including excellent multilingual support, superior knowledge capabilities and cost efficiency.


Unsloth AI (Daniel Han) ▷ #community-collaboration (3 messages):

  • Coding Model Metrics
  • StackOverflow Dataset

Unsloth AI (Daniel Han) ▷ #research (19 messages🔥):

  • AgentInstruct framework
  • GaLore & Q-GaLore
  • CoT style fine-tuning issues
  • CURLoRA
  • Dolomite Engine

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (403 messages🔥🔥):

  • GPTs Agents
  • OpenAI's sidebars
  • ComfyUI vs. A1111
  • AI for Custom Masks
  • AI Art Ethics and Legality

Links mentioned:


LM Studio ▷ #💬-general (120 messages🔥🔥):

  • CUDA llama.cpp error
  • GPUs for LLM
  • Multiple Instances of LM Studio
  • Context for LMs
  • Quantized Models for Performance

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (50 messages🔥):

  • Issues with WizardLM-2 on Mac
  • Best general-purpose vision model
  • Stopping Llama 3 from chat summary behavior
  • New recommendation models
  • Memory and vision model recommendations

Links mentioned:


LM Studio ▷ #📝-prompts-discussion-chat (5 messages):

  • Skynet joke prompt
  • Feedback loop for self-modifying systems
  • Parallel execution of thought trains

LM Studio ▷ #🎛-hardware-discussion (164 messages🔥🔥):

  • Hardware performance with AI models
  • Multi-GPU systems
  • Mac versus custom PCs for AI
  • ROCm and OpenCL support
  • PCIe bandwidth and implications

Links mentioned:


LM Studio ▷ #amd-rocm-tech-preview (19 messages🔥):

  • Vulkan support
  • ROCm integration
  • Hardware limitations
  • 4-bit quantization

Links mentioned:


LM Studio ▷ #🛠-dev-chat (20 messages🔥):

  • Rust vs C++
  • lmstudio.js design decisions
  • Python for neural network development
  • Embedding support in LM Studio SDK

OpenAI ▷ #ai-discussions (324 messages🔥🔥):

  • GPT and alternatives debate
  • Uses of various CoPilots
  • Online vs Offline Model Execution
  • Customization and training of AI models
  • Alternatives for affordable AI tools

Links mentioned:


OpenAI ▷ #gpt-4-discussions (45 messages🔥):

  • GPT-4o input/output types activation
  • DALL-E reliability issues with GPT
  • Hyperlink generation issues
  • Sam Altman's comments on GPT-5
  • Handling JSON responses in assistant API

OpenAI ▷ #prompt-engineering (4 messages):

  • Android Optimization Guide
  • Language Prompt Effect on AI Output

OpenAI ▷ #api-discussions (4 messages):

  • Android Optimization Guide
  • Prompt Language and Output Quality
  • Testing Language Prompts

Modular (Mojo 🔥) ▷ #general (182 messages🔥🔥):

  • Feature Requests
  • Mojo Documentation
  • Python GIL
  • Python JIT
  • Network Performance

Links mentioned:


Modular (Mojo 🔥) ▷ #ai (4 messages):

  • Conscious AI
  • Bernardo Kastrup
  • Joscha Bach
  • Split brain patients
  • Consciousness and computation

Links mentioned:


Modular (Mojo 🔥) ▷ #mojo (137 messages🔥🔥):

  • Mojo website down
  • Module ownership and deletion
  • Using keep and release in Mojo
  • Socket library implementation in Mojo
  • DateTime library in Mojo

Links mentioned:


Modular (Mojo 🔥) ▷ #max (6 messages):

  • MAX license typos
  • AMD Unified AI software stack
  • Modular's exclusive partnerships

Modular (Mojo 🔥) ▷ #max-gpu (11 messages🔥):

  • Writing custom kernels with Max
  • Lower-level API than graph
  • Benchmark Tensor Cores
  • Writing PyTorch for XLA devices

Link mentioned: GitHub - pytorch/xla: Enabling PyTorch on XLA Devices (e.g. Google TPU): Enabling PyTorch on XLA Devices (e.g. Google TPU). Contribute to pytorch/xla development by creating an account on GitHub.


Modular (Mojo 🔥) ▷ #nightly (13 messages🔥):

  • Mojo nightly releases
  • Bot interaction
  • Proposal for stdlib extensions
  • Contributor feedback

Link mentioned: [Proposal] Reduce the workload of stdlib's maintainers with stdlib-extensions · modularml/mojo · Discussion #3233: This discussion is here to have a place to talk about the folloowing proposal: pull request markdown document We are especially interested in the opinion of frequent contributors, as well as the st...


Perplexity AI ▷ #general (207 messages🔥🔥):

  • GPTs Agents
  • API Credits
  • Pro Plan Issues
  • Image Response Problems
  • Perplexity vs ChatGPT

Links mentioned:


Perplexity AI ▷ #sharing (12 messages🔥):

  • Health and Strength
  • Marketing Expertise
  • Cantillon Effect
  • Uniqueness of Teeth
  • Trump Assassination Attempt

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (8 messages🔥):

  • Cloudflare issues
  • Pro subscriber credit issues
  • API free credit problems
  • Perplexity AI API models

Nous Research AI ▷ #research-papers (5 messages):

  • AgentInstruct by Microsoft Research
  • Arena Learning by WizardLM

Links mentioned:


Nous Research AI ▷ #off-topic (11 messages🔥):

  • LivePortrait GitHub project
  • Egg cooking and peeling techniques

Link mentioned: Naked Eggs: Acid-Base Reaction - Science World: In this activity, students describe the effects of an acid on an eggshell. The reaction of the eggshell in vinegar is an acid-base reaction. When you submerge an egg in vinegar, the shell dissolves, l...


Nous Research AI ▷ #interesting-links (6 messages):

  • TextGrad
  • Q-star details
  • Claude artifacts
  • System prompts optimization tips

Links mentioned:


Nous Research AI ▷ #general (169 messages🔥🔥):

  • LLM Reasoning Improvement
  • OpenAI Platform Updates
  • AgentInstruct (Orca 3) Paper Discussion
  • New Vision Language Model by Google
  • Teknium Hiring Announcement

Links mentioned:


Nous Research AI ▷ #ask-about-llms (22 messages🔥):

  • Integrating LLM in apps
  • Extending context length for models
  • Model performance
  • UX for integrated chat
  • AI agents

Link mentioned: History for ) - bugbakery/audapolis: an editor for spoken-word audio with automatic transcription - History for ) - bugbakery/audapolis


Nous Research AI ▷ #rag-dataset (8 messages🔥):

  • Marker version speedup
  • Integration with synthetic RAG
  • XML in agent definition
  • Mixture of Agents models
  • Stasima diverse models

Links mentioned:


CUDA MODE ▷ #general (55 messages🔥🔥):

  • WebGPU Development Workflow
  • Flash Attention Memory Usage
  • ResNet Implementation

Links mentioned:


CUDA MODE ▷ #triton (5 messages):

  • Learning Triton
  • Triton Puzzles on GitHub
  • Triton in FP8 training
  • Triton's inline asm for elementwise operations

Links mentioned:


CUDA MODE ▷ #torch (3 messages):

  • Bootstrap estimate of accuracy stdev
  • Optimized dataloader issue
  • Torch nightly broken

CUDA MODE ▷ #cool-links (2 messages):

  • LoQT method for efficient training
  • Brian Kernighan on The Practice of Programming

Links mentioned:


CUDA MODE ▷ #beginner (23 messages🔥):

  • Accessing GPUs for Testing
  • Using Google Colab and nsight compute
  • CoreWeave vs Lambda Labs
  • Cloud GPU Services
  • Learning Triton

Link mentioned: Welcome to Triton’s documentation! — Triton documentation: no description found


CUDA MODE ▷ #pmpp-book (34 messages🔥):

  • CUDA Core Processing
  • Register Limitations
  • Occupancy Calculation
  • Block Size Optimization
  • Kernel Parameterization

CUDA MODE ▷ #torchao (2 messages):

  • FSDP support for low-bit optimization
  • Developer guide for integration

CUDA MODE ▷ #llmdotc (46 messages🔥):

  • Switching to cudaMallocManaged
  • llm.cpp updates
  • WebGPU insights
  • gpt3v1 by karpathy
  • GPT-3 model interpolation

Links mentioned:


CUDA MODE ▷ #youtube-watch-party (1 messages):

vkaul11: Hi


CUDA MODE ▷ #webgpu (25 messages🔥):

  • WebGPU resources and support
  • Running LLMs in the browser with Transformers.js
  • Building and troubleshooting Dawn on Windows
  • GPU buffers and performance

Links mentioned:


Cohere ▷ #general (141 messages🔥🔥):

  • OpenArena GitHub project
  • Cohere event link confusion
  • LlamaIndex KG deduplication
  • Karpathy on AI training costs
  • Account support issues

Links mentioned:


Cohere ▷ #project-sharing (26 messages🔥):

  • NPM module for Cohere
  • r/localllama bot using Langchain and Cohere
  • Using JSON from Reddit
  • Mult AI subreddit update

Links mentioned:


Eleuther ▷ #general (70 messages🔥🔥):

  • AI Meetups in London
  • OpenAI Collaborations
  • Model Benchmarking
  • Time Consideration in Models
  • Machine Learning Conferences

Links mentioned:


Eleuther ▷ #research (61 messages🔥🔥):

  • Hermes 2 Performance
  • RAG Systems with LangChain
  • Compute Thresholds Governance
  • RISE in LLMs
  • Model Compression and Accuracy

Links mentioned:


Eleuther ▷ #scaling-laws (1 messages):

wabi.sabi.1: Very interesting, thanks


Eleuther ▷ #lm-thunderdome (13 messages🔥):

  • lm-eval Python API
  • PRAUC metric for lm-eval
  • Quantization flips research
  • Distributed lm_evaluation
  • Custom functions in task YAML

Links mentioned:


Eleuther ▷ #gpt-neox-dev (1 messages):

bobby_mcbobface: Thanks Ryan! Just wanted to make sure I wasn’t going down an abandoned path


tinygrad (George Hotz) ▷ #general (104 messages🔥🔥):

  • MonoNN Compiler
  • tinygrad Kernel Overhead
  • MLX vs tinygrad Performance
  • Shape Changing Bitcasts
  • Monday Meeting Highlights

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (27 messages🔥):

  • count_include_pad in avg_pool2d
  • Tensor indexing and gather function
  • Improving Tinygrad documentation
  • Splitting Tensors based on ratios

Link mentioned: Quickstart - tinygrad docs: no description found


Latent Space ▷ #ai-general-chat (43 messages🔥):

  • Open LLM Leaderboard V2
  • Solving Reddit Link Hallucination
  • New Models in LMSys Arena
  • Cursor's Composer Feature
  • SpreadsheetLLM by Microsoft

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

swyxio: new podcast drop! https://x.com/swyx/status/1811898574416019562


Latent Space ▷ #ai-in-action-club (86 messages🔥🔥):

  • Memorable Acronyms
  • More demos and examples
  • Evaluation techniques
  • Logprob usages
  • State management

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (86 messages🔥🔥):

  • OpenArena Project
  • ORPO Training
  • Anthropic Prompt Integration
  • RAG Model Dataset
  • Weighting Conversation Data

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (5 messages):

  • Chat template dataset type
  • PR review process
  • Configuration flexibility
  • Training labels configuration
  • Handling token offsets

OpenAccess AI Collective (axolotl) ▷ #general-help (6 messages):

  • Eric's Spectrum Work
  • Quantizing Dolphin Vision 72b
  • 4-bit Model on 96GB Mac Pro

OpenAccess AI Collective (axolotl) ▷ #datasets (1 messages):

n_tt_n: i love capybara, have gotten awesome results with it


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (18 messages🔥):

  • Pushing Model to Hub after LoRA Merge
  • Vicuna Chat Template Support
  • Config Options for Vicuna Template

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (9 messages🔥):

  • GPTs Agents
  • OpenAI Platform's sidebars
  • Custom chat templates for axolotl training
  • Axolotl training setup
  • Jinja format for templates

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (31 messages🔥):

  • OpenAI working on Strawberry
  • New models in LMSYS arena
  • Stealth releases of models in LMSYS

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (23 messages🔥):

  • Mistral-7B instruct-tuning
  • Orca3/AgentInstruct paper
  • InFoBench benchmark
  • WizardArena/ArenaLearning paper
  • ChatbotArena competition

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (7 messages):

  • Finite State Machine
  • Paper Rewriting Controversy
  • Google Plagiarism

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (12 messages🔥):

  • OpenAI's revenue speculations
  • OpenAI Supply Co. shop
  • Shopify usage
  • Interconnects merch
  • Hackathons and free merch

Links mentioned:


Interconnects (Nathan Lambert) ▷ #reads (4 messages):

  • California AI Bill SB 1047
  • Paywall circumvention
  • Archive.is
  • Silicon Valley debates
  • Fortune article

Links mentioned:


LangChain AI ▷ #general (58 messages🔥🔥):

  • LangChain JS Usage
  • Gemini Pro vs API
  • RAG Errors
  • Using Base64 with APIs
  • OpenAI Embedding Models

Links mentioned:


LangChain AI ▷ #share-your-work (1 messages):

  • LLM Scraper
  • code generation
  • local models
  • GitHub project release
  • webpage scraping

Link mentioned: GitHub - mishushakov/llm-scraper: Turn any webpage into structured data using LLMs: Turn any webpage into structured data using LLMs. Contribute to mishushakov/llm-scraper development by creating an account on GitHub.


LlamaIndex ▷ #blog (10 messages🔥):

  • entity deduplication
  • LlamaCloud
  • GPT-4o for financial reports
  • multi-agent workflows with Redis
  • advanced RAG guide

Link mentioned: blogs/llm/llama_index_neo4j_custom_retriever.ipynb at master · tomasonjo/blogs: Jupyter notebooks that support my graph data science blog posts at https://bratanic-tomaz.medium.com/ - tomasonjo/blogs


LlamaIndex ▷ #general (18 messages🔥):

  • LlamaIndex KG node deduplication
  • Combining SQL and PDF embeddings
  • Handling chat history in FastAPI
  • Chunking data for better embeddings
  • KnowledgeGraphIndex with NebulaGraphStore

Links mentioned:


OpenInterpreter ▷ #general (13 messages🔥):

  • OpenInterpreter GUI Integration
  • OpenAI OS Rumors
  • Phi-3.1 Model Evaluation
  • Internlm2 Valuation
  • System Architecture Documentation Request

Links mentioned:


OpenInterpreter ▷ #O1 (3 messages):

  • Meta Ray-Ban Jailbreak
  • Installing O1 on Linux
  • 'Interpreter' Not Defined Error

LLM Finetuning (Hamel + Dan) ▷ #general (1 messages):

  • LLM agent
  • Adding agents in LLMs
  • Modular components in chat pipelines
  • Processing information using agents
  • Interacting with external APIs

Link mentioned: Adding Agents to Large Language Models Guide: Learn how to add agents in large language models using JSON output for flexible, scalable chat pipelines in this detailed guide


LLM Finetuning (Hamel + Dan) ▷ #asia-tz (2 messages):

  • OpenAI API Key request

LLM Finetuning (Hamel + Dan) ▷ #🟩-modal (1 messages):

healthymonkey: I’ve heard it’s about a year. I really like how easy it is to get H100s on modal lol


LLM Finetuning (Hamel + Dan) ▷ #hugging-face (1 messages):

  • Credit Denial

LLM Finetuning (Hamel + Dan) ▷ #axolotl (1 messages):

  • Training Loss Issues
  • Template Correctness
  • Meta's Template

Link mentioned: Meta Llama 3 | Model Cards and Prompt formats: Special Tokens used with Meta Llama 3. A prompt should contain a single system message, can contain multiple alternating user and assistant messages, and always ends with the last user message followe...


LLM Finetuning (Hamel + Dan) ▷ #wing-axolotl (2 messages):

  • modal error
  • axolotl troubleshooting
  • seeking help on slack

LLM Finetuning (Hamel + Dan) ▷ #langchain-langsmith (1 messages):

  • Langsmith evaluation
  • Rate limits in OpenAI

LLM Finetuning (Hamel + Dan) ▷ #openai (5 messages):

  • OpenAI Credit Expiration
  • Petition for Credit Extension

LAION ▷ #general (2 messages):

  • Hugging Face Profitability
  • Cambrian-1 Multimodal LLMs

Links mentioned:


LAION ▷ #learning-ml (1 messages):

  • MagViT2 compatibility with non-RGB motion data
  • Motion data preprocessing

DiscoResearch ▷ #disco_judge (2 messages):

  • LLM Arena
  • Ollama models
  • WizardLM paper
  • Arena Learning methodology

Links mentioned:








{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}