Frozen AI News archive

FlashAttention 3, PaliGemma, OpenAI's 5 Levels to Superintelligence

**FlashAttention-3** introduces fast and accurate attention optimized for **H100 GPUs**, advancing native **FP8 training**. **PaliGemma**, a versatile **3B Vision-Language Model (VLM)** combining a SigLIP-So400m ViT encoder with the **Gemma-2B** language model, emphasizes a prefix-LM architecture for improved image-query interaction. **OpenAI** reveals a framework on levels of superintelligence, signaling progress toward Level 2 and highlighting internal safety disagreements. On Reddit, **NuminaMath 7B**, fine-tuned from **DeepSeekMath-7B**, wins the AI Math Olympiad by solving 29 problems using iterative supervised fine-tuning and tool-integrated reasoning. Open-source LLMs like **CodeLlama-34b** and **WizardCoder-Python-34B-V1.0** are closing the coding performance gap with closed models such as **ChatGPT-3.5**.

Canonical issue URL

AI News for 7/10/2024-7/11/2024. We checked 7 subreddits, 384 Twitters and 29 Discords (463 channels, and 2240 messages) for you. Estimated reading time saved (at 200wpm): 280 minutes. You can now tag @smol_ai for AINews discussions!

Three picks for today:

FlashAttention-3: Fast and Accurate Attention with Asynchrony and Low-precision:

While FlashAttention2 was an immediate hit last year, it was only optimized for A100 GPUs. The H100 update is here:

image.png

There's lots of fancy algorithm work that is above our paygrades, but it is notable how they are preparing the industry to move toward native FP8 training:

image.png

PaliGemma: A versatile 3B VLM for transfer:

Announced at I/O, PaliGemma is a 3B open Vision-Language Model (VLM) that is based on a shape optimized SigLIP-So400m ViT encoder and the Gemma-2B language model, and the paper is out now. Lucas tried his best to make it an informative paper.

image.png

They are really stressing the Prefix-LM nature of it: "Full attention between image and prefix (=user input), auto-regressive only on suffix (=model output). The intuition is that this way, the image tokens can see the query and do task-dependent "thinking"; if it was full AR, they couldn't."

**OpenAI Levels of Superintelligence:

We typically ignore AGI debates but when OpenAI has a framework they are communicating at all-hands, it's relevant. Bloomberg got the leak:

image.png

It's notable that OpenAI thinks it is close to solving Level 2, and that Ilya left because he also thinks Superintelligence is within reach, but disagrees on the safety element.


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

Our Twitter recap is temporarily down due to scaling issues from Smol talk.


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity.

NEW: We are experimenting with new ways to combat hallucination in our summaries and improve our comment summarization. this is our work-in-progress done this week - the final output will be a lot shorter though - let us know what you think you value in a Reddit summary.

1. Advancements in Open Source AI Models

NuminaMath 7B TIR released - the first prize of the AI Math Olympiad (Score: 10, Comments: 0):

NuminaMath 7B won first place in the AI Mathematical Olympiad, solving 29 problems compared to less than 23 by other solutions. The model is a fine-tuned version of DeepSeekMath-7B. Key points:

The model uses self-consistency decoding with tool-integrated reasoning to solve problems:

  1. Generates a CoT explanation
  2. Translates to Python code and executes in a REPL
  3. Self-heals and repeats if necessary

The competition featured complex mathematical problems, demonstrating the model's advanced capabilities in problem-solving.

Open LLMs catching up to closed LLMs [coding/ELO] (Updated 10 July 2024) (Score: 56, Comments: 4):

Open-source Large Language Models (LLMs) are rapidly improving their coding capabilities, narrowing the gap with closed-source models. Key points:

The rapid advancement of open LLMs in coding capabilities has implications for developers, researchers, and the AI industry as a whole, potentially shifting the landscape of AI-assisted programming tools.

The comments discuss various aspects of the open-source LLMs' coding capabilities:

  1. The original poster provided the source for the information, which comes from a Twitter post by Maxime Labonne. The data is based on the BigCode Bench leaderboard on Hugging Face.

  2. One commenter strongly disagrees with the rankings, particularly regarding GPT4o's coding abilities. They claim that based on their extensive daily use, Sonnet 3.5 significantly outperforms other models in coding tasks.

  3. Another user expresses amazement at the rapid progress of open-source LLMs:

    • They recall when ChatGPT was considered unbeatable, with only inferior alternatives available.
    • Now, there are models surpassing ChatGPT's performance.
    • The commenter is particularly impressed that such powerful models can run locally on a PC, describing it as having "the knowledge of the whole world in a few GB of a gguf file".

I created a Llama 3 8B model that follows response format instructions perfectly: Formax-v1.0 (Score: 29, Comments: 3):

The user claims to have created a Llama 3 8B model called Formax-v1.0 that excels at following response format instructions. Key points include:

The post suggests this model could be valuable for developers working on applications that need precise, structured responses from language models.

Comments:

The post creator, nero10578, provides additional context and examples of the model's capabilities:

  1. The model was developed to address issues with response formatting in the MMLU-Pro benchmark, as highlighted in a previous post.

  2. A comparison of MMLU-Pro test results shows:

    • The new model (Formax-v1.0) significantly reduced random guesses caused by incorrect formatting.
    • It achieves near-perfect adherence to the requested answer format of "The answer is [answer]".
    • However, it shows slightly lower accuracy compared to other models, indicating a minor trade-off in knowledge and understanding.
  3. The model was trained using a custom dataset based on the dolphin dataset by cognitivecomputations.

  4. It's designed for data processing and scenarios requiring specific response formats parsable by programs.

  5. Examples of the model's capabilities include:

    • Responding in specific JSON formats for question identification tasks.
    • Creating structured stories with defined fields like "Title" and "Story".
    • Extracting information from text and presenting it in JSON format, such as identifying characters in a story.
  6. The model can handle various formatting instructions and maintain coherence in its responses, demonstrating its versatility in following complex prom

2. AI Research Partnerships and Industry Developments

Tech Giants Step Back: Microsoft and Apple Withdraw from OpenAI Amid Regulatory Pressure (Score: 25, Comments: 0): Here's a summary of the post:

Microsoft and Apple have withdrawn from their board seats at OpenAI, the leading artificial intelligence research company. This decision comes in response to increasing regulatory scrutiny and potential antitrust concerns. Key points:

This development highlights the complex dynamics between tech giants, AI research, and regulatory pressures in the evolving landscape of artificial intelligence.

OpenAI and Los Alamos National Laboratory announce bioscience research partnership (Score: 49, Comments: 0): Summary:

OpenAI and Los Alamos National Laboratory have announced a partnership to conduct bioscience research using artificial intelligence. Key points of the collaboration include:

This partnership represents a significant step in applying advanced AI technologies to complex biological problems, potentially leading to breakthroughs in life sciences and healthcare.

This is wild. Marc Andreessen just sent $50,000 in Bitcoin to an AI agent (@truth_terminal) to so it can pay humans to help it spread out into the wild (Score: 14, Comments: 0): Summary:

Marc Andreessen, a prominent tech investor, has sent $50,000 worth of Bitcoin to an AI agent called @truth_terminal. The purpose of this funding is to enable the AI agent to:

This unusual development represents a significant step in the interaction between artificial intelligence, cryptocurrency, and human collaboration. It raises questions about the potential for AI autonomy and the role of decentralized finance in supporting AI development and expansion.

3. Advancements in AI-Generated Media

Whisper Timestamped: Multilingual speech recognition w/ word-level timestamps, running locally in your browser using Transformers.js (Score: 38, Comments: 0): Here's a summary of the post:

Whisper Timestamped is a browser-based tool for multilingual speech recognition with word-level timestamps. Key features include:

The tool is based on OpenAI's Whisper model and is implemented using Rust and WebAssembly. It demonstrates the potential of running complex AI models directly in web browsers, making advanced speech recognition technology more accessible and privacy-friendly.

Tips on how to achieve this results? This is by far the best ai influencer Ive seen. Ive shown this profile to many people and no one thought It could be ai. @viva_lalina (Score: 22, Comments: 3): Summary:

This post discusses a highly convincing AI-generated Instagram influencer profile named @viva_lalina. The author claims it's the most realistic AI influencer they've encountered, noting that many people shown the profile couldn't discern it was AI-generated. The post seeks advice on how to achieve similar results, specifically inquiring about which Stable Diffusion checkpoint might be closest to producing such realistic images, suggesting either 1.5 or XL as potential options.

Comments: Summary of comments

The comments discuss various aspects of the AI-generated Instagram influencer profile:

  1. One commenter notes that many men will likely be deceived by this realistic AI-generated profile.

  2. A user suggests that the images are created using a realistic SDXL checkpoint, stating that many such checkpoints can produce similar results.

  3. The original poster responds, mentioning difficulties in achieving the same level of realism, particularly in skin texture, eyes, and lips, even when using adetailer.

  4. A more detailed analysis suggests that the images might be created using:

    • Depth maps from existing Instagram profiles
    • SDXL for image generation
    • Possibly different checkpoints for various images
    • IPAdapter face swap for consistency in facial features
  5. The commenter notes variance in skin texture and body across images, suggesting a mix of techniques.

  6. The original poster asks for clarification on how to identify the use of different checkpoints in the images.

Overall, the comments indicate that while the AI-generated profile is highly convincing, it likely involves a combination of advanced techniques and tools beyond a single Stable Diffusion checkpoint.


AI Discord Recap

A summary of Summaries of Summaries

1. AI Model Releases and Updates

2. AI Hardware and Infrastructure

3. AI Research and Techniques


PART 1: High level Discord summaries

HuggingFace Discord


Unsloth AI (Daniel Han) Discord


CUDA MODE Discord


Nous Research AI Discord


LM Studio Discord


Latent Space Discord


Perplexity AI Discord


Stability.ai (Stable Diffusion) Discord


Modular (Mojo 🔥) Discord


LangChain AI Discord


OpenRouter (Alex Atallah) Discord


OpenAI Discord


LlamaIndex Discord


LAION Discord


Eleuther Discord


OpenInterpreter Discord


OpenAccess AI Collective (axolotl) Discord


Interconnects (Nathan Lambert) Discord


Cohere Discord


LLM Finetuning (Hamel + Dan) Discord


tinygrad (George Hotz) Discord


Mozilla AI Discord


MLOps @Chipro Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI Stack Devs (Yoko Li) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Torchtune Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ▷ #announcements (1 messages):

  • qdurllm demo
  • Leveraging Knowledge Graphs for RAG
  • Intel CPUs for HF Models
  • Self-reviewing coding assistant
  • LlamaIndex for personal data

Links mentioned:


HuggingFace ▷ #general (367 messages🔥🔥):

  • GPU Obsolescence
  • Managing Large LLMs
  • Quantization Techniques
  • Job Application AI
  • Cloud Compute Costs

Links mentioned:


HuggingFace ▷ #today-im-learning (2 messages):

  • Triplet collapse in embedding models
  • Pre-training a base with softmax for transfer learning

HuggingFace ▷ #cool-finds (6 messages):

  • Eval Dataset Fights
  • Model Accuracy Check
  • Feature Importances
  • LeRobot on Twitter

Link mentioned: Tweet from undefined: no description found


HuggingFace ▷ #i-made-this (8 messages🔥):

  • LLM Based Autonomous Agents
  • Ideogram Outputs Collection
  • Next.JS Website Refactor
  • Recent ML Research Blog
  • DPO Dataset for Python Code Quality

Links mentioned:


HuggingFace ▷ #reading-group (17 messages🔥):

  • Paper Presentation Scheduling
  • Understanding LLM Understanding Summer School
  • ResNets vs Highway Networks

Link mentioned: Understanding LLM Understanding: DEDICATED TO THE MEMORY OF DANIEL C. DENNETT : 1942 – 2024 Summer School: June 3 – June 14, 2024 VIDEOS of all the 33 talks and 7 panels Speakers — Abstracts — Timetable &#...


HuggingFace ▷ #NLP (3 messages):

  • llama-3 8b model performance
  • tensorFlow model for detecting homophobic messages
  • RAG for limited data classification
  • fine-tuning LLMs for harmful message detection

HuggingFace ▷ #diffusion-discussions (2 messages):

  • rm -rf command in Unix-based systems

Unsloth AI (Daniel Han) ▷ #general (310 messages🔥🔥):

  • Ghost 8B Beta experience
  • Qwen2 1.5b Model Discussion
  • Hardware for Fine-tuning
  • Finetuning Tips and Strategies
  • Phi-3 Models Fine-tuning Concerns

Links mentioned:


Unsloth AI (Daniel Han) ▷ #off-topic (10 messages🔥):

  • Sam Altmann Investment
  • Open Diloco
  • Distributed Training

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (7 messages):

  • Continued Pretraining without using Lora
  • Unsloth and multiple GPUs
  • Decoder Architecture for Embedding Model
  • Xformers compatibility issue with Unsloth

Unsloth AI (Daniel Han) ▷ #showcase (2 messages):

  • Ghost 8B Beta
  • Context length capabilities

Link mentioned: Ghost 8B Beta (β, 128k) - a Hugging Face Space by lamhieu: no description found


Unsloth AI (Daniel Han) ▷ #community-collaboration (10 messages🔥):

  • New message types
  • Modular Model Spec
  • Training directly on new tokens
  • Partially Trainable Config in PyTorch
  • Finetuning Gemma-2-27b for coding

Link mentioned: Modular Model Spec: no description found


Unsloth AI (Daniel Han) ▷ #research (4 messages):

  • Model Compression in LLMs
  • Norm Tweaking for Quantization
  • FlashAttention-3 Performance Boost
  • Pingpong Scheduler Implementation

Links mentioned:


CUDA MODE ▷ #general (18 messages🔥):

  • Hackathon Team Formation
  • FlashAttention discussion
  • Shared Memory Usage

Links mentioned:


CUDA MODE ▷ #triton (1 messages):

  • User-defined Triton kernels
  • torch.compile for optimization
  • Triton kernel tutorial

Link mentioned: Using User-Defined Triton Kernels with torch.compile — PyTorch Tutorials 2.3.0+cu121 documentation: no description found


CUDA MODE ▷ #torch (17 messages🔥):

  • bf16/fp16 model checkpoint issues
  • Lottery ticket hypothesis with bfloat16
  • flex_attention function
  • Optimization in test-time-training repo

Link mentioned: GitHub - test-time-training/ttt-lm-pytorch: Official PyTorch implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States: Official PyTorch implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States - test-time-training/ttt-lm-pytorch


CUDA MODE ▷ #algorithms (1 messages):

  • Adam Mini
  • GrokFast
  • MobileLLM
  • JEST

Link mentioned: AI Unplugged 14: Adam mini, GrokFast, MobileLLM, JEST: Insights over information


CUDA MODE ▷ #cool-links (2 messages):

  • AMD and Silo AI Acquisition
  • FlashAttention and GPU Performance

Links mentioned:


CUDA MODE ▷ #beginner (11 messages🔥):

  • CUDA environment setup
  • NCU segmentation fault
  • GPU driver update for WSL
  • Docker usage for CUDA

Links mentioned:


CUDA MODE ▷ #torchao (2 messages):

  • Support for Smooth Quant and AWQ
  • Implementation of to_calibrating_ Function

CUDA MODE ▷ #hqq (1 messages):

  • BitBlas backend
  • torch.compile support

Link mentioned: add bitblas backend for 4-bit/2-bit · mobiusml/hqq@6249449: no description found


CUDA MODE ▷ #llmdotc (252 messages🔥🔥):

  • Bias Handling in Models
  • Parameterized GPT2 Training
  • Custom Attention Implementations
  • AdamW Optimizer Precision
  • FlashAttention-3

Links mentioned:


CUDA MODE ▷ #sparsity (4 messages):

  • Quantization and Sparsity
  • Speed-up Techniques
  • SparseGPT
  • WANDA Pruning
  • Distillation with Sparsified Models

Links mentioned:


Nous Research AI ▷ #interesting-links (2 messages):

  • Orca 3
  • Generative Teaching
  • synthetic data for language models

Link mentioned: Tweet from arindam mitra (@Arindam1408): #Orca I'm thrilled to announce our latest work on Generative Teaching: generating vast amount of diverse high-quality synthetic data for language models to teach a specific skill (e.g. RC, text cl...


Nous Research AI ▷ #general (177 messages🔥🔥):

  • Hermes Model Performance
  • Open-Source AI
  • Dataset Availability
  • Guardrails for AI
  • Arena Learning for LLMs

Links mentioned:


Nous Research AI ▷ #ask-about-llms (7 messages):

  • Hermes 2 Theta Llama 3 70B Finetunes
  • Hermes 2 Pro
  • Storytelling Focused Finetunes

Nous Research AI ▷ #rag-dataset (13 messages🔥):

  • Anthropic Workbench
  • Prompt Engineering Job Replacement
  • Grounded vs Ungrounded Tags
  • Hermes RAG Templates
  • Synthetic Generations Export

Link mentioned: Hermes RAG Templates: Cohere-Hermes Format: [interstellarninja] System Prompt: ____________________________________________________________________________________ # Role You are an AI assistant that answers user queri...


LM Studio ▷ #💬-general (90 messages🔥🔥):

  • Feature Requests for LM Studio
  • GPU Compatibility Issues
  • Context Overflow Bug
  • Setup and Configuration Tips
  • Model and Proxy Issues

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (23 messages🔥):

  • Whisper and LM Studio integration
  • Gemma-2 Flash Attention issue
  • Handling system prompts for non-supporting models
  • Installing models using Ollama and LM Studio
  • Salesforce introduces xLAM-1B

Link mentioned: Tweet from Marc Benioff (@Benioff): Meet Salesforce Einstein “Tiny Giant.” Our 1B parameter model xLAM-1B is now the best micro model for function calling, outperforming models 7x its size, including GPT-3.5 & Claude. On-device agentic ...


LM Studio ▷ #🎛-hardware-discussion (30 messages🔥):

  • 8cx
  • Windows updates
  • Dual 4090 GPUs vs. Waiting for 5090
  • RX 580 setup
  • Arc 770 performance

Links mentioned:


LM Studio ▷ #🛠-dev-chat (5 messages):

  • Rust development
  • Etiquette of asking questions
  • The XY problem

Links mentioned:


Latent Space ▷ #ai-general-chat (36 messages🔥):

  • timestamped whisper
  • useful OpenAI API integrations
  • Blackstone's investment in AI data centers
  • PaliGemma report
  • OpenAI's revenue and progress towards AGI

Links mentioned:


Latent Space ▷ #llm-paper-club-west (93 messages🔥🔥):

  • ColBERT paper discussion
  • AI Agent survey paper
  • ImageBind modalities
  • SBERT design and training
  • Multi-agent systems in AI

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):

  • Perplexity and AWS collaboration
  • Launch of Perplexity Enterprise Pro on AWS Marketplace
  • Benefits of Amazon Bedrock for Perplexity

Link mentioned: Perplexity collaborates with Amazon Web Services to launch Enterprise Pro: We’re taking another major step in giving organizations the ability to leverage AI-powered tools for greater efficiency and productivity.


Perplexity AI ▷ #general (110 messages🔥🔥):

  • Perplexity AI features and limitations
  • Pharmacy and medication cost queries
  • Perplexity Pro and Education plans
  • Programming with Perplexity AI
  • Claude LLM model updates

Link mentioned: Bringing Perplexity to education and not-for-profits : Perplexity Enterprise Pro, with special rates for philanthropic organizations, public servants, and schools


Perplexity AI ▷ #sharing (6 messages):

  • Demographic and Pornography Use
  • Family Concepts
  • Preventing Spam Phone Calls
  • YouTube Dislike Information
  • Docker Compose Dependencies

Links mentioned:


Perplexity AI ▷ #pplx-api (3 messages):

  • Perplexity Discord integration
  • Latency issues with online models
  • Account balance check

Stability.ai (Stable Diffusion) ▷ #general-chat (116 messages🔥🔥):

  • Image Enhancements
  • Character Loras
  • Comfy-portable
  • Stable Diffusion issues
  • CivitAI banning SD3 content

Links mentioned:


Modular (Mojo 🔥) ▷ #general (2 messages):

  • mdBook advantages
  • ModularBot level advancements

Modular (Mojo 🔥) ▷ #💬︱twitter (2 messages):

  • Modular Twitter update
  • Modular status announcement

Modular (Mojo 🔥) ▷ #mojo (44 messages🔥):

  • Setitem syntax issue
  • NuMojo compatibility with nightly
  • Mojo open-source timeline
  • Kernel bypass networking in Mojo
  • Dynamic operands in mlir_ops

Links mentioned:


Modular (Mojo 🔥) ▷ #max (1 messages):

  • Channel Name Changes
  • GPU Programming Channel

Modular (Mojo 🔥) ▷ #max-gpu (1 messages):

  • MAX-related discussion
  • Dedicated GPU programming information

Modular (Mojo 🔥) ▷ #nightly (51 messages🔥):

  • New Mojo Compiler Nightly Releases
  • ArrowIntVector Conditional Conformance
  • Mojo Build Issues
  • Variant Type in Mojo

Links mentioned:


Modular (Mojo 🔥) ▷ #mojo-marathons (5 messages):

  • Mojo compiler performance
  • AVX2 and AVX-512 utilization
  • Handwritten kernels vs compiler
  • Assembly code review

LangChain AI ▷ #general (71 messages🔥🔥):

  • LangSmith Cost Calculation
  • Voice Bot Implementation
  • Vector Store Retriever Tool
  • Chroma DB Initialization
  • OpenAI Vector Store

Links mentioned:


LangChain AI ▷ #langserve (14 messages🔥):

  • Asyncio.run() RuntimeError
  • uvicorn.run() issues
  • Stream content type error
  • LangServe replacement
  • LangGraph Cloud

OpenRouter (Alex Atallah) ▷ #announcements (1 messages):

  • Magnum 72B
  • Hermes 2 Theta
  • Model Deprecations
  • Router Resilience Update

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (62 messages🔥🔥):

  • Noromaid model removal
  • LLaMA-Guard benefits
  • VoiceFlow integration with OpenRouter
  • Maintaining conversation context
  • OpenRouter and assistant API

Links mentioned:


OpenAI ▷ #ai-discussions (53 messages🔥):

  • Decentralized AI
  • BOINC
  • Sharded Computing
  • Parallel GPU Usage
  • OpenAI's New Models

Links mentioned:


OpenAI ▷ #prompt-engineering (3 messages):

  • Prompt library rename
  • Reminder about different channels

OpenAI ▷ #api-discussions (3 messages):

  • Prompt Library Rename
  • Channel Difference Reminder

LlamaIndex ▷ #blog (3 messages):

  • llama-agents launch
  • NebulaGraph integration
  • LlamaTrace collaboration

LlamaIndex ▷ #general (32 messages🔥):

  • Llamaparse and OCR
  • Setting language for prompt templates
  • Accessing additional_kwargs in CompletionResponse
  • Voice chat with GPT models
  • ReACT agent variable mapping issues

Links mentioned:


LAION ▷ #research (29 messages🔥):

  • Experimental Architectures
  • Sign Gradient
  • Residual Connections
  • Memory Efficiency in Training

Eleuther ▷ #general (11 messages🔥):

  • Diffusion Models
  • Local AI Projects
  • DoLa Decoding Strategy
  • Hugging Face Datasets
  • LLM Hallucinations

Links mentioned:


Eleuther ▷ #research (8 messages🔥):

  • Training on the test task
  • BitNet b1.58 LLM
  • Emergent behavior in models
  • Reproduction studies of LLM papers
  • Understanding of large models

Links mentioned:


OpenInterpreter ▷ #general (3 messages):

  • GPT-4o profiles
  • Llama3 local standards

OpenInterpreter ▷ #O1 (15 messages🔥):

  • LLM-Service Flag Issue
  • Profile Workaround for 01
  • Remote Experience Script for 01
  • Community Contributions in 01 Development
  • Commercial Applications of 01

OpenAccess AI Collective (axolotl) ▷ #general (17 messages🔥):

  • Axolotl dataset formats link
  • TurBcat 72B usage
  • Testing TurBcat API
  • WizardLM ArenaLearning
  • FlashAttention-3 on H100 GPUs

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (7 messages):

  • Data Curation
  • FlashAttention
  • LMSYS Chatbot Arena

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (4 messages):

  • Synthetic Instruction Data
  • RPO Preference Tuning
  • Nemotron
  • Instruction Backtranslation
  • Reward-Aware Preference Optimization

Interconnects (Nathan Lambert) ▷ #random (3 messages):

  • OpenAI's revenue breakdown
  • Subscription model of ChatGPT
  • Free usage of GPT-4

Link mentioned: Tweet from Jeremy Nixon (@JvNixon): The report on OpenAI's revenue by futureresearch is out, showing: $1.9B for ChatGPT Plus (7.7M subscribers at $20/mo), $714M from ChatGPT Enterprise (1.2M at $50/mo), $510M from the API, and $290...


Interconnects (Nathan Lambert) ▷ #rlhf (1 messages):

emily_learner: Super nice. Thanks so much. Will take look.


Cohere ▷ #general (8 messages🔥):

  • GPT Agents
  • Command R Plus
  • Fine-tuning models

LLM Finetuning (Hamel + Dan) ▷ #general (4 messages):

  • Prompt/Reply Logging Tools
  • OpenPipe for OpenAI
  • Fireworks.ai Lecture

LLM Finetuning (Hamel + Dan) ▷ #fireworks (1 messages):

  • Credits Check
  • Account ID Query

tinygrad (George Hotz) ▷ #learn-tinygrad (4 messages):

  • NVDLA vs NV accelerator
  • Runtime operations in NV
  • Unexpected UOps in simple NN graph

Link mentioned: nvdla: NVDLA Open Source Project. nvdla has 17 repositories available. Follow their code on GitHub.


Mozilla AI ▷ #llamafile (4 messages):

  • US Senate AI hearing
  • Mozilla blog on privacy law

Links mentioned:


MLOps @Chipro ▷ #events (1 messages):

  • Hugging Face Workshop
  • Business Impact of LLMs
  • Prema Roman
  • Patrick Deziel

Link mentioned: What's in an LLM? Demystifying HuggingFace models & How to Leverage Them For Business Impact | July 30, 2024: Join us on July 30 via Zoom.


MLOps @Chipro ▷ #general-ml (2 messages):

  • Recsys Community
  • Search/IR Community
  • Cohere's Sentence Transformer Team
  • Vespa
  • Elastic






{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}