Frozen AI News archive

Shipping and Dipping: Inflection + Stability edition

**Inflection AI** and **Stability AI** recently shipped major updates (**Inflection AI 2.5** and **Stable Diffusion 3**) but are now experiencing significant executive departures, signaling potential consolidation in the GPU-rich startup space. **Mustafa Suleyman** has joined **Microsoft AI** as CEO, overseeing consumer AI products like Copilot, Bing, and Edge. **Microsoft Azure** is collaborating with **NVIDIA** on the Grace Blackwell 200 Superchip. **Google DeepMind** announced **TacticAI**, an AI assistant for football tactics developed with Liverpool FC, using geometric deep learning and achieving 90% expert approval in blind tests. **Anthropic** released **Claude 3 Haiku** and **Claude 3 Sonnet** on Google Cloud's Vertex AI, with **Claude 3 Opus** coming soon. Concerns about AI job displacement arise as **NVIDIA** introduces AI nurses that outperform humans at bedside manner at 90% lower cost.

Canonical issue URL

It's often said that a key sign a startup has gone bad is "shipping and dipping" - the tendency of top performers who have decided to leave, to try to ship something and leave on a high note. This has just happened to both:

Senior departures are a fact of life in chaotic startups, but these do feel more major than most. It could be the start of a consolidation/cooling wave in the "hot"/"GPU-rich" startup area, but we're not quiiiite ready to call it yet. Consider us on alert though.


Table of Contents

[TOC]


PART X: AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs

Microsoft AI

Inflection AI

Google DeepMind

Anthropic

AI Safety and Risks

AI Benchmarks and Evaluations

AI Assistants and Agents

AI Coding Assistants

AI Avatars and Video

Memes and Humor


PART 0: Summary of Summaries of Summaries

we are concluding that Claude Opus is just the best model for top level summaries so we're discontinuing the A/B/C tests (see archives for our struggles/record). We'll be exposing parallel runs for all 3 + more models (incl Gemini 1.5!!) as this problem is topologically similar to our personalization app we'll be launching.


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord


Perplexity AI Discord


Unsloth AI (Daniel Han) Discord


LM Studio Discord


Nous Research AI Discord


Eleuther Discord


OpenAI Discord


HuggingFace Discord


LlamaIndex Discord


Latent Space Discord

Yann LeCun's Visual vs Linguistic Reasoning Debate Heats Up: LeCun theorizes that visual models may have an edge over language-focused ones due to the ability to map directly to actions, a view attributed to his lack of an inner monologue as indicated in tweets and an interview.

Revelations in Resolution: The UPSCALER tool by Scene-scenario is shaking up the image enhancement market with promises of 10k resolution uplifts, as shared in a tweet, potentially prompting Magnific to reconsider its pricing structure.

Grok-1's Grand Entrance: xAI's Grok-1 with 314 billion parameters is released under Apache 2.0, spurring discussions on its potential compared to other models, as seen on Grok-1's release page and various Twitter threads.

Paper Club Dives Deep Into Large Language Models: Enthusiasts in the #llm-paper-club-west channel discuss the mechanics and efficiency of attention mechanisms in transformers, helpful for understanding the design and scalability of current LLMs.

AI In Action Club Strategizes Learning: The #ai-in-action-club showcases structured AI discussions using a shared Google spreadsheet, offering insights on contrastive embeddings and suggesting LLMs for improved vector comparisons.


LAION Discord


OpenAccess AI Collective (axolotl) Discord


CUDA MODE Discord

Photonics Chips Blaze New Paths: A YouTube video titled "New Chip Breakthrough: x1000 faster" introduces photonic chips claimed to offer a 1000x performance increase, presenting insights from Lightmatter's advancements in photonic supercomputing.

Triton Trounces CUDA Puzzles: A new set of challenging Triton Puzzles was released to help users sharpen their skills, and a visualizer for Triton debugging was launched to simplify the understanding of complex load/store functions.

CUDA Community Courts Optimizations: Discussions in the CUDA channel ran deep into warp schedulers, active warps, and memories management, indicating the collective's drive to maximize CUDA efficiency and sharing insights on project structuring for better performance.

New Strides in Machine Learning Hardware: The research group led by Prof. Mohamed Abdelfattah at Cornell University is highlighted for their work in reconfigurable computing and efficient machine learning, with an accompanying master's level course ECE 5545 (CS 5775) detailed that dives into ML optimization for hardware systems.

Ring Flash Attention Clears the Air: Extensive deliberations occurred around memory requirements for attention mechanisms like FlashAttention and RingAttention, featuring knowledge sharing and a look at Striped Attention's stride towards better workloads described in an associated paper.

MLSys 2024: Where Machine Learning Meets Systems: Enthusiasm builds up for the MLSys 2024 conference in May, focusing on the merging of machine learning with systems, inviting an interdisciplinary approach to pushing the boundaries of AI efficiency.


OpenRouter (Alex Atallah) Discord

Links mentioned:


LangChain AI Discord


Interconnects (Nathan Lambert) Discord


Alignment Lab AI Discord


LLM Perf Enthusiasts AI Discord


DiscoResearch Discord


Datasette - LLM (@SimonW) Discord


Skunkworks AI Discord

[N.B.: The shared YouTube video from the #off-topic channel did not contain enough context to assess its relevance to the technical discussions.]


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ▷ #announcements (1 messages):

Link mentioned: Introducing Stable Video 3D: Quality Novel View Synthesis and 3D Generation from Single Images — Stability AI: When we released Stable Video Diffusion, we highlighted the versatility of our video model across various applications. Building upon this foundation, we are excited to release Stable Video 3D. This n...


Stability.ai (Stable Diffusion) ▷ #general-chat (988 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):


Perplexity AI ▷ #general (795 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (35 messages🔥):


Perplexity AI ▷ #pplx-api (64 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (853 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #announcements (1 messages):

Link mentioned: GitHub - unslothai/unsloth: 2-5X faster 70% less memory QLoRA & LoRA finetuning: 2-5X faster 70% less memory QLoRA & LoRA finetuning - unslothai/unsloth


Unsloth AI (Daniel Han) ▷ #random (25 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (568 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (21 messages🔥):

Links mentioned:


LM Studio ▷ #💬-general (301 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (138 messages🔥🔥):

<ul>
  <li><strong>Command-R Model Awaits Merge for LM Studio:</strong> A member mentioned that support for the Command-R 35B model in LM Studio is imminent, pending the merge of <a href="https://github.com/ggerganov/llama.cpp/pull/6033">llama.cpp pull request #6033</a> and an update to LM Studio. Once merged, CohereAI/Command-R should work.</li>
  <li><strong>Seeking Model Recommendations for Local Use:</strong> Various members discussed searching for appropriate models to run locally based on individual system constraints. A productive source for these inquiries includes visiting <a href="https://www.reddit.com/r/LocalLLaMA/">reddit's LocalLLaMA community</a> for insights.</li>
  <li><strong>Yi-9B-200K is a New Base Model:</strong> Clarification was provided that Yi-9B-200K operates with a 200k context limit and stems from a new base model series, separate from the Llama models. An extensive resource with more information can be found on its <a href="https://huggingface.co/01-ai/Yi-9B-200K">Hugging Face model card</a>.</li>
  <li><strong>Grok-1 Release Spurs Debate:</strong> Discussion about xAI's release of the Grok-1 model, a 314B parameter Mixture-of-Experts that's not fine-tuned for any specific task, reveals skepticism regarding its immediate utility without further tuning. Details about the Grok-1 base model release can be read on the <a href="https://x.ai/blog/grok-os">xAI blog</a> and the model's raw weights are shared under the Apache 2.0 license.</li>
  <li><strong>Running LLMs with Limited Hardware:</strong> Members exchanged advice for operating language models with GPUs that have restricted capacity, such as the Nvidia 1660 Super with 6GB VRAM. One suggestion included running smaller models like Gemma 2b, often requiring operational compromises or hardware adjustments.</li>
</ul>

Links mentioned:


LM Studio ▷ #🧠-feedback (12 messages🔥):

Link mentioned: andrewcanis/c4ai-command-r-v01-GGUF · Hugging Face: no description found


LM Studio ▷ #🎛-hardware-discussion (480 messages🔥🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (4 messages):

Link mentioned: GitHub - lmstudio-ai/configs: LM Studio JSON configuration file format and a collection of example config files.: LM Studio JSON configuration file format and a collection of example config files. - lmstudio-ai/configs


LM Studio ▷ #langchain (1 messages):


LM Studio ▷ #avx-beta (5 messages):


LM Studio ▷ #amd-rocm-tech-preview (5 messages):

Link mentioned: GitHub - brknsoul/ROCmLibs: Prebuild Windows ROCM Libs for gfx1031 and gfx1032: Prebuild Windows ROCM Libs for gfx1031 and gfx1032 - brknsoul/ROCmLibs


LM Studio ▷ #crew-ai (1 messages):


Nous Research AI ▷ #off-topic (56 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (16 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (656 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (25 messages🔥):

Links mentioned:


Nous Research AI ▷ #bittensor-finetune-subnet (18 messages🔥):


Nous Research AI ▷ #rag-dataset (100 messages🔥🔥):

Link mentioned: scratchTHOUGHTS/commanDUH.py at main · EveryOneIsGross/scratchTHOUGHTS: 2nd brain scratchmemory to avoid overrun errors with self. - EveryOneIsGross/scratchTHOUGHTS


Eleuther ▷ #general (273 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (245 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (11 messages🔥):


Eleuther ▷ #interpretability-general (13 messages🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (31 messages🔥):

Links mentioned:


Eleuther ▷ #gpt-neox-dev (3 messages):


OpenAI ▷ #ai-discussions (193 messages🔥🔥):

Link mentioned: Enterprise privacy: no description found


OpenAI ▷ #gpt-4-discussions (34 messages🔥):


OpenAI ▷ #prompt-engineering (79 messages🔥🔥):


OpenAI ▷ #api-discussions (79 messages🔥🔥):


HuggingFace ▷ #general (96 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (12 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (12 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (18 messages🔥):

Link mentioned: Introduction - Hugging Face NLP Course: no description found


LlamaIndex ▷ #blog (7 messages):

Links mentioned:


LlamaIndex ▷ #general (303 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (4 messages):

Link mentioned: RAG with LlamaParse, Qdrant and Groq | Step By Step: In this video, I will show you how to create a effective RAG with LlamaParse, Qdrant and Groq. I will explain what LlamaParse is and briefly walk you through...


Latent Space ▷ #ai-general-chat (202 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (2 messages):

Link mentioned: no title found: no description found


Latent Space ▷ #llm-paper-club-west (20 messages🔥):


Latent Space ▷ #ai-in-action-club (36 messages🔥):

Link mentioned: AI In Action: Weekly Jam Sessions: 2024 Topic,Date,Facilitator,Resources,@dropdown UI/UX patterns for GenAI,1/26/2024,nuvic,<a href="https://maggieappleton.com/squish-structure">https://maggieappleton.com/squish-struct...


LAION ▷ #general (168 messages🔥🔥):

Links mentioned:


LAION ▷ #research (13 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (99 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (24 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (35 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #datasets (8 messages🔥):

Link mentioned: GitHub - NVIDIA/NeMo-Curator: Scalable toolkit for data curation: Scalable toolkit for data curation. Contribute to NVIDIA/NeMo-Curator development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #rlhf (1 messages):

duh_kola: Is it possible to use different lora adapter to do dpo on another model


CUDA MODE ▷ #general (43 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (7 messages):

Link mentioned: Google Colaboratory: no description found


CUDA MODE ▷ #cuda (68 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #suggestions (5 messages):

Links mentioned:


CUDA MODE ▷ #jobs (1 messages):

vim410: Depends. But yes.


CUDA MODE ▷ #beginner (5 messages):

Link mentioned: no title found: no description found


CUDA MODE ▷ #pmpp-book (6 messages):


CUDA MODE ▷ #ring-attention (14 messages🔥):

Links mentioned:


CUDA MODE ▷ #off-topic (5 messages):

Link mentioned: MLSys 2024: no description found


CUDA MODE ▷ #gtc-meetup (9 messages🔥):

Link mentioned: I Snuck Into A Secret Arms-Dealer Conference: Get an exclusive video every month at https://www.patreon.com/Boy_BoyWe made this in collaboration with the legendary Australian political satire group The C...


OpenRouter (Alex Atallah) ▷ #general (159 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #general (95 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #langserve (45 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (11 messages🔥):

Links mentioned:


LangChain AI ▷ #tutorials (2 messages):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #other-papers (8 messages🔥):

Link mentioned: Logits of API-Protected LLMs Leak Proprietary Information: The commercialization of large language models (LLMs) has led to the common practice of high-level API-only access to proprietary models. In this work, we show that even with a conservative assumption...


Interconnects (Nathan Lambert) ▷ #ml-drama (19 messages🔥):

Link mentioned: Tweet from Stella Biderman (@BlancheMinerva): @natolambert @felix_red_panda You're wrong though :P


Interconnects (Nathan Lambert) ▷ #random (63 messages🔥🔥):

Links mentioned:


Alignment Lab AI ▷ #general-chat (6 messages):


Alignment Lab AI ▷ #oo (32 messages🔥):

Link mentioned: keirp/hungarian_national_hs_finals_exam · Datasets at Hugging Face: no description found


LLM Perf Enthusiasts AI ▷ #general (1 messages):


LLM Perf Enthusiasts AI ▷ #claude (7 messages):

Link mentioned: Tweet from roon (@tszzl): anthropic is controlled opposition to put the fear of god in the members of technical staff


LLM Perf Enthusiasts AI ▷ #reliability (16 messages🔥):

Links mentioned:


LLM Perf Enthusiasts AI ▷ #openai (1 messages):

res6969: https://x.com/leopoldasch/status/1768868127138549841?s=46


DiscoResearch ▷ #general (21 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (4 messages):


Datasette - LLM (@SimonW) ▷ #ai (20 messages🔥):

Links mentioned:


Datasette - LLM (@SimonW) ▷ #llm (1 messages):

obra: Is it possible to recover the seed used by the openai models for a previous api request?


Skunkworks AI ▷ #general (17 messages🔥):


Skunkworks AI ▷ #off-topic (1 messages):

pradeep1148: https://www.youtube.com/watch?v=ZlJbaYQ2hm4