Frozen AI News archive

Grok-1 in Bio

**Grok-1**, a **314B parameter Mixture-of-Experts (MoE) model** from **xAI**, has been released under an Apache 2.0 license, sparking discussions on its architecture, finetuning challenges, and performance compared to models like **Mixtral** and **Miqu 70B**. Despite its size, its **MMLU benchmark performance** is currently unimpressive, with expectations that **Grok-2** will be more competitive. The model's weights and code are publicly available, encouraging community experimentation. **Sam Altman** highlighted the growing importance of compute resources, while **Grok's** potential deployment on **Groq hardware** was noted as a possible game-changer. Meanwhile, **Anthropic's Claude** continues to attract attention for its "spiritual" interaction experience and consistent ethical framework. The release also inspired memes and humor within the AI community.

Canonical issue URL


After Elon promised to release it last week, Grok-1 is now open, with a characteristically platform native announcement:

image.png

If you don't get the "in bio" thing, just ignore it, it's a silly in-joke/doesn't matter.

the GH repo offers a few more details:

image.png

Unsloth's Daniel Han went thru the architecture and called out a few notable differences, but nothing groundbreaking it seems.

Grok-1 is great that it appears to be a brand new, from-scratch open LLM that people can use, but its size makes it difficult to finetune, which Arthur Mensch of Mistral is slyly poking at:

image.png

However folks like Perplexity have already pledged to finetune it and undoubtedly the capabilities of Grok-1 will be mapped out now that it is in the wild. Ultimately the MMLU performance doesn't seem impressive, and (since we have no details on the dataset) the speculation is that it is an upcycled Grok-0, undertrained for its size and Grok-2 will be more interesting.


Table of Contents

[TOC]


PART X: AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs

Model Releases

Model Performance & Benchmarking

Compute & Hardware

Anthropic Claude

Memes & Humor

In summary, the release of Grok-1, a 314B parameter MoE model from xAI, generated significant discussion around model performance, compute requirements, and comparisons to other open-source models like Mixtral and Miqu. The spiritual experience of interacting with Anthropic's Claude also captured attention, with users noting its self-consistent histories and strong moral framework. Memes and humor around Grok's capabilities and potential misuse added levity to the technical discussions.


PART 0: Summary of Summaries of Summaries

Since Claude 3 Haiku was released recently, we're adding them to this summary run for you to compare. We'll keep running these side by side for a little longer while we build the AINews platform for a better UX.

Claude 3 Haiku (3B?)

More instability in Haiku today. It just started spitting back the previous day's prompt, and it needed a couple turns of prompting to follow instructions right.

Advancements in 3D Content Generation: Stability.ai announced the release of Stable Video 3D, a new model that can generate high-quality novel view synthesis and 3D meshes from single images, building upon their previous Stable Video Diffusion technology. This represents a significant advancement in 3D content generation capabilities. Source

Debates on Model Efficiency and Optimization: Across multiple Discords, there were ongoing discussions about the relative efficiency and performance of different Stable Diffusion models, Mistral variants, and large language models like Grok-1. Users explored techniques like quantization, sparse attention, and model scaling to improve efficiency. Sources, Sources

Concerns over Proprietary AI Models and Partnerships: The Stability.ai Discord saw users expressing concerns about the company's recent partnerships with blockchain entities, fearing a potential shift towards more proprietary models and restricted access to AI tools. Source

Navigating API and Deployment Challenges: Across multiple Discords, users discussed issues with API integration, model deployment, and service disruptions. This included problems with OpenAI's API, inconsistencies in Perplexity's API responses, and challenges in running large models like Grok-1 locally. Sources, Sources

Finetuning and Dataset Curation: The Unsloth AI, Nous Research AI, and DiscoResearch Discords saw extensive discussions around the challenges and best practices for finetuning large language models, including dataset preparation, hyperparameter tuning, and model compatibility. Users shared resources and experiences to help navigate these complex tasks. Sources, Sources, Sources

Advancements in Retrieval-Augmented Generation (RAG): The LlamaIndex Discord saw active discussions around enhancing RAG pipelines, including techniques for handling complex queries, improving structured output, and leveraging memory management. Members shared resources and project updates related to these developments. Sources

Exploring Open-Source AI Models and Tooling: Discords like Eleuther, CUDA MODE, and Interconnects highlighted the release and exploration of large open-source models like Grok-1, as well as the development of new tools and techniques for CUDA programming and AI model evaluation. Sources, Sources, Sources

Claude 3 Sonnet (14B?)

Sonnet was better at following instructions today but continued not working out of the box like we had running a few days ago. Needed to implement a check step like "you should summarize shared themes across discords and make sure to directly link to source urls where relevant."

1. Massive AI Models and Their Implications:

2. Advancements in Multimodal and Retrieval-Augmented AI:

3. Fine-tuning and Optimizing Large Language Models:

4. Prompt Engineering and Enhancing LLM Capabilities:

Claude 3 Opus (>220B?)

By far the best off the shelf summarizer model. Incredible prompt adherence. We like the Opus.

Some other noteworthy discussions include:

ChatGPT (GPT4T)

ChatGPT proved particularly stubborn today - no amount of prompting tricks were able to improve the quality of the link sourcing in today's output. We will cut over to the new pipeline this week which should solve this problem but it is disappointing that prompts alone don't do what we want here.

These themes encapsulate the dynamic nature of AI development and application, from enhancing content creation and improving model efficiency to tackling hardware limitations and exploring AI's societal implications.


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord

Revolutionizing 3D Content Generation: Stable Video 3D has been announced by Stability AI, leveraging capabilities for high-quality novel view synthesis from single images and creating 3D meshes, surpassing former models like Stable Zero123 in quality and multi-view outputs.

Anticipation High for SD3: Engineers are eagerly awaiting Stable Diffusion 3 (SD3), with beta access speculated to start rolling out soon, and the official release expected next month, promising new advancements.

Efficiency on Trial: Ongoing debates focus on the efficiency of various Stable Diffusion models, where some engineers find Stable Cascade slower but more adept at handling complex prompts compared to SDXL.

Blockchain Ventures Spark Concern: Stability AI's partnerships with blockchain entities have stirred discussions, with some AI engineers worrying about the potential move towards proprietary models and restricted access to AI tools.

Safety in File Handling: Amidst security discussions, an inquiry about converting .pt files to SAFETENSOR format led to the share of a converter tool link, while most UIs are confirmed to avoid executing unsafe code - GitHub converter tool.


Perplexity AI Discord


Unsloth AI (Daniel Han) Discord


LM Studio Discord


Nous Research AI Discord


Eleuther Discord


OpenAI Discord


HuggingFace Discord

Discord's AI Scholars Share Latest Insights:


LlamaIndex Discord


Latent Space Discord


LAION Discord


OpenAccess AI Collective (axolotl) Discord

Relevant links found in the discussions:


CUDA MODE Discord


OpenRouter (Alex Atallah) Discord


LangChain AI Discord


Interconnects (Nathan Lambert) Discord


Alignment Lab AI Discord


LLM Perf Enthusiasts AI Discord


DiscoResearch Discord


Datasette - LLM (@SimonW) Discord


Skunkworks AI Discord


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ▷ #announcements (1 messages):

Link mentioned: Introducing Stable Video 3D: Quality Novel View Synthesis and 3D Generation from Single Images — Stability AI: When we released Stable Video Diffusion, we highlighted the versatility of our video model across various applications. Building upon this foundation, we are excited to release Stable Video 3D. This n...


Stability.ai (Stable Diffusion) ▷ #general-chat (988 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):


Perplexity AI ▷ #general (795 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (35 messages🔥):


Perplexity AI ▷ #pplx-api (64 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (853 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #announcements (1 messages):

Link mentioned: GitHub - unslothai/unsloth: 2-5X faster 70% less memory QLoRA & LoRA finetuning: 2-5X faster 70% less memory QLoRA & LoRA finetuning - unslothai/unsloth


Unsloth AI (Daniel Han) ▷ #random (25 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (568 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (21 messages🔥):

Links mentioned:


LM Studio ▷ #💬-general (301 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (138 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (12 messages🔥):

Link mentioned: andrewcanis/c4ai-command-r-v01-GGUF · Hugging Face: no description found


LM Studio ▷ #🎛-hardware-discussion (480 messages🔥🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (4 messages):

Link mentioned: GitHub - lmstudio-ai/configs: LM Studio JSON configuration file format and a collection of example config files.: LM Studio JSON configuration file format and a collection of example config files. - lmstudio-ai/configs


LM Studio ▷ #langchain (1 messages):


LM Studio ▷ #avx-beta (5 messages):


LM Studio ▷ #amd-rocm-tech-preview (5 messages):

Link mentioned: GitHub - brknsoul/ROCmLibs: Prebuild Windows ROCM Libs for gfx1031 and gfx1032: Prebuild Windows ROCM Libs for gfx1031 and gfx1032 - brknsoul/ROCmLibs


LM Studio ▷ #crew-ai (1 messages):


Nous Research AI ▷ #off-topic (56 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (16 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (656 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (25 messages🔥):

Links mentioned:


Nous Research AI ▷ #bittensor-finetune-subnet (18 messages🔥):


Nous Research AI ▷ #rag-dataset (100 messages🔥🔥):

Link mentioned: scratchTHOUGHTS/commanDUH.py at main · EveryOneIsGross/scratchTHOUGHTS: 2nd brain scratchmemory to avoid overrun errors with self. - EveryOneIsGross/scratchTHOUGHTS


Eleuther ▷ #general (273 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (245 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (11 messages🔥):


Eleuther ▷ #interpretability-general (13 messages🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (31 messages🔥):

Links mentioned:


Eleuther ▷ #gpt-neox-dev (3 messages):


OpenAI ▷ #ai-discussions (193 messages🔥🔥):

Link mentioned: Enterprise privacy: no description found


OpenAI ▷ #gpt-4-discussions (34 messages🔥):


OpenAI ▷ #prompt-engineering (79 messages🔥🔥):


OpenAI ▷ #api-discussions (79 messages🔥🔥):


HuggingFace ▷ #general (96 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (12 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (12 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (18 messages🔥):

Link mentioned: Introduction - Hugging Face NLP Course: no description found


LlamaIndex ▷ #blog (7 messages):

Links mentioned:


LlamaIndex ▷ #general (303 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (4 messages):

Link mentioned: RAG with LlamaParse, Qdrant and Groq | Step By Step: In this video, I will show you how to create a effective RAG with LlamaParse, Qdrant and Groq. I will explain what LlamaParse is and briefly walk you through...


Latent Space ▷ #ai-general-chat (202 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (2 messages):

Link mentioned: A ChatGPT for Music Is Here. Inside Suno, the Startup Changing Everything | Hacker News: no description found


Latent Space ▷ #llm-paper-club-west (20 messages🔥):


Latent Space ▷ #ai-in-action-club (36 messages🔥):

Link mentioned: AI In Action: Weekly Jam Sessions: 2024 Topic,Date,Facilitator,Resources,@dropdown UI/UX patterns for GenAI,1/26/2024,nuvic,<a href="https://maggieappleton.com/squish-structure">https://maggieappleton.com/squish-struct...


LAION ▷ #general (168 messages🔥🔥):

Links mentioned:


LAION ▷ #research (13 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (99 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (24 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (35 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #datasets (8 messages🔥):

Link mentioned: GitHub - NVIDIA/NeMo-Curator: Scalable toolkit for data curation: Scalable toolkit for data curation. Contribute to NVIDIA/NeMo-Curator development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #rlhf (1 messages):

duh_kola: Is it possible to use different lora adapter to do dpo on another model


CUDA MODE ▷ #general (43 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (7 messages):

Link mentioned: Google Colaboratory: no description found


CUDA MODE ▷ #cuda (68 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #suggestions (5 messages):

Links mentioned:


CUDA MODE ▷ #jobs (1 messages):

vim410: Depends. But yes.


CUDA MODE ▷ #beginner (5 messages):

Link mentioned: no title found: no description found


CUDA MODE ▷ #pmpp-book (6 messages):


CUDA MODE ▷ #ring-attention (14 messages🔥):

Links mentioned:


CUDA MODE ▷ #off-topic (5 messages):

Link mentioned: MLSys 2024: no description found


CUDA MODE ▷ #gtc-meetup (9 messages🔥):

Link mentioned: I Snuck Into A Secret Arms-Dealer Conference: Get an exclusive video every month at https://www.patreon.com/Boy_BoyWe made this in collaboration with the legendary Australian political satire group The C...


OpenRouter (Alex Atallah) ▷ #general (159 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #general (95 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #langserve (45 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (11 messages🔥):

Links mentioned:


LangChain AI ▷ #tutorials (2 messages):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #other-papers (8 messages🔥):

Link mentioned: Logits of API-Protected LLMs Leak Proprietary Information: The commercialization of large language models (LLMs) has led to the common practice of high-level API-only access to proprietary models. In this work, we show that even with a conservative assumption...


Interconnects (Nathan Lambert) ▷ #ml-drama (19 messages🔥):

Link mentioned: Tweet from Stella Biderman (@BlancheMinerva): @natolambert @felix_red_panda You're wrong though :P


Interconnects (Nathan Lambert) ▷ #random (63 messages🔥🔥):

Links mentioned:


Alignment Lab AI ▷ #general-chat (6 messages):


Alignment Lab AI ▷ #oo (32 messages🔥):

Link mentioned: keirp/hungarian_national_hs_finals_exam · Datasets at Hugging Face: no description found


LLM Perf Enthusiasts AI ▷ #general (1 messages):


LLM Perf Enthusiasts AI ▷ #claude (7 messages):

Link mentioned: Tweet from roon (@tszzl): anthropic is controlled opposition to put the fear of god in the members of technical staff


LLM Perf Enthusiasts AI ▷ #reliability (16 messages🔥):

Links mentioned:


LLM Perf Enthusiasts AI ▷ #openai (1 messages):

res6969: https://x.com/leopoldasch/status/1768868127138549841?s=46


DiscoResearch ▷ #general (21 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (4 messages):


Datasette - LLM (@SimonW) ▷ #ai (20 messages🔥):

Links mentioned:


Datasette - LLM (@SimonW) ▷ #llm (1 messages):

obra: Is it possible to recover the seed used by the openai models for a previous api request?


Skunkworks AI ▷ #general (17 messages🔥):


Skunkworks AI ▷ #off-topic (1 messages):

pradeep1148: https://www.youtube.com/watch?v=ZlJbaYQ2hm4