Frozen AI News archive

Google AI: Win some (Gemma, 1.5 Pro), Lose some (Image gen)

**Google's Gemma open models** (2-7B parameters) outperform **Llama 2** and **Mistral** in benchmarks but face criticism for an unusual license and poor image generation quality, which Google partially acknowledges. The upcoming **Gemini Pro 1.5** model features a 1 million token context window, excelling in video understanding and needle-in-haystack tasks. Discord communities like **TheBloke** and **LM Studio** discuss mixed reception of Gemma models, anticipation for **Llama 3** release, challenges in dataset editing, and hardware considerations such as **NVIDIA GeForce RTX 3090** and **RTX 4090** GPUs. LM Studio users report issues with version 0.2.15 Beta and ongoing integration of Gemma models, with resources shared on **Hugging Face**.

Canonical issue URL

Google is at the top of conversations for a lot of good and bad reasons today. The new Gemma open models (2-7B in size, presumably the smaller version of Gemini models) showed better benchmarks than Llama2 and Mistral:

image.png

but comes with an unusual license and doesnt pass the human vibe check.

Meanwhile, literally everybody is dogpiling on Gemini's clumsily diverse image generation, a problem partially acknowledged by Google.

image.png

But in what seems like a pure win, the long context of the still-waitlisted Gemini Pro 1.5 (with 1m token context) is video understanding and needle in haystack tests.


Table of Contents

[TOC]

PART 1: High level Discord summaries

TheBloke Discord Summary


LM Studio Discord Summary


Nous Research AI Discord Summary


Eleuther Discord Summary


HuggingFace Discord Summary


LlamaIndex Discord Summary


Mistral Discord Summary

Relevant Links:


OpenAI Discord Summary


Latent Space Discord Summary


OpenAccess AI Collective (axolotl) Discord Summary


LAION Discord Summary


Perplexity AI Discord Summary


CUDA MODE Discord Summary


LangChain AI Discord Summary


DiscoResearch Discord Summary


LLM Perf Enthusiasts AI Discord Summary


Skunkworks AI Discord Summary


Alignment Lab AI Discord Summary


Datasette - LLM (@SimonW) Discord Summary


AI Engineer Foundation Discord Summary


PART 2: Detailed by-Channel summaries and links

TheBloke ▷ #general (1156 messages🔥🔥🔥):

Links mentioned:


TheBloke ▷ #characters-roleplay-stories (189 messages🔥🔥):

Links mentioned:


TheBloke ▷ #training-and-fine-tuning (6 messages):


TheBloke ▷ #coding (166 messages🔥🔥):

Links mentioned:


LM Studio ▷ #💬-general (375 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (65 messages🔥🔥):

Links mentioned:


LM Studio ▷ #announcements (3 messages):

Links mentioned:


LM Studio ▷ #🧠-feedback (10 messages🔥):

Links mentioned:

HuggingChat - Assistants: Browse HuggingChat assistants made by the community.


LM Studio ▷ #🎛-hardware-discussion (96 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (301 messages🔥🔥):

Links mentioned:


LM Studio ▷ #autogen (1 messages):

senecalouck: https://github.com/microsoft/UFO


LM Studio ▷ #crew-ai (2 messages):


Nous Research AI ▷ #ctx-length-research (67 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #off-topic (22 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (49 messages🔥):

Links mentioned:


Nous Research AI ▷ #announcements (2 messages):

Links mentioned:


Nous Research AI ▷ #general (594 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (37 messages🔥):

Links mentioned:


Nous Research AI ▷ #collective-cognition (3 messages):


Nous Research AI ▷ #project-obsidian (3 messages):


Eleuther ▷ #general (146 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (350 messages🔥🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (38 messages🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (76 messages🔥🔥):

Links mentioned:


Eleuther ▷ #gpt-neox-dev (5 messages):

Links mentioned:

gpt-neox/megatron/model/transformer.py at f7373f806689cb270677dd48bffddf4a32bfadce · EleutherAI/gpt-neox,): An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library. - EleutherAI/gpt-neox


HuggingFace ▷ #announcements (1 messages):


HuggingFace ▷ #general (250 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (6 messages):

Links mentioned:

Agents Need Not Know Their Purpose: Ensuring artificial intelligence behaves in such a way that is aligned with human values is commonly referred to as the alignment challenge. Prior work has shown that rational agents, behaving in such...


HuggingFace ▷ #cool-finds (8 messages🔥):

Links mentioned:

Zero Prime @ Data Council '24: Join Zero Prime Ventures @ Data Council Austin 2024 for a unique chance to showcase your AI startup. Apply now for exposure to top investors and elite founders.


HuggingFace ▷ #i-made-this (33 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (4 messages):


HuggingFace ▷ #diffusion-discussions (35 messages🔥):


HuggingFace ▷ #computer-vision (4 messages):

Links mentioned:


HuggingFace ▷ #NLP (33 messages🔥):


HuggingFace ▷ #diffusion-discussions (35 messages🔥):


LlamaIndex ▷ #announcements (1 messages):

Links mentioned:


LlamaIndex ▷ #blog (4 messages):


LlamaIndex ▷ #general (379 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (7 messages):

Links mentioned:


Mistral ▷ #general (197 messages🔥🔥):

Links mentioned:


Mistral ▷ #models (72 messages🔥🔥):

Links mentioned:

FUSIONL AI: FUSIONL AI is a pioneer of SMLM Model (Smart Minimalistic Language Model) for learning in smart and minimalistic way.


Mistral ▷ #deployment (38 messages🔥):

Links mentioned:

vLLM | Mistral AI Large Language Models: vLLM can be deployed using a docker image we provide, or directly from the python package.


Mistral ▷ #finetuning (29 messages🔥):


Mistral ▷ #showcase (25 messages🔥):

Links mentioned:


Mistral ▷ #random (2 messages):

Links mentioned:


Mistral ▷ #la-plateforme (10 messages🔥):

Links mentioned:

no title found: no description found


OpenAI ▷ #ai-discussions (104 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (54 messages🔥):


OpenAI ▷ #prompt-engineering (94 messages🔥🔥):


OpenAI ▷ #api-discussions (94 messages🔥🔥):


Latent Space ▷ #ai-general-chat (92 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (3 messages):

Links mentioned:

Latent Space (Paper Club & Other Events) · Luma: View and subscribe to events from Latent Space (Paper Club & Other Events) on Luma. Latent.Space events. PLEASE CLICK THE RSS LOGO JUST ABOVE THE CALENDAR ON THE RIGHT TO ADD TO YOUR CAL. "Ad...


Latent Space ▷ #llm-paper-club-west (173 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (165 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (23 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (27 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #rlhf (1 messages):


OpenAccess AI Collective (axolotl) ▷ #runpod-help (5 messages):


LAION ▷ #general (185 messages🔥🔥):

Links mentioned:


LAION ▷ #research (30 messages🔥):

Links mentioned:


Perplexity AI ▷ #general (111 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (7 messages):


Perplexity AI ▷ #pplx-api (18 messages🔥):

Links mentioned:

Supported Models: no description found


CUDA MODE ▷ #general (3 messages):

Links mentioned:


CUDA MODE ▷ #triton (8 messages🔥):

Links mentioned:

Mamba: The Hard Way: no description found


CUDA MODE ▷ #cuda (5 messages):


CUDA MODE ▷ #torch (15 messages🔥):

Links mentioned:


CUDA MODE ▷ #beginner (14 messages🔥):

Links mentioned:


CUDA MODE ▷ #youtube-recordings (1 messages):

Links mentioned:


CUDA MODE ▷ #jax (3 messages):

Links mentioned:


CUDA MODE ▷ #ring-attention (52 messages🔥):

Links mentioned:


LangChain AI ▷ #general (46 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (3 messages):

Links mentioned:

Tweet from Michael Daigler (@michaeldaigler_): The one-man media company is closer than you think. A newsletter draft and tweets from a single topic with AI. Here's how it works:


LangChain AI ▷ #tutorials (4 messages):

Links mentioned:


DiscoResearch ▷ #general (18 messages🔥):

Links mentioned:


DiscoResearch ▷ #benchmark_dev (9 messages🔥):


LLM Perf Enthusiasts AI ▷ #general (7 messages):


LLM Perf Enthusiasts AI ▷ #opensource (1 messages):

potrock: https://blog.google/technology/developers/gemma-open-models/


LLM Perf Enthusiasts AI ▷ #embeddings (1 messages):


Skunkworks AI ▷ #general (1 messages):


Skunkworks AI ▷ #off-topic (4 messages):

Links mentioned:


Alignment Lab AI ▷ #general-chat (4 messages):


Datasette - LLM (@SimonW) ▷ #ai (2 messages):

Links mentioned:


Datasette - LLM (@SimonW) ▷ #llm (2 messages):


AI Engineer Foundation ▷ #general (3 messages):

Links mentioned:

Groq: no description found


AI Engineer Foundation ▷ #events (1 messages):

Links mentioned: