Frozen AI News archive

MetaVoice & RIP Bard

**Coqui**, a TTS startup that recently shut down, inspired a new **TTS model** supporting voice cloning and longform synthesis from a small startup called **MetaVoice**. **Google** discontinued the **Bard** brand in favor of **Gemini**. On **TheBloke Discord**, discussions focused on AI training with models like **Mixtral**, **Nous Mixtral DPO**, and **Miqu 70B**, comparing them to **OpenAI's GPT** models, and debated prompt engineering, lorebooks, and removing safety features via **LoRA fine-tuning** on models such as **Llama2 70B instruct**. Technical topics included transformer layer offloading limitations and adapting **LLaMa 2** for Apple Silicon. On **OpenAI Discord**, **DALL-E** images now include **C2PA metadata** for content authenticity, sparking debates on AI censorship, metadata manipulation, and open-source AI models versus commercial giants like **GPT-4**. Users discussed GPT-4 usability, limitations, and practical applications.

Canonical issue URL

Remember Coqui, the TTS startup that died last month? Well, a new TTS model that supports voice cloning and longform synthesis is here (try it).

image.png

It's a small startup but a promising first ship.

In other news, Google killed the Bard brand for Gemini.


Table of Contents

[TOC]

PART 1: High level Discord summaries

TheBloke Discord Summary


OpenAI Discord Summary


HuggingFace Discord Summary


LM Studio Discord Summary


Nous Research AI Discord Summary

Mistral Outshines in Programming: @carsonpoole discovered that Mistral outperforms phi2 significantly on the code section of OpenHermes 2.5 under identical sampling scenarios. The discussions included implications for GPT-4's programming capabilities and sparked curiosity surrounding the expected skillset of a 2-billion-parameter model, with cited expectations from Microsoft Research.

Sparsetral Unveiled and Math Benchmarking Excitement: The introduction of Sparsetral, a sparse MoE model, comes complete with resources such as the original paper and GitHub repos. Meanwhile, .benxh celebrated Deepseek, which incorporates a technique called DPO to achieve new proficiency levels in math-focused assessments.

Quant Tune and EQ-Bench: @tsunemoto has quantized Senku-70B, a finetuned version of the hypothetical Mistral-70B, yielding an EQ-Bench score of 84.89, and shared it on HuggingFace. This sparked a broader discourse on the significance of mathematics in appraising language models' abilities and hosting LLM-powered robotics hackathons.

Language Model Quirks and Mixtral Issues Noted: Users experienced that Mixtral, directed in Chinese, presents mixed-language responses, and similar issues with OpenHermes. Cloudflare’s AI platform adoption of these models was highlighted through tweets.

Support for Robot-Control Framework: @babycommando sought suggestions on finetuning multi-modal models and released MachinaScript for Robots with a GitHub repository. They asked for guidance on finetuning Obsidian and technical specifications for robot interactions using their LLM-driven framework.


Mistral Discord Summary


LAION Discord Summary


Eleuther Discord Summary

GPT Rivalries and Bots: GPT-3.5 showed surprising prowess in generating code for obscure languages over GPT-4, while the Eleuther server debates the trade-offs between openness and spambot disruptions.

MetaVoice TTS Model Unveiled: MetaVoice 1B, a new TTS model, was released with open source licensing, sparking discussions about its performance which includes features like zero-shot voice cloning and emotional speech synthesis as detailed in a tweet.

Evaluating Model Extrapolation and Optimization: A variety of methods for understanding and pushing model capabilities were reviewed, from analyzing loss vs sequence length to SELF-DISCOVER framework outperforming traditional methods on reasoning benchmarks as described in this paper.

Infinite Limits and Interpretability: Queries about deep learning infinite depth limits and loss landscapes sparked interest in existing research, while a new method called Evolutionary Prompt Optimization (EPO) for language model interpretation was proposed in a research paper.

Dissecting LLM Prompt Influence: The search for reliable input saliency methods in LLM prompts continued with skepticism against Integrated Gradients, underscored by a concerning paper that casts doubt on the attribution methods' ability to infer model behavior.


OpenAccess AI Collective (axolotl) Discord Summary


Perplexity AI Discord Summary


LlamaIndex Discord Summary


LangChain AI Discord Summary


Latent Space Discord Summary


CUDA MODE Discord Summary


DiscoResearch Discord Summary


LLM Perf Enthusiasts AI Discord Summary


Alignment Lab AI Discord Summary


Datasette - LLM (@SimonW) Discord Summary


PART 2: Detailed by-Channel summaries and links

TheBloke ▷ #general (1312 messages🔥🔥🔥):

Links mentioned:


TheBloke ▷ #characters-roleplay-stories (503 messages🔥🔥🔥):

Links mentioned:


TheBloke ▷ #training-and-fine-tuning (11 messages🔥):

Links mentioned:

LoRA Fine-tuning Efficiently Undoes Safety Training from Llama 2-Chat 70B — LessWrong: Produced as part of the SERI ML Alignment Theory Scholars Program - Summer 2023 Cohort, under the mentorship of Jeffrey Ladish.  …


TheBloke ▷ #coding (6 messages):

Links mentioned:

GitHub - facebookresearch/Sphere: Web-scale retrieval for knowledge-intensive NLP: Web-scale retrieval for knowledge-intensive NLP. Contribute to facebookresearch/Sphere development by creating an account on GitHub.


OpenAI ▷ #annnouncements (1 messages):


OpenAI ▷ #ai-discussions (300 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (46 messages🔥):


OpenAI ▷ #prompt-engineering (5 messages):


OpenAI ▷ #api-discussions (5 messages):


HuggingFace ▷ #announcements (1 messages):

Links mentioned:


HuggingFace ▷ #general (170 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #cool-finds (7 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (10 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (101 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #diffusion-discussions (7 messages):


HuggingFace ▷ #NLP (5 messages):


HuggingFace ▷ #diffusion-discussions (7 messages):


LM Studio ▷ #💬-general (141 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (73 messages🔥🔥):

Links mentioned:


LM Studio ▷ #announcements (1 messages):

Links mentioned:

👾 LM Studio - Discover and run local LLMs: Find, download, and experiment with local LLMs


LM Studio ▷ #🧠-feedback (12 messages🔥):

Links mentioned:

LM Studio: experience the magic of LLMs with Zero technical expertise: Your guide to Zero configuration Local LLMs on any computer.


LM Studio ▷ #🎛-hardware-discussion (27 messages🔥):

Links mentioned:

I Saw W Gus Fring GIF - I Saw W Gus Fring Gus - Discover & Share GIFs: Click to view the GIF


LM Studio ▷ #🧪-beta-releases-chat (15 messages🔥):


LM Studio ▷ #langchain (1 messages):


LM Studio ▷ #crew-ai (1 messages):


LM Studio ▷ #open-interpreter (1 messages):

phoenix2574: <@294336444393324545> I'm using Mixtral and it seems to work alright


Nous Research AI ▷ #off-topic (8 messages🔥):

Links mentioned:

Shocked Shocked Cat GIF - Shocked Shocked cat Silly cat - Discover & Share GIFs: Click to view the GIF


Nous Research AI ▷ #interesting-links (16 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (198 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (11 messages🔥):

Links mentioned:


Mistral ▷ #general (159 messages🔥🔥):

Links mentioned:


Mistral ▷ #models (47 messages🔥):

Links mentioned:

Mistral 7B is 187x cheaper compared to GPT-4 : Mistral AI 7B model can be a great alternative to GPT 3.5 or 4 models with 187x cheaper in cost. Find calculation inside to find out cost comparison between the models.


Mistral ▷ #finetuning (16 messages🔥):

Links mentioned:


Mistral ▷ #showcase (2 messages):


Mistral ▷ #office-hour (1 messages):

Links mentioned:

Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


LAION ▷ #general (114 messages🔥🔥):

Links mentioned:


LAION ▷ #research (9 messages🔥):

Links mentioned:


Eleuther ▷ #general (42 messages🔥):

Links mentioned:


Eleuther ▷ #research (71 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (3 messages):


Eleuther ▷ #interpretability-general (2 messages):

Links mentioned:


Eleuther ▷ #lm-thunderdome (3 messages):


OpenAccess AI Collective (axolotl) ▷ #general (21 messages🔥):

Links mentioned:

GitHub - mlabonne/llm-autoeval: Automatically evaluate your LLMs in Google Colab: Automatically evaluate your LLMs in Google Colab. Contribute to mlabonne/llm-autoeval development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (26 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (17 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #rlhf (9 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #runpod-help (7 messages):

Links mentioned:

Cloud motd by winglian · Pull Request #1235 · OpenAccess-AI-Collective/axolotl: Description sometimes in runpod, the extra disk gets mounted and it clobbers the axolotl dir. add a motd to help users w a solution


Perplexity AI ▷ #general (48 messages🔥):

Links mentioned:


Perplexity AI ▷ #sharing (9 messages🔥):


Perplexity AI ▷ #pplx-api (9 messages🔥):

Links mentioned:

Shortcuts: no description found


LlamaIndex ▷ #announcements (1 messages):

Links mentioned:


LlamaIndex ▷ #blog (5 messages):

Links mentioned:


LlamaIndex ▷ #general (38 messages🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (6 messages):

Links mentioned:


LangChain AI ▷ #general (15 messages🔥):

Links mentioned:


LangChain AI ▷ #langserve (8 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (8 messages🔥):

Links mentioned:


Latent Space ▷ #ai-general-chat (23 messages🔥):

Links mentioned:


CUDA MODE ▷ #general (3 messages):


CUDA MODE ▷ #cuda (7 messages):

Links mentioned:


CUDA MODE ▷ #torch (1 messages):

Links mentioned:

pytorch/torch/_inductor/kernel/mm.py at main · pytorch/pytorch: Tensors and Dynamic neural networks in Python with strong GPU acceleration - pytorch/pytorch


CUDA MODE ▷ #beginner (4 messages):

Links mentioned:

Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


CUDA MODE ▷ #jax (2 messages):


DiscoResearch ▷ #general (6 messages):


DiscoResearch ▷ #embedding_dev (4 messages):


DiscoResearch ▷ #discolm_german (5 messages):


LLM Perf Enthusiasts AI ▷ #general (2 messages):


LLM Perf Enthusiasts AI ▷ #gpt4 (1 messages):

.psychickoala: does azure have a gpt 4 vision model


LLM Perf Enthusiasts AI ▷ #speed (2 messages):

Links mentioned:

Tweet from Varun Shenoy (@varunshenoy_): Introducing 𝗦𝘂𝗽𝗲𝗿 𝗝𝗦𝗢𝗡 𝗠𝗼𝗱𝗲, a framework for low latency structured output generation from LLMs. Generate JSON up to 𝟮𝟬𝘅 𝗳𝗮𝘀𝘁𝗲𝗿 from OpenAI and open source models. ❌ No need to...


LLM Perf Enthusiasts AI ▷ #cost (1 messages):


LLM Perf Enthusiasts AI ▷ #reliability (3 messages):


Alignment Lab AI ▷ #general-chat (2 messages):


Alignment Lab AI ▷ #join-in (4 messages):

Links mentioned:

Threat Prompt - AI Security: no description found


Datasette - LLM (@SimonW) ▷ #ai (2 messages):


AI Engineer Foundation ▷ #events (1 messages):

._z: @everyone Weekly meeting is beginning now. 😄