Frozen AI News archive

Trust in GPTs at all time low

**Discord communities** were analyzed with **21 guilds**, **312 channels**, and **8530 messages** reviewed, saving an estimated **628 minutes** of reading time. Discussions highlighted challenges with **GPTs** and the **GPT store**, including critiques of the **knowledge files capability** and context management issues. The **CUDA MODE Discord** was introduced for CUDA coding support. Key conversations in the **TheBloke Discord** covered **Xeon** GPU server cost-effectiveness, **Llama3** and **Mistral Medium** model comparisons, **LLaVA-1.6**'s visual reasoning and OCR capabilities, and the leaked **Miqu** 70B model. Technical topics included fine-tuning **TinyLlama** and **MiquMaid+Euryale** models, and model merging with examples like **Harmony-4x7B-bf16** and **Smaug-34B-v0.1**. The **Nous Research AI Discord** discussed style influence in LLMs, quantization issues, **Bittensor** incentives for AI model improvements, and the identification of **MIQU** as **Mistral Medium**. The release of the **Open Hermes 2.5 dataset** on **Hugging Face** was also announced. *"Discussions pointed towards the need for better context management in GPTs, contrasting with OpenAI's no-code approach."*

Canonical issue URL

It's been about 3 months since GPTs were released and ~a month since the GPT store was launched. But the reviews have been brutal:

image.png

Nick Dobos (of Grimoire fame) also blasted the entire knowledge files capability - it seems the RAG system naively includes 40k characters' worth of context from docs every time, reducing available context and adherence to system prompts.

image.png

All pointing towards needing greater visibility for context management in GPTs, which is somewhat at odds with OpenAI's clear no-code approach.


In meta (pun?) news, warm welcome to our newest Discord scraped - Saroufim et al's CUDA MODE discord! Lots of nice help for those new to CUDA coding

image.png

image.png


Table of Contents

[TOC]


PART 1: High level Discord summaries

TheBloke Discord Summary


Nous Research AI Discord Summary


Mistral Discord Summary


LM Studio Discord Summary


OpenAI Discord Summary


OpenAccess AI Collective (axolotl) Discord Summary


HuggingFace Discord Summary

Training Headaches and Quantum Leaps: A user faced a loss flatline during model training; a batch size reduction to 1 and the potential use of EarlyStoppingCallback were suggested solutions. Another proposed solution was 4bit quantization to tackle training instability, which might help conserve VRAM albeit at some cost to model accuracy.

Seeking Specialized Language Models: There was an inquiry about language models tailored to tech datasets around Arduino, ESP32, and Raspberry Pi, suggesting a demand for LLMs with specialized knowledge.

Tech Enthusiast's Project Spotlight: Showcasing a range of projects from seeking feedback on a thesis tweet, to offering access to a Magic: The Gathering model space, as well as a custom pipeline solution for the moondream1 model with a related pull request.

Experimental Models Run Lean: A NeuralBeagle14-7b model was successfully demonstrated on a local 8GB GPU, piquing the interest of those looking to optimize resource usage, which is key for maintainable AI solutions here.

Scholarly Papers and AI Explorations: A paper on language model compression algorithms has been shared, discussing the balance between efficiency and accuracy in methods such as pruning, quantization, and distillation, which could be very pertinent in the ongoing dialogue about optimizing model performance.


LAION Discord Summary


Perplexity AI Discord Summary


LangChain AI Discord Summary


LlamaIndex Discord Summary


Eleuther Discord Summary


CUDA MODE (Mark Saroufim) Discord Summary


DiscoResearch Discord Summary


Latent Space Discord Summary

LLM Perf Enthusiasts AI Discord Summary


Skunkworks AI Discord Summary


AI Engineer Foundation Discord Summary


Alignment Lab AI Discord Summary


Datasette - LLM (@SimonW) Discord Summary


The Ontocord (MDEL discord) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

TheBloke ▷ #general (1315 messages🔥🔥🔥):

Links mentioned:


TheBloke ▷ #characters-roleplay-stories (885 messages🔥🔥🔥):

Links mentioned:


TheBloke ▷ #training-and-fine-tuning (30 messages🔥):

Links mentioned:


TheBloke ▷ #model-merging (15 messages🔥):

Links mentioned:


TheBloke ▷ #coding (2 messages):

Links mentioned:

Students who grew up with search engines might change STEM education forever: Professors are struggling to teach Gen Z


Nous Research AI ▷ #ctx-length-research (6 messages):


Nous Research AI ▷ #off-topic (13 messages🔥):

Links mentioned:


Nous Research AI ▷ #benchmarks-log (2 messages):


Nous Research AI ▷ #interesting-links (77 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

Links mentioned:


Nous Research AI ▷ #general (673 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (32 messages🔥):

Links mentioned:


Mistral ▷ #general (363 messages🔥🔥):

Links mentioned:


Mistral ▷ #models (17 messages🔥):

Links mentioned:


Mistral ▷ #ref-implem (1 messages):


Mistral ▷ #finetuning (33 messages🔥):


Mistral ▷ #la-plateforme (8 messages🔥):

Links mentioned:

👾 LM Studio - Discover and run local LLMs: Find, download, and experiment with local LLMs


LM Studio ▷ #💬-general (123 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (100 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (21 messages🔥):


LM Studio ▷ #🎛-hardware-discussion (157 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (1 messages):

mike_50363: Is the non-avx2 beta version going to be updated to 2.12?


OpenAI ▷ #ai-discussions (137 messages🔥🔥):

Links mentioned:

What is Microsoft's Approach to AI? | Microsoft Source: We believe AI is the defining technology of our time. Read about our approach to AI for infrastructure, research, responsibility and social good.


OpenAI ▷ #gpt-4-discussions (85 messages🔥🔥):

Links mentioned:

Pricing: Simple and flexible. Only pay for what you use.


OpenAI ▷ #prompt-engineering (62 messages🔥🔥):


OpenAI ▷ #api-discussions (62 messages🔥🔥):


OpenAccess AI Collective (axolotl) ▷ #general (109 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (26 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (42 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #rlhf (4 messages):


OpenAccess AI Collective (axolotl) ▷ #runpod-help (5 messages):


HuggingFace ▷ #general (149 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (3 messages):


HuggingFace ▷ #i-made-this (11 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (1 messages):

Links mentioned:

A Comprehensive Survey of Compression Algorithms for Language Models: How can we compress language models without sacrificing accuracy? The number of compression algorithms for language models is rapidly growing to benefit from remarkable advances of recent language mod...


HuggingFace ▷ #diffusion-discussions (1 messages):


HuggingFace ▷ #computer-vision (1 messages):

merve3234: it's not an error but rather a warning, feel free to ignore


HuggingFace ▷ #NLP (2 messages):

Links mentioned:

GitHub - joshuasundance-swca/llamacpp-langchain-neuralbeagle-demo: a small demo repo to show how I got neuralbeagle14-7b running locally on my 8GB GPU: a small demo repo to show how I got neuralbeagle14-7b running locally on my 8GB GPU - GitHub - joshuasundance-swca/llamacpp-langchain-neuralbeagle-demo: a small demo repo to show how I got neuralbe...


HuggingFace ▷ #diffusion-discussions (1 messages):


LAION ▷ #general (72 messages🔥🔥):

Links mentioned:


LAION ▷ #research (19 messages🔥):

Links mentioned:


Perplexity AI ▷ #general (37 messages🔥):

Links mentioned:


Perplexity AI ▷ #sharing (2 messages):

Links mentioned:

How I Use Perplexity AI to Source Content Ideas for LinkedIn: A dive deep into how I've been using Perplexity AI to revolutionize my content creation process for social media. Whether you're a content creator, business ...


Perplexity AI ▷ #pplx-api (31 messages🔥):


LangChain AI ▷ #general (38 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (9 messages🔥):

Links mentioned:

Unveiling the Future of AI Collaboration with LangGraph: Embracing Multi-Agent Workflows: Ankush k Singal


LangChain AI ▷ #tutorials (1 messages):

Links mentioned:

How To USE New OpenAI Embeddings Model with LangChain 🦜🔗: OpenAI introduces new Embeddings model. They are releasing new models, reducing prices for GPT-3.5 Turbo, and introducing new ways for developers to manage A...


LlamaIndex ▷ #blog (2 messages):

Links mentioned:

no title found: no description found


LlamaIndex ▷ #general (43 messages🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):

leonms123: Hi would anyone be willing to teach me ML using python 😄


Eleuther ▷ #general (10 messages🔥):

Links mentioned:

Tweet from Arthur Mensch (@arthurmensch): An over-enthusiastic employee of one of our early access customers leaked a quantised (and watermarked) version of an old model we trained and distributed quite openly. To quickly start working with ...


Eleuther ▷ #research (14 messages🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (3 messages):

Links mentioned:

Infini-gram: Scaling Unbounded n-gram Language Models to a Trillion Tokens: Are n-gram language models still relevant in this era of neural large language models (LLMs)? Our answer is yes, and we show their values in both text analysis and improving neural LLMs. Yet this nece...


Eleuther ▷ #lm-thunderdome (12 messages🔥):

Links mentioned:


Eleuther ▷ #multimodal-general (4 messages):


CUDA MODE (Mark Saroufim) ▷ #general (17 messages🔥):

Links mentioned:


CUDA MODE (Mark Saroufim) ▷ #triton (1 messages):

iloveh8: Got it thank you


CUDA MODE (Mark Saroufim) ▷ #cuda (7 messages):

Links mentioned:


CUDA MODE (Mark Saroufim) ▷ #torch (1 messages):


CUDA MODE (Mark Saroufim) ▷ #beginner (12 messages🔥):

Links mentioned:


CUDA MODE (Mark Saroufim) ▷ #pmpp-book (3 messages):


DiscoResearch ▷ #mixtral_implementation (21 messages🔥):

Links mentioned:


DiscoResearch ▷ #general (10 messages🔥):

Links mentioned:


DiscoResearch ▷ #embedding_dev (8 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (1 messages):

philipmay: How did you generate this plot for DiscoResearch/DiscoLM_German_7b_v1 ?


LLM Perf Enthusiasts AI ▷ #gpt4 (2 messages):


LLM Perf Enthusiasts AI ▷ #opensource (3 messages):

Links mentioned:

miqudev/miqu-1-70b · Hugging Face: no description found


LLM Perf Enthusiasts AI ▷ #openai (5 messages):


Skunkworks AI ▷ #general (5 messages):

Links mentioned:


AI Engineer Foundation ▷ #general (5 messages):

Links mentioned:

Agents & Tools: no description found


Alignment Lab AI ▷ #general-chat (2 messages):


Alignment Lab AI ▷ #alignment-lab-announcements (1 messages):

Links mentioned:


Datasette - LLM (@SimonW) ▷ #llm (1 messages):