Frozen AI News archive

Sama says: GPT-5 soon

**Sam Altman** at Davos highlighted that his top priority is launching the new model, likely called **GPT-5**, while expressing uncertainty about **Ilya Sutskever**'s employment status. **Itamar from Codium** introduced the concept of **Flow Engineering** with **AlphaCodium**, gaining attention from **Andrej Karpathy**. On the **TheBloke Discord**, engineers discussed a **multi-specialty mixture-of-experts (MOE) model** combining seven distinct 7 billion parameter models specialized in law, finance, and medicine. Debates on **8-bit fine-tuning** and the use of **bitsandbytes** with GPU support were prominent. Discussions also covered **model merging** using tools like **Mergekit** and compatibility with **Alpaca format**. Interest in optimizing AI models on **AMD** hardware using **AOCL blas and lapack libraries** with **llama.cpp** was noted. Users experimented with AI for command line tasks, and the **Mixtral MoE model** was refined to surpass larger models in coding ability. Comparisons among LLMs such as **GPT-3.5**, **Mixtral**, **Gemini Pro**, and **GPT-4** focused on knowledge depth, problem-solving, and speed, especially for coding tasks.

Canonical issue URL

https://www.youtube.com/watch?v=QFXp_TU-bO8

Sama at Davos:

Separately, Itamar from Codium coined Flow Engineering with AlphaCodium, picked up by Karpathy.

--

Table of Contents

[TOC]

TheBloke Discord Summary

TheBloke Channel Summaries

▷ #general (1151 messages🔥🔥🔥):

Links mentioned:

▷ #characters-roleplay-stories (425 messages🔥🔥🔥):

Links mentioned:

▷ #training-and-fine-tuning (29 messages🔥):

Links mentioned:

jondurbin: Weights & Biases, developer tools for machine learning

▷ #model-merging (10 messages🔥):

▷ #coding (2 messages):


Nous Research AI Discord Summary

Nous Research AI Channel Summaries

▷ #off-topic (22 messages🔥):

Links mentioned:

▷ #interesting-links (13 messages🔥):

Links mentioned:

▷ #general (338 messages🔥🔥):

Links mentioned:

▷ #ask-about-llms (44 messages🔥):

Links mentioned:


LM Studio Discord Summary

GPU Tango: VRAM and Resource Management in Focus: Engineers in the guild discussed GPU offload settings in LM Studio, noting that setting the GPU offload to -1 utilizes all layers but may show low GPU utilization. Recommendations were made for using Nvidia P40 GPUs as a cost-effective performance solution, and concerns were raised about potential VRAM allocation conflicts when running AI models alongside intensive applications like gaming.

LM Studio Beta V4 Debuts: Beta V4 (0.2.11 release candidate) of LM Studio has been released, featuring a model search page with VRAM fit estimates and support for new 2bit quants. Download links were provided, and it was stated that plans for open sourcing or adding a plugin system are in development, assuring LM Studio will remain free for personal use.

Dispatches from the Hardware Front: Relevant hardware discussions included power supply considerations for dual RTX 3090 setups, where a 1200W+ PSU was advised. Creative solutions for fitting large GPUs into small cases were exchanged, emphasizing the ingenuity of the engineers in optimizing their AI computing rigs.

CrewAI: Framework and Performance Insights: The CrewAI Multi-Agent Framework and its integration with the LM Studio API were highlighted, with a mention of leveraging specific agents for dedicated tasks like internet search. Benchmarks for multiple models using CrewAI were promised, along with sample code once the user's work is completed.

Model Performance and Usage: It was reported that local models, albeit operational for repeated function calls, are not as impressive as the 3.5T model. The Skyrim ChatGPT mod's image recognition was spotlighted as a parallel task that competes for GPU resources with other processes. LM Studio installation issues and an unspecified model error on a 24G RAM laptop also emerged, with the latter redirected to technical support channels for further assistance.

LM Studio Channel Summaries

▷ #💬-general (200 messages🔥🔥):

Links mentioned:

▷ #🤖-models-discussion-chat (11 messages🔥):

▷ #🧠-feedback (6 messages):

▷ #🎛-hardware-discussion (79 messages🔥🔥):

▷ #🧪-beta-releases-chat (37 messages🔥):

▷ #autogen (1 messages):

yagilb: https://discord.com/channels/1110598183144399058/1197707651438624849

▷ #langchain (1 messages):

▷ #avx-beta (1 messages):

▷ #crew-ai (10 messages🔥):


Perplexity AI Discord Summary

Perplexity AI Channel Summaries

▷ #announcements (1 messages):

▷ #general (186 messages🔥🔥):

Links mentioned:

▷ #sharing (5 messages):

Links mentioned:

▷ #pplx-api (6 messages):


OpenAccess AI Collective (axolotl) Discord Summary

OpenAccess AI Collective (axolotl) Channel Summaries

▷ #general (140 messages🔥🔥):

Links mentioned:

▷ #axolotl-dev (26 messages🔥):

Links mentioned:

▷ #general-help (2 messages):

▷ #datasets (6 messages):

▷ #rlhf (5 messages):

▷ #replicate-help (2 messages):


OpenAI Discord Summary

OpenAI Channel Summaries

▷ #ai-discussions (110 messages🔥🔥):

Links mentioned:

▷ #gpt-4-discussions (38 messages🔥):

Links mentioned:

OpenAI Status: no description found

▷ #prompt-engineering (16 messages🔥):

▷ #api-discussions (16 messages🔥):


Mistral Discord Summary

Self-hosting and API wonders with Mistral 7B: Discussions across channels showed interest in self-hosting Mistral 7B and utilizing it with Python applications, with various users offering assistance and tool suggestions. Concerns around commercial application data privacy and technical issues with quantization affecting performance were raised.

The Quandary of Long Texts: Users debated on processing long texts with Mistral and the 32K token limit. While documentation mentions this limit, the practical token cap varies based on model size and task-specific conditions.

Frustrations and Recommendations in Fine-Tuning: The community reported challenges when fine-tuning Mistral 7B, such as persistence of old prompt responses and GPU memory difficulties on an RTX 4090. Additionally, the correct implementation for Mistral in the HF trainer and finding a good GGUF format model were subjects of inquiry.

Hearty Discussions on Deployment and Tool Integration: Participants exchanged experiences with integrating tools such as Deep Chat, highlighting its simplicity over more complex setups like Open Copilot. Personal experiences related to open-source projects and international moves in the tech sector were also shared among members.

Guidance for Aspiring Coders and LLaMa Musings: Recommendations for beginner coders pointed towards Harvard’s CS50 course and learning through hands-on experience. Curiosity was piqued by a Reddit discussion about Meta AI's LLaMa 3 being trained on an impressive array of 600,000 H100s.

Mistral Channel Summaries

▷ #general (31 messages🔥):

Links mentioned:

mistralai/Mixtral-8x7B-Instruct-v0.1 · Hugging Face: no description found

▷ #models (22 messages🔥):

Links mentioned:

cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser · Hugging Face: no description found

▷ #deployment (8 messages🔥):

▷ #finetuning (7 messages):

▷ #showcase (104 messages🔥🔥):

Links mentioned:

▷ #random (4 messages):

Links mentioned:

Reddit - Dive into anything: no description found


HuggingFace Discord Discord Summary

HuggingFace Discord Channel Summaries

▷ #announcements (1 messages):

Links mentioned:

TL;DR: Watch the…"](https://huggingface.co/posts/SkalskiP/421333989856413): no description found

Everyone loves…"](https://huggingface.co/posts/s3nh/683576905550627): no description found

▷ #general (79 messages🔥🔥):

Links mentioned:

▷ #today-im-learning (4 messages):

▷ #cool-finds (4 messages):

▷ #i-made-this (7 messages):

Links mentioned:

▷ #diffusion-discussions (3 messages):

▷ #computer-vision (27 messages🔥):

▷ #NLP (2 messages):

▷ #diffusion-discussions (3 messages):


LAION Discord Summary

LAION Channel Summaries

▷ #general (78 messages🔥🔥):

Links mentioned:

▷ #research (5 messages):

Links mentioned:


Eleuther Discord Summary

Eleuther Channel Summaries

▷ #general (28 messages🔥):

Links mentioned:

Build software better, together: GitHub is where people build software. More than 100 million people use GitHub to discover, fork, and contribute to over 420 million projects.

▷ #research (26 messages🔥):

Links mentioned:

▷ #interpretability-general (1 messages):

jsai_51448: What is mech interp vs. concept interp vs. dev interp?

▷ #lm-thunderdome (9 messages🔥):

Links mentioned:

▷ #multimodal-general (1 messages):

Links mentioned:

▷ #gpt-neox-dev (2 messages):

Links mentioned:

Minor changes by segyges · Pull Request #1125 · EleutherAI/gpt-neox: Changes default output for docker container Renames docker pythia config to indicate it is docker pythia config Fix unit test for evaluate function


LlamaIndex Discord Discord Summary

LlamaIndex Discord Channel Summaries

▷ #blog (4 messages):

Links mentioned:

LlamaIndex RAG Hackathon (in-person only): Think Beyond Chatbots: Unleashing the Potential of AI Agents

▷ #general (34 messages🔥):

Links mentioned:

▷ #ai-discussion (3 messages):


Latent Space Discord Summary

Latent Space Channel Summaries

▷ #ai-general-chat (27 messages🔥):

Links mentioned:

▷ #llm-paper-club (1 messages):

Links mentioned:

How EleutherAI Trains and Releases LLMs: Interview with Stella Biderman — Gradient Dissent: Exploring Machine Learning, AI, Deep Learning, Computer Vision — Overcast: no description found


LangChain AI Discord Summary

LangChain AI Channel Summaries

▷ #general (14 messages🔥):

Links mentioned:

Chains | 🦜️🔗 Langchain: Chains refer to sequences of calls - whether to an LLM, a tool, or a

▷ #langserve (7 messages):

Links mentioned:

▷ #share-your-work (4 messages):

Links mentioned:


DiscoResearch Discord Summary

DiscoResearch Channel Summaries

▷ #general (4 messages):

Links mentioned:

▷ #benchmark_dev (4 messages):

▷ #embedding_dev (1 messages):

▷ #discolm_german (13 messages🔥):


Skunkworks AI Discord Summary

Skunkworks AI Channel Summaries

▷ #general (9 messages🔥):

Links mentioned:

Break the Sequential Dependency of LLM Inference Using Lookahead Decoding | LMSYS Org: TL;DR: We introduce lookahead decoding, a new, exact, and parallel decoding algorithm to accelerate LLM inference. Look...

▷ #off-topic (1 messages):

pradeep1148: https://www.youtube.com/watch?v=POgLwYxDGYk


LLM Perf Enthusiasts AI Discord Summary

LLM Perf Enthusiasts AI Channel Summaries

▷ #gpt4 (3 messages):

▷ #opensource (1 messages):

thisisnotawill: yeah im using the instruct model from anyscale

▷ #resources (1 messages):

▷ #feedback-meta (3 messages):

▷ #openai (1 messages):


Alignment Lab AI Discord Summary

Only 1 channel had activity, so no need to summarize...

imonenext: Does anyone have a Gemini Pro key?


Datasette - LLM (@SimonW) Discord Summary

Only 1 channel had activity, so no need to summarize...

Links mentioned:

fix: allow local models to work without internet connection by hydrosquall · Pull Request #18 · simonw/llm-gpt4all: Motivation Currently, the library tries to download the model even if it already exists locally, which prevents offline use. Fixes #10 , applying a code hint and investigation from @rotterb Changes...


The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.