Frozen AI News archive

GPT4Turbo A/B Test: gpt-4-1106-preview

**OpenAI** released a new **GPT-4 Turbo** version, prompting a natural experiment in summarization comparing the November 2023 and January 2024 versions. The **TheBloke** Discord discussed troubleshooting model loading errors with **OpenHermes-2.5-Mistral-7B-4.0bpw** and **exllamav2**, debates on **RHEL** in ML, dataset generation for understanding GPT flaws, and running LLMs like **Llama** and **Mistral** on consoles. **LangChain** fine-tuning challenges for **Llama2** were also noted. The **OpenAI** Discord highlighted **GPT-4** speed inconsistencies, API vs web performance, prompt engineering with **GPT-3.5** and **GPT-4 Turbo**, and **DALL-E** typo issues in image text. Discussions included NLP tools like *semantic-text-splitter* and collaboration concerns with **GPT-4 Vision** on **Azure**. The **Nous Research AI** Discord focused on extending context windows with **Mistral instruct v0.2**, **MistralLite**, and **LLaMA-2-7B-Chat** achieving 16,384 token context, plus alternatives like **SelfExtend** for context extension without fine-tuning. The societal impact of AI technology was also considered.

Canonical issue URL

OpenAI released a new GPT4 Turbo version yesterday (our notes here). We're using this opportunity to conduct a natural experiment for summarization. This version is generated with the "old" GPT4T from Nov 2023 (Dev Day), stay tuned for the next email with the 2024 Jan 25th version for comparison and commentary.

--

Table of Contents

[TOC]

PART 1: High level Discord summaries

TheBloke Discord Summary


OpenAI Discord Summary


Nous Research AI Discord Summary


OpenAccess AI Collective (axolotl) Discord Summary


LM Studio Discord Summary


Mistral Discord Summary

GPU Rental to Mistral Integration: GPU rental options including runpod, vast, and lambda were discussed, with Kaggle also mentioned as offering free access up to 30 hours weekly. Mistral 7B use cases and integration challenges were shared, seeking insights for effective implementations, referencing Hugging Face's Mistral 7B model.

Memory Matters in Model Finetuning: Discourse around Mixtral's large memory appetite for inference highlighted that 26GB is required across four T4 GPUs, with actual usage potentially higher than expected. Efficiency debates compared exllamav2 and bnb 4 bit for quantization, with a nod to exllamav2 GitHub for running LLMs efficiently.

Evaluating LLMs Beyond Traditional Metrics: Emphasis was placed on the inadequacy of BLEU and ROUGE metrics for LLMs, suggesting elo rankings (arena.lmsys.org) and benchmarks like MMLU and Alpaca eval for better performance measurements. The introduction of a normalized Alpaca eval market version was mentioned without further details.

Creative Showcases and Random RAG Tips: A tool named SoContextual.com that integrates AI for browser searches including DOM references was showcased, working with MistralAI and spotted on Hacker News. Meanwhile, prompt optimization for RAG applications was touched upon, recommending DSPy and sharing a prompting guide.

Platform Puzzles and API Anomalies: A billing page bug causing the monthly limit to reset to €150 was reported, while API bugs concerning the 'max_tokens' parameter and early stopping issues were discussed, including a posted GitHub issue. Hosting queries affirmed Mistral's API is located on Azure in Sweden.


Eleuther Discord Summary


LAION Discord Summary


Perplexity AI Discord Summary


HuggingFace Discord Summary


LlamaIndex Discord Summary


Latent Space Discord Summary


DiscoResearch Discord Summary


LLM Perf Enthusiasts AI Discord Summary


LangChain AI Discord Summary

Please note that any direct references to usernames were included as they were considered contextually relevant based on the information provided.


Datasette - LLM (@SimonW) Discord Summary

PART 2: Detailed by-Channel summaries and links

TheBloke ▷ #general (1212 messages🔥🔥🔥):

Links mentioned:


TheBloke ▷ #characters-roleplay-stories (74 messages🔥🔥):

Links mentioned:


TheBloke ▷ #training-and-fine-tuning (20 messages🔥):


TheBloke ▷ #model-merging (19 messages🔥):


TheBloke ▷ #coding (1 messages):


OpenAI ▷ #ai-discussions (35 messages🔥):

Links mentioned:

TuringsSolutions/PFAF750 · Datasets at Hugging Face: no description found


OpenAI ▷ #gpt-4-discussions (105 messages🔥🔥):


OpenAI ▷ #prompt-engineering (558 messages🔥🔥🔥):

Links mentioned:

How do you maintain historical context in repeat API calls?: Each time I make a call to the API it starts off with no prior context, unlike the chat.openai.com scenario. Is there a way to maintain state of the model during a session? response = openai.Completi...


OpenAI ▷ #api-discussions (558 messages🔥🔥🔥):

Links mentioned:

How do you maintain historical context in repeat API calls?: Each time I make a call to the API it starts off with no prior context, unlike the chat.openai.com scenario. Is there a way to maintain state of the model during a session? response = openai.Completi...


Nous Research AI ▷ #ctx-length-research (7 messages):

Links mentioned:

config.json · mistralai/Mistral-7B-Instruct-v0.2 at main: no description found


Nous Research AI ▷ #off-topic (5 messages):

Links mentioned:


Nous Research AI ▷ #benchmarks-log (2 messages):

Links mentioned:

TheBloke/Everyone-Coder-33B-Base-GGUF · Hugging Face: no description found


Nous Research AI ▷ #interesting-links (2 messages):

Links mentioned:


Nous Research AI ▷ #general (362 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (35 messages🔥):

Links mentioned:


Nous Research AI ▷ #project-obsidian (3 messages):


OpenAccess AI Collective (axolotl) ▷ #general (219 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (1 messages):


OpenAccess AI Collective (axolotl) ▷ #general-help (47 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #datasets (7 messages):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #rlhf (2 messages):


OpenAccess AI Collective (axolotl) ▷ #community-showcase (1 messages):

pradeep1148: https://www.youtube.com/watch?v=wlPxEq_Mtkc


LM Studio ▷ #💬-general (115 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (54 messages🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (4 messages):


LM Studio ▷ #🎛-hardware-discussion (11 messages🔥):


LM Studio ▷ #🧪-beta-releases-chat (46 messages🔥):


LM Studio ▷ #autogen (1 messages):


LM Studio ▷ #open-interpreter (10 messages🔥):


Mistral ▷ #general (163 messages🔥🔥):

Links mentioned:


Mistral ▷ #ref-implem (9 messages🔥):


Mistral ▷ #finetuning (3 messages):


Mistral ▷ #showcase (1 messages):

Links mentioned:

no title found: no description found


Mistral ▷ #random (8 messages🔥):

Links mentioned:

Prompt Engineering Guide: A Comprehensive Overview of Prompt Engineering


Mistral ▷ #la-plateforme (35 messages🔥):

Links mentioned:


Eleuther ▷ #general (29 messages🔥):

Links mentioned:

DiLoCo: Distributed Low-Communication Training of Language Models: Large language models (LLM) have become a critical component in many applications of machine learning. However, standard approaches to training LLM require a large number of tightly interconnected acc...


Eleuther ▷ #research (125 messages🔥🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (2 messages):

Links mentioned:

feat: Add Weights and Biases support by ayulockin · Pull Request #1339 · EleutherAI/lm-evaluation-harness: In #359 @parambharat did proposed to add support for W&B logging. However it was done before the big refactor that got in. As a user of both lm-evaluation-harness and wandb, I have opened this PR ...


Eleuther ▷ #gpt-neox-dev (16 messages🔥):

Links mentioned:

Tests fail when run with pytest --forked · Issue #1132 · EleutherAI/gpt-neox: Describe the bug When tests are run with pytest --forked per the instructions in /test/README.md, a large number of tests fail with the error: RuntimeError: Cannot re-initialize CUDA in forked subp...


LAION ▷ #general (59 messages🔥🔥):

Links mentioned:


LAION ▷ #research (71 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #general (87 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (4 messages):

Links mentioned:

I use Perplexity MORE than Google and ChatGPT: Main Takaways From this Video: "I use Perplexity more than ChatGPT, BARD, and Microsoft Copilots for five main reasons, including its use in content creation...


Perplexity AI ▷ #pplx-api (5 messages):


HuggingFace ▷ #announcements (3 messages):

Links mentioned:

I launched my first competition !

Goal : Use AI to…"](https://huggingface.co/posts/Tonic/783827682062088): no description found

Well, yes, if the models are…"](https://huggingface.co/posts/vicgalle/320544784279721): no description found


HuggingFace ▷ #general (40 messages🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (1 messages):


HuggingFace ▷ #cool-finds (7 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (12 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (3 messages):

Links mentioned:

Lumiere: A Space-Time Diffusion Model for Video Generation: We introduce Lumiere -- a text-to-video diffusion model designed for synthesizing videos that portray realistic, diverse and coherent motion -- a pivotal challenge in video synthesis. To this end, we ...


HuggingFace ▷ #diffusion-discussions (1 messages):

spikespiegel5112: How to load LoRA model in local?


HuggingFace ▷ #computer-vision (5 messages):

Links mentioned:

Gemini Pro Vision AI API Documentation (swift-api-swift-api-default) | RapidAPI: no description found


HuggingFace ▷ #NLP (15 messages🔥):

Links mentioned:

talks.cam : Replicating and auditing black-box Language Models.: no description found


HuggingFace ▷ #diffusion-discussions (1 messages):

spikespiegel5112: How to load LoRA model in local?


HuggingFace ▷ #gradio-announcements (1 messages):

Links mentioned:

gradio/CHANGELOG.md at main · gradio-app/gradio: Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work! - gradio-app/gradio


LlamaIndex ▷ #announcements (1 messages):

Links mentioned:

LlamaIndex Webinar: Efficient Parallel Function Calling Agents with LLMCompiler · Zoom · Luma: LLMs are great at reasoning and taking actions. But previous frameworks for agentic reasoning (e.g. ReAct) were primarily focused on sequential reasoning, leading to higher...


LlamaIndex ▷ #blog (7 messages):

Links mentioned:


LlamaIndex ▷ #general (38 messages🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (5 messages):

Links mentioned:


Latent Space ▷ #ai-general-chat (36 messages🔥):

Links mentioned:


Latent Space ▷ #ai-event-announcements (1 messages):

Links mentioned:


Latent Space ▷ #llm-paper-club (8 messages🔥):

Links mentioned:

Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling: How do large language models (LLMs) develop and evolve over the course of training? How do these patterns change as models scale? To answer these questions, we introduce \textit{Pythia}, a suite of 16...


DiscoResearch ▷ #mixtral_implementation (2 messages):

Links mentioned:

Mixtral branch: What option should I choose when I want to do some finetuning after the merge? · Issue #116 · cg123/mergekit: The parameter description of "hidden" and "random" does not exactly explain what to do when I want to finetune later. Is it even useful (possible) to finetune after merging with &q...


DiscoResearch ▷ #general (23 messages🔥):

Links mentioned:


DiscoResearch ▷ #embedding_dev (12 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (6 messages):

Links mentioned:

Tweet from Nils Reimers (@Nils_Reimers): @OttoZastrow @jerryjliu0 Yes, embeddings is a massive focus for us, with amazing launches upcoming. E.g. OpenAI 54.3 on MIRACL with 3072 dimensions versus our upcoming 256 dimensional-like model wit...


LLM Perf Enthusiasts AI ▷ #embeddings (2 messages):

Links mentioned:

New embedding models and API updates: We are launching a new generation of embedding models, new GPT-4 Turbo and moderation models, new API usage management tools, and soon, lower pricing on GPT-3.5 Turbo.


LLM Perf Enthusiasts AI ▷ #announcements (1 messages):

mat_mto: Thanks Jeff! love all the work you're doing so far


LLM Perf Enthusiasts AI ▷ #openai (16 messages🔥):

Links mentioned:

New embedding models and API updates: We are launching a new generation of embedding models, new GPT-4 Turbo and moderation models, new API usage management tools, and soon, lower pricing on GPT-3.5 Turbo.


LangChain AI ▷ #general (12 messages🔥):

Links mentioned:

Finetuning Large Language Models: no description found


LangChain AI ▷ #langserve (3 messages):

Links mentioned:


LangChain AI ▷ #share-your-work (2 messages):


Datasette - LLM (@SimonW) ▷ #llm (3 messages):

Links mentioned:


Skunkworks AI ▷ #off-topic (1 messages):

pradeep1148: https://www.youtube.com/watch?v=wlPxEq_Mtkc


Skunkworks AI ▷ #bakklava-1 (1 messages):

arielnlee: Anyone working on bakklava-2?!