Frozen AI News archive

Less Lazy AI

The AI Discord summaries for early 2024 cover various community discussions and developments. Highlights include **20** guilds, **308** channels, and **10449** messages analyzed, saving an estimated **780 minutes** of reading time. Key topics include **Polymind Plugin Puzzle** integrating PubMed API, roleplay with **HamSter v0.2**, VRAM challenges in **Axolotl** training, fine-tuning tips for **FLAN-T5**, and innovative **model merging** strategies. The **Nous Research AI** community discussed GPT-4's lyricism issues, quantization techniques using `llama.cpp`, **frankenmerging** with models like **miqu-1-120b-GGUF**, anticipation for **Qwen2**, and tools like `text-generation-webui` and **ExLlamaV2**. The **LM Studio** community reported a bug where the app continues running after UI closure, with a workaround to forcibly terminate the process. These discussions reflect ongoing challenges and innovations in AI model training, deployment, and interaction.

Canonical issue URL

We've anecdotally gotten examples of refusal to follow instructions approximating laziness:

image.png

but it is hard to tell when it is luck of a bad draw or shameless self promotion.

This is why it's rare to get official confirmation from the top:

image.png

Still, laziness isn't a well defined technical term. It is frustrating to know that OpenAI has identified a problem and fixed it, but is not sharing what exactly it is.


Table of Contents

[TOC]

PART 1: High level Discord summaries

TheBloke Discord Summary


Nous Research AI Discord Summary


LM Studio Discord Summary


Mistral Discord Summary


LAION Discord Summary


HuggingFace Discord Summary


OpenAI Discord Summary


OpenAccess AI Collective (axolotl) Discord Summary


CUDA MODE (Mark Saroufim) Discord Summary

CUDA Curiosity Peaks: CUDA's dominance over OpenCL is attributed to its widespread popularity and Nvidia's support; Python continues to be a viable option for GPU computing, offering a balance between high-level programming ease and the nitty-gritty of kernel writing, as detailed in the CUDA MODE GitHub repository. Members also discussed the impact of compiler optimizations on CUDA performance, emphasizing the significance of even minute details in code, while advocating for robust CUDA learning through shared resources like tiny-cuda-nn.

PyTorch Parsers Perspire: Tips were shared on how to efficiently use the torch.compile API by specifying compiled layers, as seen in the gpt-fast repository. There's a bonafide interest in controlling the Torch compiler's behavior more finely, with the PyTorch documentation offering guidance. Amidst PyTorch preferences, TensorFlow also got a nod, mainly for Google's hardware and pricing.

Lecture Hype: Anticipation grows as CUDA MODE's fourth lecture on compute and memory architecture is heralded, with materials found in a repository jokingly criticized for its "increasingly inaccurately named" title, lecture2 repo. The lecture promises to delve into the nitty-gritty of blocks, warps, and memory hierarchies.

Job Market Buzzes: Aleph Alpha and Mistral AI are on the hunt for CUDA gurus, with roles integrating language model research into practical applications. Positions with a focus on GPU optimization and custom CUDA kernel development are up for grabs, detailed in the Aleph Alpha job listing and Mistral AI's opportunity.

CUDA Beginners Unite: Rust gained some spotlight in lower-level graphics programming and the discussion tilted towards its viability in CUDA programming, garnering interest for CUDA GPU projects in Rust, like rust-gpu for shaders. The Rust neural network scene is warming up, with projects like Kyanite and burn to ignite the coding fire.


Eleuther Discord Summary


Perplexity AI Discord Summary


LangChain AI Discord Summary


LlamaIndex Discord Summary


Latent Space Discord Summary

Relevant Links:


Datasette - LLM (@SimonW) Discord Summary


DiscoResearch Discord Summary


Alignment Lab AI Discord Summary


Skunkworks AI Discord Summary


LLM Perf Enthusiasts AI Discord Summary


AI Engineer Foundation Discord Summary


PART 2: Detailed by-Channel summaries and links

TheBloke ▷ #general (1738 messages🔥🔥🔥):

Links mentioned:


TheBloke ▷ #characters-roleplay-stories (678 messages🔥🔥🔥):

Links mentioned:


TheBloke ▷ #training-and-fine-tuning (18 messages🔥):

Links mentioned:


TheBloke ▷ #model-merging (5 messages):

Links mentioned:

tinyllama-merge.ipynb: GitHub Gist: instantly share code, notes, and snippets.


TheBloke ▷ #coding (6 messages):

Links mentioned:

GitHub - h2oai/h2ogpt: Private Q&A and summarization of documents+images or chat with local GPT, 100% private, Apache 2.0. Supports Mixtral, llama.cpp, and more. Demo: https://gpt.h2o.ai/ https://codellama.h2o.ai/: Private Q&A and summarization of documents+images or chat with local GPT, 100% private, Apache 2.0. Supports Mixtral, llama.cpp, and more. Demo: https://gpt.h2o.ai/ https://codellama.h2o.ai/ -...


Nous Research AI ▷ #off-topic (56 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (42 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (550 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (90 messages🔥🔥):

Links mentioned:


LM Studio ▷ #💬-general (225 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (149 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (11 messages🔥):


LM Studio ▷ #🎛-hardware-discussion (217 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (42 messages🔥):

Links mentioned:


LM Studio ▷ #autogen (4 messages):

Links mentioned:


LM Studio ▷ #langchain (2 messages):


Mistral ▷ #general (278 messages🔥🔥):

Links mentioned:


Mistral ▷ #models (45 messages🔥):

Links mentioned:


Mistral ▷ #deployment (17 messages🔥):

Links mentioned:


Mistral ▷ #finetuning (17 messages🔥):


Mistral ▷ #showcase (7 messages):

Links mentioned:

ExLlamaV2: The Fastest Library to Run LLMs: A fast inference library for running LLMs locally on modern consumer-class GPUshttps://github.com/turboderp/exllamav2https://colab.research.google.com/github...


Mistral ▷ #random (2 messages):


Mistral ▷ #la-plateforme (4 messages):


LAION ▷ #general (361 messages🔥🔥):

Links mentioned:


LAION ▷ #research (3 messages):

Links mentioned:

Qwen-VL-Max - a Hugging Face Space by Qwen: no description found


HuggingFace ▷ #general (289 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (5 messages):


HuggingFace ▷ #cool-finds (11 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (9 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (9 messages🔥):

Links mentioned:


HuggingFace ▷ #diffusion-discussions (12 messages🔥):

Links mentioned:


HuggingFace ▷ #computer-vision (7 messages):

Links mentioned:


HuggingFace ▷ #NLP (9 messages🔥):

Links mentioned:

merge-crew (Merge Crew): no description found


HuggingFace ▷ #diffusion-discussions (12 messages🔥):

Links mentioned:


OpenAI ▷ #ai-discussions (77 messages🔥🔥):

Links mentioned:

Beyond Consciousness in Large Language Models: An Investigation into the Existence of a “Soul” in…: Author: David Côrtes Cavalcante Publication Date: February 3, 2024 © CC BY Creative Commons Attribution


OpenAI ▷ #gpt-4-discussions (59 messages🔥🔥):


OpenAI ▷ #prompt-engineering (51 messages🔥):


OpenAI ▷ #api-discussions (51 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #general (119 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (88 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #other-llms (1 messages):

cf0913: https://huggingface.co/chatdb/natural-sql-7b


OpenAccess AI Collective (axolotl) ▷ #general-help (18 messages🔥):

Links mentioned:

GitHub - ctlllll/axolotl: Go ahead and axolotl questions: Go ahead and axolotl questions. Contribute to ctlllll/axolotl development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #rlhf (7 messages):


OpenAccess AI Collective (axolotl) ▷ #runpod-help (5 messages):


CUDA MODE (Mark Saroufim) ▷ #general (25 messages🔥):

Links mentioned:


CUDA MODE (Mark Saroufim) ▷ #cuda (99 messages🔥🔥):

Links mentioned:


CUDA MODE (Mark Saroufim) ▷ #torch (5 messages):

Links mentioned:


CUDA MODE (Mark Saroufim) ▷ #announcements (2 messages):


CUDA MODE (Mark Saroufim) ▷ #jobs (2 messages):

Links mentioned:


CUDA MODE (Mark Saroufim) ▷ #beginner (10 messages🔥):

Links mentioned:


CUDA MODE (Mark Saroufim) ▷ #pmpp-book (2 messages):


Eleuther ▷ #general (88 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (32 messages🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (1 messages):


Eleuther ▷ #lm-thunderdome (5 messages):


Eleuther ▷ #multimodal-general (9 messages🔥):


Eleuther ▷ #gpt-neox-dev (2 messages):


Perplexity AI ▷ #general (96 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (14 messages🔥):

Links mentioned:


Perplexity AI ▷ #pplx-api (7 messages):


LangChain AI ▷ #general (34 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (7 messages):

Links mentioned:


LangChain AI ▷ #tutorials (5 messages):

<ul>
  <li><strong>Next.js and LangChain for SMART Portfolios</strong>: User <code>@flo_walther</code> shared a <a href="https://www.youtube.com/watch?v=1LZltsK5nKI">YouTube video</a> on building a SMART portfolio website using <strong>Next.js 14, Langchain, Vercel AI SDK</strong>, and more, highlighting an AI chatbot that can be trained on your data.</li>
  <li><strong>Tutorial Troubles</strong>: <code>@stuartjatkinson</code> expressed frustration that steps in YouTube tutorials for LangChain have changed or yield errors when followed directly.</li>
  <li><strong>LangChain Starting Stumbles</strong>: <code>@treym1112</code> encountered errors while following the langchain quick tutorial on the LangChain website, specifically with the <strong>Ollama model</strong>, resulting in an <em>AttributeError</em> concerning the missing 'verbose' attribute.</li>
  <li><strong>LangChain Guide Launched</strong>: <code>@mehulgupta7991</code> announced the release of their book "<em>LangChain in your Pocket: Beginner's Guide to Building Generative AI Applications using LLMs</em>" on <a href="https://amzn.eu/d/dqQJzV1">Amazon</a>, describing it as a hands-on guide covering a range of use cases and LangServe deployment.</li>
  <li><strong>Meet the Data Scientist and Content Creator</strong>: <code>@mehulgupta7991</code> shared their professional background as a data scientist at DBS Bank and mentioned their Medium+YouTube channel "<em>Data Science in your Pocket</em>" which features around 600 tutorials, including a <a href="https://youtube.com/playlist?list=PLnH2pfPCPZsKJnAIPimrZaKwStQrLSNIQ">LangChain playlist</a>.</li>
</ul>

Links mentioned:


LlamaIndex ▷ #blog (5 messages):


LlamaIndex ▷ #general (19 messages🔥):

Links mentioned:

Usage Pattern - LlamaIndex 🦙 0.9.44: no description found


LlamaIndex ▷ #ai-discussion (7 messages):

Links mentioned:


Latent Space ▷ #ai-general-chat (29 messages🔥):

Links mentioned:


Latent Space ▷ #llm-paper-club-east (1 messages):

swyxio: check out the summary


Datasette - LLM (@SimonW) ▷ #ai (28 messages🔥):


DiscoResearch ▷ #general (8 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (1 messages):

flozi00: I am working on it at the moment to provide an german hosting service


Alignment Lab AI ▷ #general-chat (5 messages):

Links mentioned:

Open-Orca/SlimOrca · Datasets at Hugging Face: no description found


Skunkworks AI ▷ #off-topic (1 messages):

pradeep1148: https://www.youtube.com/watch?v=N5lDUZRI8sc


Skunkworks AI ▷ #bakklava-1 (1 messages):


LLM Perf Enthusiasts AI ▷ #reliability (1 messages):


LLM Perf Enthusiasts AI ▷ #prompting (1 messages):

Links mentioned:


AI Engineer Foundation ▷ #general (1 messages):

Links to the Google Slides presentations and the LinkedIn post are not fully provided, so they cannot be accurately shared.

Links mentioned: