Frozen AI News archive

Gemini Ultra is out, to mixed reviews

**Google** released **Gemini Ultra** as a paid tier for "Gemini Advanced with Ultra 1.0" following the discontinuation of Bard. Reviews noted it is "slightly faster/better than ChatGPT" but with reasoning gaps. The **Steam Deck** was highlighted as a surprising AI workstation capable of running models like Solar 10.7B. Discussions in AI communities covered topics such as multi-GPU support for OSS Unsloth, training data contamination from OpenAI outputs, ethical concerns over model merging, and new alignment techniques like Listwise Preference Optimization (LiPO). The **Mojo** programming language was praised for high-performance computing. In research, the **Subformer** model uses sandwich-style parameter sharing and SAFE for efficiency, and **BiLLM** introduced 1-bit post-training quantization to reduce resource use. The **OpenHermes** dataset viewer tool was launched, and GPU scheduling with Slurm was discussed. Fine-tuning challenges for models like **OpenHermes-2.5-Mistral-7B** and VRAM requirements were also topics of interest.

Canonical issue URL

Business as usual, it's been pretty quiet overall in AI. With Bard well and truly dead, Gemini Ultra was released today as a paid tier for "Gemini Advanced with Ultra 1.0". The reviews industrial complex is getting to work:

image.png

image.png


Table of Contents

[TOC]

PART 1: High level Discord summaries

TheBloke Discord Summary


Nous Research AI Discord Summary


LM Studio Discord Summary


Latent Space Discord Summary


Mistral Discord Summary


HuggingFace Discord Summary


OpenAI Discord Summary


Perplexity AI Discord Summary


LlamaIndex Discord Summary


OpenAccess AI Collective (axolotl) Discord Summary


LAION Discord Summary


CUDA MODE Discord Summary


LangChain AI Discord Summary


LLM Perf Enthusiasts AI Discord Summary


Datasette - LLM (@SimonW) Discord Summary


DiscoResearch Discord Summary


PART 2: Detailed by-Channel summaries and links

TheBloke ▷ #general (1295 messages🔥🔥🔥):

Links mentioned:


TheBloke ▷ #characters-roleplay-stories (342 messages🔥🔥):

Links mentioned:


TheBloke ▷ #training-and-fine-tuning (5 messages):

Links mentioned:

LiPO: Listwise Preference Optimization through Learning-to-Rank: no description found


TheBloke ▷ #coding (24 messages🔥):

Links mentioned:


Nous Research AI ▷ #off-topic (2 messages):


Nous Research AI ▷ #interesting-links (43 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (221 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (49 messages🔥):

Links mentioned:


LM Studio ▷ #💬-general (138 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (45 messages🔥):

Links mentioned:


LM Studio ▷ #🎛-hardware-discussion (98 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (10 messages🔥):

Links mentioned:

Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


LM Studio ▷ #autogen (3 messages):


LM Studio ▷ #avx-beta (1 messages):


LM Studio ▷ #crew-ai (1 messages):


LM Studio ▷ #open-interpreter (13 messages🔥):

Links mentioned:

GitHub - KillianLucas/open-interpreter: A natural language interface for computers: A natural language interface for computers. Contribute to KillianLucas/open-interpreter development by creating an account on GitHub.


Latent Space ▷ #ai-general-chat (12 messages🔥):

Links mentioned:

LoRA Fine-tuning Efficiently Undoes Safety Training from Llama 2-Chat 70B — LessWrong: Produced as part of the SERI ML Alignment Theory Scholars Program - Summer 2023 Cohort, under the mentorship of Jeffrey Ladish.  …


Latent Space ▷ #ai-announcements (1 messages):

Links mentioned:

LLM Paper Club (West) · Luma: We have moved to use the new Discord Stage feature here: https://discord.com/channels/822583790773862470/1197350122112168006 see you soon!


Latent Space ▷ #llm-paper-club-west (209 messages🔥🔥):

Links mentioned:


Mistral ▷ #general (128 messages🔥🔥):

Links mentioned:

Chat with Open Large Language Models: no description found


Mistral ▷ #models (4 messages):

Links mentioned:

intfloat/e5-mistral-7b-instruct · Hugging Face: no description found


Mistral ▷ #showcase (69 messages🔥🔥):

Links mentioned:


Mistral ▷ #random (2 messages):


Mistral ▷ #la-plateforme (1 messages):

Links mentioned:

databricks/databricks-dolly-15k · Datasets at Hugging Face: no description found


HuggingFace ▷ #announcements (1 messages):


HuggingFace ▷ #general (115 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (2 messages):


HuggingFace ▷ #cool-finds (4 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (9 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (4 messages):

Links mentioned:

mamba/mamba_ssm at main · state-spaces/mamba: Contribute to state-spaces/mamba development by creating an account on GitHub.


HuggingFace ▷ #diffusion-discussions (1 messages):


HuggingFace ▷ #computer-vision (4 messages):

Links mentioned:

GitHub - VikParuchuri/surya: Accurate line-level text detection and recognition (OCR) in any language: Accurate line-level text detection and recognition (OCR) in any language - GitHub - VikParuchuri/surya: Accurate line-level text detection and recognition (OCR) in any language


HuggingFace ▷ #NLP (11 messages🔥):

Links mentioned:

How to detect bad data in your instruction tuning dataset (for better LLM fine-tuning): Overview of automated tools for catching: low-quality responses, incomplete/vague prompts, and other problematic text (toxic language, PII, informal writing, bad grammar/spelling) lurking in a instru...


HuggingFace ▷ #diffusion-discussions (1 messages):


OpenAI ▷ #ai-discussions (79 messages🔥🔥):


OpenAI ▷ #gpt-4-discussions (30 messages🔥):


OpenAI ▷ #prompt-engineering (8 messages🔥):

Links mentioned:

Chat model feedback: no description found


OpenAI ▷ #api-discussions (8 messages🔥):

Links mentioned:

Chat model feedback: no description found


Perplexity AI ▷ #general (31 messages🔥):

Links mentioned:


Perplexity AI ▷ #sharing (5 messages):

Links mentioned:

Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Perplexity AI ▷ #pplx-api (43 messages🔥):

Links mentioned:


LlamaIndex ▷ #blog (4 messages):

Links mentioned:

Setting up Query Pipeline For Advanced RAG Workflow using LlamaIndex: What is QueryPipelines?


LlamaIndex ▷ #general (65 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):


OpenAccess AI Collective (axolotl) ▷ #general (12 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (24 messages🔥):

Links mentioned:

Scheduler implementation of Continual Pre-Training of Large Language Models: How to (re)warm your model? by jinwonkim93 · Pull Request #1273 · OpenAccess-AI-Collective/axolotl: Scheduler implementation of Continual Pre-Training of Large Language Models: How to (re)warm your model? (https://arxiv.org/pdf/2308.04014.pdf) Description almost identical to consine min lr but i...


OpenAccess AI Collective (axolotl) ▷ #general-help (10 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #rlhf (12 messages🔥):

Links mentioned:

DreamGenTrain/dpo.py at master · DreamGenX/DreamGenTrain: Contribute to DreamGenX/DreamGenTrain development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #runpod-help (2 messages):


LAION ▷ #general (22 messages🔥):

Links mentioned:

OpenAI is adding new watermarks to DALL-E 3: The watermarks can still be erased, however.


LAION ▷ #research (27 messages🔥):

Links mentioned:


CUDA MODE ▷ #general (20 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (6 messages):

Links mentioned:

PyTorch 2 paper and tutorial @ ASPLOS 2024: The PyTorch team is excited to share that our paper on PyTorch 2 has been accepted for presentation at the ACM International Conference on Architectural Support for Programming Languages and Operating...


CUDA MODE ▷ #torch (2 messages):


CUDA MODE ▷ #beginner (6 messages):


CUDA MODE ▷ #jax (5 messages):

Links mentioned:

Day 1 Talks: JAX, Flax & Transformers 🤗: Day 1 Talks: JAX, Flax & Transformers 🤗0:00:00 Skye Wanderman-Milne (Google Brain): Intro to JAX on Cloud TPUs0:42:49 Marc van Zee (Google Brain): Introduct...


LangChain AI ▷ #general (13 messages🔥):


LangChain AI ▷ #langserve (6 messages):

Links mentioned:


LangChain AI ▷ #langchain-templates (1 messages):


LangChain AI ▷ #share-your-work (4 messages):

Links mentioned:


LangChain AI ▷ #tutorials (7 messages):


LLM Perf Enthusiasts AI ▷ #general (3 messages):


LLM Perf Enthusiasts AI ▷ #finetuning (1 messages):


LLM Perf Enthusiasts AI ▷ #offtopic (3 messages):


LLM Perf Enthusiasts AI ▷ #speed (7 messages):


LLM Perf Enthusiasts AI ▷ #cost (2 messages):


Datasette - LLM (@SimonW) ▷ #ai (7 messages):

Links mentioned:

Drew Breunig (@[email protected]): Attached: 1 video Emoji suggest is now working for StepList. Some quiet AI UX, it works by generating an embedding for your new list's title and comparing it to an approved database of emoji embe...


DiscoResearch ▷ #general (3 messages):