Frozen AI News archive

1/9/2024: Nous Research lands $5m for Open Source AI

**Nous Research** announced a **$5.2 million seed financing** focused on **Nous-Forge**, aiming to embed transformer architecture into chips for powerful servers supporting real-time voice agents and **trillion parameter models**. **Rabbit R1** launched a demo at CES with mixed reactions. **OpenAI** shipped the **GPT store** and briefly leaked an upcoming personalization feature. A new paper on **Activation Beacon** proposes a solution to extend LLMs' context window significantly, with code to be released on GitHub. Discussions also covered **QLORA**, **fine-tuning**, **synthetic data**, and **custom architectures** for LLMs.

Canonical issue URL

Nous announced their seed, and the business focus is Nous Forge: image.png

Rabbit R1 also launched their demo at CES and opinions were very divided.

In other news, OpenAI shipped the GPT store today, and briefly leaked their upcoming personalization feature.

image.png

--

Table of Contents

[TOC]

Nous Research AI Discord Summary

Nous Research AI Channel Summaries

▷ #ctx-length-research (3 messages):

Links mentioned:

▷ #off-topic (95 messages🔥🔥):

Links mentioned:

▷ #interesting-links (39 messages🔥):

Links mentioned:

▷ #announcements (1 messages):

Links mentioned:

Etched | The World's First Transformer Supercomputer: Transformers etched into silicon. By burning the transformer architecture into our chips, we're creating the world's most powerful servers for transformer inference.

▷ #general (377 messages🔥🔥):

Links mentioned:

▷ #ask-about-llms (44 messages🔥):

Links mentioned:

GitHub - SciPhi-AI/synthesizer: A multi-purpose LLM framework for RAG and data creation.: A multi-purpose LLM framework for RAG and data creation. - GitHub - SciPhi-AI/synthesizer: A multi-purpose LLM framework for RAG and data creation.

▷ #project-obsidian (4 messages):


OpenAI Discord Summary

OpenAI Channel Summaries

▷ #ai-discussions (160 messages🔥🔥):

Links mentioned:

▷ #gpt-4-discussions (154 messages🔥🔥):

Links mentioned:

Brand guidelines: Language and assets for using the OpenAI brand in your marketing and communications.

▷ #prompt-engineering (6 messages):

▷ #api-discussions (6 messages):


LM Studio Discord Summary

LM Studio Channel Summaries

▷ #💬-general (71 messages🔥🔥):

Links mentioned:

▷ #🤖-models-discussion-chat (18 messages🔥):

Links mentioned:

GitHub - stanfordnlp/dspy: Stanford DSPy: The framework for programming—not prompting—foundation models: Stanford DSPy: The framework for programming—not prompting—foundation models - GitHub - stanfordnlp/dspy: Stanford DSPy: The framework for programming—not prompting—foundation models

▷ #🧠-feedback (10 messages🔥):

▷ #🎛-hardware-discussion (29 messages🔥):

Links mentioned:

▷ #autogen (1 messages):

thelefthandofurza: Has anyone used autogen studio with lmstudio?


Eleuther Discord Summary

Eleuther Channel Summaries

▷ #general (61 messages🔥🔥):

Links mentioned:

▷ #research (25 messages🔥):

Links mentioned:

AI agents help explain other AI systems: FIND (function interpretation and description) is a new technique for evaluating automated interpretability methods. Developed at MIT, the system uses artificial intelligence to automate the explanati...

▷ #interpretability-general (8 messages🔥):

Links mentioned:

▷ #gpt-neox-dev (13 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) Discord Summary

OpenAccess AI Collective (axolotl) Channel Summaries

▷ #general (7 messages):

Links mentioned:

Tweet from Stas Bekman (@StasBekman): Heads up to Accelerate/Deepspeed multi-gpu users There was a regression in accelerate==0.23 (deepspeed integration) which would make your training much slower. The fix has just been merged - so you c...

▷ #axolotl-dev (37 messages🔥):

Links mentioned:

▷ #other-llms (1 messages):

leoandlibe: I use the exllamav2 convert.py to make EXL2 quants 😄

▷ #general-help (10 messages🔥):

▷ #rlhf (3 messages):

▷ #community-showcase (1 messages):

▷ #shearedmistral (8 messages🔥):

Links mentioned:


HuggingFace Discord Discord Summary

HuggingFace Discord Channel Summaries

▷ #announcements (1 messages):

Links mentioned:

▷ #general (22 messages🔥):

Links mentioned:

▷ #today-im-learning (7 messages):

Links mentioned:

Welcome to the 🤗 Deep Reinforcement Learning Course - Hugging Face Deep RL Course

▷ #cool-finds (3 messages):

Links mentioned:

▷ #i-made-this (3 messages):

Links mentioned:

▷ #reading-group (5 messages):

Links mentioned:

The Variational Inference Book: A comprehensive review of variational inference in one concise book.

▷ #diffusion-discussions (3 messages):

Links mentioned:

Goodbye cold boot - how we made LoRA Inference 300% faster

▷ #computer-vision (2 messages):

Links mentioned:

skew_detection: Explore and run machine learning code with Kaggle Notebooks | Using data from rdocuments

▷ #NLP (18 messages🔥):

Links mentioned:

▷ #diffusion-discussions (3 messages):

Links mentioned:

Goodbye cold boot - how we made LoRA Inference 300% faster


Perplexity AI Discord Summary

Perplexity AI Channel Summaries

▷ #general (51 messages🔥):

Links mentioned:

▷ #sharing (8 messages🔥):

▷ #pplx-api (6 messages):


LAION Discord Summary

LAION Channel Summaries

▷ #general (44 messages🔥):

Links mentioned:

Phase1 Collect Underpants GIF - Phase1 Collect Underpants Gnome - Discover & Share GIFs: Click to view the GIF

▷ #research (18 messages🔥):

Links mentioned:


Mistral Discord Summary

Mistral Channel Summaries

▷ #general (13 messages🔥):

▷ #deployment (9 messages🔥):

Links mentioned:

▷ #finetuning (13 messages🔥):

Links mentioned:

▷ #announcements (1 messages):

sophiamyang: New paper on Mixtral of Experts: https://arxiv.org/pdf/2401.04088.pdf

▷ #showcase (3 messages):

Links mentioned:

GitHub - vanna-ai/vanna: 🤖 Chat with your SQL database 📊. Accurate Text-to-SQL Generation via LLMs using RAG 🔄.: 🤖 Chat with your SQL database 📊. Accurate Text-to-SQL Generation via LLMs using RAG 🔄. - GitHub - vanna-ai/vanna: 🤖 Chat with your SQL database 📊. Accurate Text-to-SQL Generation via LLMs using R...

▷ #la-plateforme (7 messages):


LangChain AI Discord Summary

LangChain AI Channel Summaries

▷ #general (28 messages🔥):

Links mentioned:

▷ #langserve (6 messages):

Links mentioned:

langserve/examples/passthrough_dict/server.py at main · langchain-ai/langserve: LangServe 🦜️🏓. Contribute to langchain-ai/langserve development by creating an account on GitHub.

▷ #share-your-work (3 messages):

Links mentioned:

Langchain with FastAPI in Docker with Traefik [Code Included]: A tutorial on how to use langchain with FastAPI.h

▷ #tutorials (1 messages):


LlamaIndex Discord Discord Summary

LlamaIndex Discord Channel Summaries

▷ #blog (3 messages):

Links mentioned:

undefined

▷ #general (22 messages🔥):

▷ #ai-discussion (9 messages🔥):

Links mentioned:


DiscoResearch Discord Summary

DiscoResearch Channel Summaries

▷ #mixtral_implementation (1 messages):

sebastian.bodza: Paper for mixtral is released: https://arxiv.org/abs/2401.04088

▷ #embedding_dev (16 messages🔥):

Links mentioned:


Latent Space Discord Summary

Latent Space Channel Summaries

▷ #ai-general-chat (11 messages🔥):

Links mentioned:

▷ #llm-paper-club (5 messages):

Links mentioned:


Skunkworks AI Discord Summary

Skunkworks AI Channel Summaries

▷ #general (7 messages):

▷ #core-moe (4 messages):


LLM Perf Enthusiasts AI Discord Summary

LLM Perf Enthusiasts AI Channel Summaries

▷ #general (5 messages):

▷ #rag (1 messages):

jeffreyw128: we rely on detecting if there's bad text from non-OCR methods

▷ #datasets (2 messages):


Datasette - LLM (@SimonW) Discord Summary

Only 1 channel had activity, so no need to summarize...


Alignment Lab AI Discord Summary

Alignment Lab AI Channel Summaries

▷ #ai-and-ml-discussion (1 messages):

burnydelic: https://news.mit.edu/2024/ai-agents-help-explain-other-ai-systems-0103

▷ #general-chat (2 messages):

▷ #oo (1 messages):

teknium: https://fxtwitter.com/NousResearch/status/1744865872563618128


YAIG (a16z Infra) Discord Summary

Only 1 channel had activity, so no need to summarize...

Links mentioned:

DiLoCo: Distributed Low-Communication Training of Language Models: Large language models (LLM) have become a critical component in many applications of machine learning. However, standard approaches to training LLM require a large number of tightly interconnected acc...