Frozen AI News archive

Mergestral, Meta MTIAv2, Cohere Rerank 3, Google Infini-Attention

**Meta** announced their new **MTIAv2 chips** designed for training and inference acceleration with improved architecture and integration with PyTorch 2.0. **Mistral** released the **8x22B Mixtral** model, which was merged back into a dense model to effectively create a 22B Mistral model. **Cohere** launched **Rerank 3**, a foundation model enhancing enterprise search and retrieval-augmented generation (RAG) systems supporting 100+ languages. **Google** published a paper on **Infini-attention**, an ultra-scalable linear attention mechanism demonstrated on 1B and 8B models with 1 million sequence length. Additionally, **Meta's Llama 3** is expected to start rolling out soon. Other notable updates include **Command R+**, an open model surpassing GPT-4 in chatbot performance with 128k context length, and advancements in Stable Diffusion models and RAG pipelines.

Canonical issue URL

Bunch of minor updates today, all worthwhile but nothing clearly The Story Of The Day:

All minor compared to Llama 3 which is slated to start rolling out next week.


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence. Comment crawling works now but has lots to improve!

New Models and Architectures

Stable Diffusion and Image Generation

Retrieval-Augmented Generation (RAG) and Context Handling

Open-Source Efforts and Local Deployment

Prompt Engineering and Fine-Tuning

Benchmarks, Comparisons, and Evaluations

Memes and Humor


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

LLM Developments

Efficient LLMs

Robotics and Embodied AI

Hardware and Systems

Miscellaneous


AI Discord Recap

A summary of Summaries of Summaries


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord


Unsloth AI (Daniel Han) Discord


LM Studio Discord

Bold Statements on LLM Performance: Users in LM Studio reported performance issues with the GPT "GPT Builder" and discussed optimal prompts, preferring manual writing to system-generated ones. Additionally, there was light on the new release in Mistral's lineage, the 8x22b model, which is pending GGUF quantizations to be operative in LM Studio.

Comparing Code Capabilities: Conversations around model capabilities for Python programming prioritized GPT-4, Claude3 Opus, and Phind, specifically calling out Phind's unique function to access the internet. The Release of Mixtral 8x22B has provoked talks due to its comparison with Command R+, noted for its advantage in low-level programming and math problems, formatting responses with LaTeX.

Pushing the Hardware Boundaries: Members exchanged knowledge on hardware suitability for AI models, mentioning successes and crashes with Codegemma and the impressiveness of the Max Studio (192 GB RAM) in running intense models. Discussions on cloud costs point towards alternatives like on-premises deployments using consumer-grade hardware for cost-effectiveness and AWS's recent removal of data egress fees.

Beta Releases Beckon Fixes and Features: Users in LM Studio noted the need for troubleshooting with the 0.2.19 beta, mentioning challenges like LaTeX rendering and interfacing with other tools like n8n. Specific troubles were highlighted with running models on AMC ROCm platform, with beta releases after 2.17 not functioning up to par.

Model Deployment Strategies Surface: Through the conversations, a narrative on model deployment arose, highlighting ways to optimize like considering cloud versus on-premises deployments and the practicality of local hardware augmentation, such as utilizing eGPUs and exploring cloud GPU services integration. There has been an appeal for a 12GB AI model suitable for hosting on 3080 GPUs for AutoGen tasks but no immediate solutions.


Nous Research AI Discord

Google Code Assistant on Board: Google's CodeGemma, a 7B model, aims to enhance developer productivity through advanced code completion, reflecting growing AI facilitation in programming.

The Return of World-Sim: World-Sim enthusiasts gear up for its relaunch, speculating on possible new features and applications, ranging from educational to AGI development, amid a flurry of teknium's cryptic messages.

Bridging AI Communication Gaps: The community discusses the benefits of bidirectional attention in LLMs, citing SD3's text rendering success and examining Infini-attention (research paper) as a way to efficiently handle long inputs in Transformer models.

Model Fine-Tuning Financial Feats: Engagements around fine-tuning large models like Nous-Hermes-8x22b expose cost issues, with QLoRA and LISA being evaluated against full-parameter fine-tuning, and cloud services like Vast offering expensive yet powerful GPU options.

Anticipated Model Developments Stir Excitement: With Meta announcing Llama 3's upcoming release (TechCrunch article), and MistralAI expected to release an instruct version of their Mixtral-8x22b, the community remains eager for new AI milestones.


Latent Space Discord


HuggingFace Discord


Perplexity AI Discord


CUDA MODE Discord


Eleuther Discord

Knowledge Scaling: A recent paper posits that language models max out at 2 bits of knowledge per parameter, igniting discussions on how various factors like training duration and model architecture might influence this limit. The community finds the implications non-trivial and is considering in-depth discussions to clarify the paper's insights.

RNNs Rise Again: Research indicates that interpretability tools developed for transformers are applicable to modern RNNs, showcasing effectiveness with Mamba and RWKV models. This revelation is backed by an accompanying paper and codebase, highlighting the resurgence of RNNs in language modeling with strong community collaboration in the study.

Fine-Tuning Finesse: A new technique, Subset fine-tuning (SubTuning), is making waves, enabling competitive performance by only adjusting a subset of layers, potentially easing computational demands for tasks like multi-task learning. The paper details this method, aligning with discussions prioritizing finetuning budget constraints.

Model Evaluation Expo: The Mixtral 8x22B model is turning heads with its AGIEval results, raising anticipation for its community release. Meanwhile, concerns about leveraging AI like deepfakes in election security cropped up alongside queries regarding downloading The Pile for research, emphasizing academic integrity.

Chat Templating Evolution: Pull requests for chat templating in the lm-evaluation-harness project are drawing attention, particularly Hailey's PR for HF models and another open PR. The community sees an opportunity to enhance the project by adding batchwise operations support for apply_chat_template.


Modular (Mojo 🔥) Discord

A New Phase for Mojo: Developments in the Mojo language include a roadmap highlight revealing future enhancements and prioritization of core programming features. Meanwhile, discussions have brewed on integrating Objective-C or AppKit for a new UI library aimed at MacOS, and the community has engaged in debates over GUI design patterns and error handling practices in Mojo, underscoring a vibrant ecosystem teetering on the cusp of substantial growth.

Advanced Storage Strategies Analyzed: A Modular blog post explores the impact of row-major and column-major memory arrangements on performance. It sheds light on the trade-offs developers face and the influence storage order has when working with languages and libraries like Mojo and NumPy.

Community Engagement and Contributions Rise: Open-source engagement has risen with significant contributions to projects like the Modular standard library and the Lightbug framework, which now boasts superior performance over Python's Flask. Mojo's lexical flexibility is on display with the addition of keyboard event handling in mojo-ui-html and the creation of lightbug_api, suggesting community-driven momentum is in full swing.

Innovating with Mojo in UI Development: The application of Mojo in UI development has been showcased through a sleek terminal text rendering tool inspired by lipgloss—available on GitHub—and the visual prowess of Basalt. These developments indicate a push towards elevating the aesthetic and functional capabilities of terminal applications using Mojo.

Modular Minds Stay Informed: The Modverse community is kept informed through sources like the "Modverse Weekly - Issue 29" newsletter, available at Modular Newsletters, and tweets providing bite-sized updates, all of which sustain the knowledge exchange within this technical hub.


OpenAccess AI Collective (axolotl) Discord

Quantum Leaping in Quantization: Discussions centered on the challenges of fitting models like Mistral onto a single 24GB card with 16k context when quantized, with testimonials validating Mixtral 8x7b performance.

Curiosity About MLLMs: Community members expressed curiosity about multimodal large models, such as LLaVA and Qwen VLM, yet faced limited resources for license navigation and fine-tuning guidance.

GPU Dilemmas for Inference Servers: Engineers debated the viability of Nvidia 4090s over 3090s for inference servers, considering the lack of NVLink and PCIe 5, suggesting that better inter-card bandwidth might make 3090s more suitable.

Hackathon Alert: The Samsung Next 2024 Generative AI Hackathon on May 11th was highlighted, emphasizing Health & Wellness and Mediatech sectors.

Diving Into Docs: The Axolotl community was encouraged to contribute to the evolving Axolotl documentation, with insights shared on dynamic programming optimization (DPO) potentially steering generated responses more effectively than supervised fine-tuning (SFT).


OpenRouter (Alex Atallah) Discord

Mixtral Joins the Router Fleet: Mixtral 8x22B has landed on OpenRouter, offering strong performance with instruction templates, and is currently available for a free trial.

Gemma's New Variant and Pricing Revisions: OpenRouter has replaced Gemma 7B with the upgraded Gemma 1.1-7B and adjusted pricing across several models—including LZLV 70B and Databricks DBRX 132B—while noting that Gemini 1.5 currently lacks a free tier.

Feedback Spurs Quick Fixes and Clarifications: User feedback prompted OpenRouter to correct issues with the "Updated" tag on models and deploy a fix for rate limit issues. The platform also clarified that tokens are counted as individual characters for Gemini models, affecting the "context" cost.

Diving Into Model Limitations: Heavily rate-limited models on OpenRouter are restrained to around 10 requests per minute, similar to the free tiers found elsewhere.

Community Weighs In on Mixtral vs. GPT-4: Comparisons between Mixtral 8x22b and GPT-4 in the community revealed a preference for Mixtral's reasoning capabilities and cost efficiency, although GPT-4 was viewed as more eloquent.


OpenInterpreter Discord


OpenAI Discord


LlamaIndex Discord


tinygrad (George Hotz) Discord


LAION Discord


LangChain AI Discord

Keep an Eye on Your Tokens: Engineers were tipped off about monitoring OpenAIAssistant token usage with tiktoken to multiply by pricing for cost estimation, perfect for those dense in economizing API calls.

Metadata Filters in Action: Vector databases harness metadata filters for precise queries, like finding companies with negative vacation leave policies. A member shed light on customizing retrievers for metadata inclusion, ensuring richer context in results.

Beta Features in the Spotlight: Questions about the with_structured_output in ChatOpenAI class uncovered that while not deprecated, it remains in beta. Code examples are fluttering around, and related tools like Instructor for Python are touted for structuring LLM outputs.

LangChain's Open-Source Compatibility Conundrum: LangChains’s architecture proudly supports various LLM providers, but members are seeking clear-cut examples for utilizing non-OpenAI LLMs, possibly to be found in the LangChain documentation.

Galaxy of New AI Tools Emerges: The dawning of apps like GPT AI with GPT-4 and Vision AI, Galaxy AI proffering free premium AI APIs, and the upgraded Appstorm v1.6.0 for intuitive app-building, demonstrates an expanding universe of AI tools ready at engineers' fingertips.


DiscoResearch Discord


Interconnects (Nathan Lambert) Discord


Datasette - LLM (@SimonW) Discord

Audio Intelligence Takes a Leap Forward: Gemini has enhanced its AI by gaining the ability to answer questions about audio in video content, addressing a previous gap where Gemini could only describe video visuals.

Google's Copy-Paste Plagued By Pasting Pains: Engineers are calling for an improvement in Google's text formatting capabilities when pasting text into their playground because it currently alters the original formatting.

Stanford Storms into Knowledge Curation: The Stanford Storm project presents a significant leap for AI in knowledge curation, with an LLM-powered system that researches topics and generates extensive reports complete with citations.

Shell Command Showdown on MacOS: A peculiar MacOS iTerm2 issue causing the llm cmd to hang turned out to be a need for user input, remedied by a fix provided on GitHub, which ensures the command no longer hangs and correctly responds to input.

Homebrew or Pipx: LLM Shells Still Stump Users: Troubleshooting llm cmd issues on different shells, one user discovered the problem wasn't the highly customized shell itself but rather the interaction required by the command, not seen in logs.


Mozilla AI Discord

Bridge the Gap with Gradio UI for Figma: Mozilla introduces Gradio UI for Figma to facilitate fast prototyping and experimentation for design phases; it's accessible through Figma's page for Mozilla. For deeper discussions, Mozilla encourages joining the thread at their Discord discussion channel.

GPU Constraints Make Waves: Engineers tackled GPU memory limitations by using -ngl 3 to offload some layers to CPU memory, though acknowledging a significant performance cost, and proposed the development of a feature in llamafile that dynamically offloads layers to manage VRAM inadequacies.

Kernel Conversations Can Crash: Engaging with tensors might lead to a kernel panic, as evidenced by an M2 MacBook freezing when converting .safetensors to .gguf due to a likely overload of its 16GB RAM capacity.

A Lesson in Language Model Memory Management: Discussion included referencing the ollama project on GitHub, which details methods for handling large language models, a potential guide to enhancing llamafile's memory handling capabilities. Visit ollama's GitHub page for more details.

Boost Text Predictions with Quiet-STaR: Interest flared around Quiet-STaR, a technique that has language models providing rationales at each token to refine text predictions; resources shared include the research paper and the GitHub repository, along with a related Hugging Face repository.


Skunkworks AI Discord


LLM Perf Enthusiasts AI Discord

GPT's Coding Game Still Strong: User experiences debunk sequel slumps - GPT maintains its robust coding capabilities when employed through cursor, providing fast performance and comprehensive code output.

Cursor vs. Claude: The Tool Time Talk: While cursor is favored for generating boilerplate code due to its efficient command-K feature powered by GPT-4, some users still prefer Claude opus for chat interactions, despite a reported incident of Claude hallucinating code for the first time.

Gemini 1.5 Rises: Members of the guild are buzzing over Gemini 1.5 with positive remarks about its coding capabilities, albeit without diving into specifics.

Copilot++ Takes Off: The introduction of Copilot++ has been met with acclaim for its top-tier performance in coding tasks, standing out even within a field of advanced tools.

Claude's Rare Slip Up: For the first time, a user reported an unexpected case where Claude crafted a piece of code out of thin air, straying from the conventionally accurate performance observed in GPT-4.


AI21 Labs (Jamba) Discord


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ▷ #general-chat (691 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (276 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (7 messages):


Unsloth AI (Daniel Han) ▷ #help (244 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (15 messages🔥):

The community showed enthusiasm and appreciation for the shared projects, highlighting their innovative contributions to the field of AI.

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (6 messages):

Link mentioned: AdaLomo: Low-memory Optimization with Adaptive Learning Rate: Large language models have achieved remarkable success, but their extensive parameter size necessitates substantial memory for training, thereby setting a high threshold. While the recently proposed l...


LM Studio ▷ #💬-general (183 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (197 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (1 messages):

sanjuhs123: this is awesome, then i just have to download the beta 0.2.19 or wait till it releases,


LM Studio ▷ #📝-prompts-discussion-chat (2 messages):


LM Studio ▷ #🎛-hardware-discussion (40 messages🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (52 messages🔥):


LM Studio ▷ #autogen (2 messages):


LM Studio ▷ #langchain (1 messages):


LM Studio ▷ #amd-rocm-tech-preview (31 messages🔥):

Links mentioned:


LM Studio ▷ #crew-ai (3 messages):


Nous Research AI ▷ #off-topic (8 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (8 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (278 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (10 messages🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (109 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-general-chat (74 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (6 messages):

Links mentioned:


Latent Space ▷ #llm-paper-club-west (294 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #announcements (8 messages🔥):

Links mentioned:


HuggingFace ▷ #general (258 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (1 messages):

Link mentioned: The Root of AI Security is Rootless. On Podman: For GenAI Microservices: An overview on installing @Podman from the terminal. #Podman #containers #AI #genAI #Docker #Linux #EdTech #deeplearning #microservices


HuggingFace ▷ #cool-finds (7 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (11 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (6 messages):


HuggingFace ▷ #core-announcements (1 messages):

Link mentioned: Distributed inference with multiple GPUs: no description found


HuggingFace ▷ #computer-vision (14 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (10 messages🔥):


HuggingFace ▷ #diffusion-discussions (8 messages🔥):

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):

Link mentioned: One interface, many LLMs - Raycast Blog: Raycast AI gets more powerful with Anthropic Claude 3, Perplexity, and more models - making it the perfect UI for AI


Perplexity AI ▷ #general (266 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (14 messages🔥):


Perplexity AI ▷ #pplx-api (16 messages🔥):

Link mentioned: Discover Typeform, where forms = fun: Create a beautiful, interactive form in minutes with no code. Get started for free.


CUDA MODE ▷ #general (1 messages):

Link mentioned: no title found: no description found


CUDA MODE ▷ #triton (1 messages):

mobicham: https://github.com/BobMcDear/attorch


CUDA MODE ▷ #cuda (63 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #torch (3 messages):

Link mentioned: No factory functions for strided quantized tensors · Issue #74540 · pytorch/pytorch: 🐛 Describe the bug For non-quantized tensors, there is both empty and empty_strided. However, for quantized tensors there are only empty variants for functions. This means that it is difficult for .....


CUDA MODE ▷ #beginner (7 messages):

Link mentioned: Join the PMPP UI lectures timezones Discord Server!: Check out the PMPP UI lectures timezones community on Discord - hang out with 4 other members and enjoy free voice and text chat.


CUDA MODE ▷ #ring-attention (3 messages):


CUDA MODE ▷ #off-topic (3 messages):


CUDA MODE ▷ #hqq (76 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #triton-viz (3 messages):


CUDA MODE ▷ #llmdotc (67 messages🔥🔥):

Links mentioned:


Eleuther ▷ #announcements (1 messages):

Links mentioned:


Eleuther ▷ #general (83 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (132 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (4 messages):

Link mentioned: Physics of Language Models: Part 3.3, Knowledge Capacity Scaling Laws: Scaling laws describe the relationship between the size of language models and their capabilities. Unlike prior studies that evaluate a model's capability via loss or benchmarks, we estimate the n...


Eleuther ▷ #interpretability-general (2 messages):


Eleuther ▷ #lm-thunderdome (2 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #general (103 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (2 messages):


Modular (Mojo 🔥) ▷ #✍︱blog (1 messages):

Link mentioned: Modular: Row-major vs. column-major matrices: a performance analysis in Mojo and NumPy: We are building a next-generation AI developer platform for the world. Check out our latest post: Row-major vs. column-major matrices: a performance analysis in Mojo and NumPy


Modular (Mojo 🔥) ▷ #🔥mojo (58 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (5 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #📰︱newsletter (1 messages):

Zapier: Modverse Weekly - Issue 29 https://www.modular.com/newsletters/modverse-weekly-29


OpenAccess AI Collective (axolotl) ▷ #general (131 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (5 messages):

Link mentioned: Introduction | Continuum Training Platform | Axolotl Training Platform: no description found


OpenAccess AI Collective (axolotl) ▷ #general-help (3 messages):


OpenAccess AI Collective (axolotl) ▷ #datasets (2 messages):


OpenAccess AI Collective (axolotl) ▷ #rlhf (1 messages):

Link mentioned: Introduction | Continuum Training Platform | Axolotl Training Platform: no description found


OpenAccess AI Collective (axolotl) ▷ #minotaur (1 messages):

Link mentioned: Introduction | Continuum Training Platform | Axolotl Training Platform: no description found


OpenAccess AI Collective (axolotl) ▷ #bots (1 messages):

Link mentioned: Introduction | Continuum Training Platform | Axolotl Training Platform: no description found


OpenAccess AI Collective (axolotl) ▷ #community-showcase (3 messages):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (5 messages):

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


OpenRouter (Alex Atallah) ▷ #announcements (10 messages🔥):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (136 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #general (97 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (30 messages🔥):

Link mentioned: 01 Light Setup - Flash and Connect to Server: Quick video to show how to flash the 01 Light and get it set up on a server running 01OS locally with OpenAI models.I'm following the instructions at h ttps:...


OpenInterpreter ▷ #ai-content (1 messages):

Link mentioned: GitHub - xenova/transformers.js: State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!: State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server! - xenova/transformers.js


OpenAI ▷ #ai-discussions (67 messages🔥🔥):

Link mentioned: OpenAI Status: no description found


OpenAI ▷ #gpt-4-discussions (29 messages🔥):

Link mentioned: OpenAI Status: no description found


OpenAI ▷ #prompt-engineering (14 messages🔥):


OpenAI ▷ #api-discussions (14 messages🔥):


LlamaIndex ▷ #blog (3 messages):


LlamaIndex ▷ #general (89 messages🔥🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #general (68 messages🔥🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (23 messages🔥):


LAION ▷ #general (78 messages🔥🔥):

Links mentioned:


LAION ▷ #research (7 messages):

Link mentioned: Reddit - Dive into anything: no description found


LAION ▷ #learning-ml (1 messages):


LangChain AI ▷ #general (60 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #langserve (5 messages):


LangChain AI ▷ #langchain-templates (1 messages):

lhc1921: https://python.langchain.com/docs/integrations/llms/azure_openai/


LangChain AI ▷ #share-your-work (3 messages):

Links mentioned:


LangChain AI ▷ #tutorials (5 messages):

Link mentioned: Hot dog or not with Ollama, Mistral and LLava: In this tutorial we take a look at whether an image is hot dog or not using Ollama, mistral and lava#python #pythonprogramming #llm #ml #ai #aritificialintel...


DiscoResearch ▷ #mixtral_implementation (10 messages🔥):

Links mentioned:


DiscoResearch ▷ #general (22 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (41 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (23 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (7 messages):

Link mentioned: Stanford CS224N | 2023 | Lecture 10 - Prompting, Reinforcement Learning from Human Feedback: For more information about Stanford's Artificial Intelligence professional and graduate programs visit: https://stanford.io/aiTo learn more about this course...


Interconnects (Nathan Lambert) ▷ #ml-drama (5 messages):


Interconnects (Nathan Lambert) ▷ #random (14 messages🔥):


Datasette - LLM (@SimonW) ▷ #ai (4 messages):

Link mentioned: GitHub - stanford-oval/storm: An LLM-powered knolwedge curation system that researches a topic and generates a full-length report with citations.: An LLM-powered knolwedge curation system that researches a topic and generates a full-length report with citations. - stanford-oval/storm


Datasette - LLM (@SimonW) ▷ #llm (12 messages🔥):

Links mentioned:


Mozilla AI ▷ #announcements (1 messages):

Link mentioned: Figma (@futureatmozilla) | Figma: The latest files and plugins from Mozilla Innovation Projects (@futureatmozilla) — We're building products that focus on creating a more personal, private and open-source internet


Mozilla AI ▷ #llamafile (11 messages🔥):

Links mentioned:


Skunkworks AI ▷ #general (1 messages):

Link mentioned: Tweet from Jan P. Harries (@jphme): @MistralAI first AGIEval results look great 👇 - thanks for releasing this beast, guys! 👏 https://x.com/jphme/status/1778028110954295486 ↘️ Quoting Jan P. Harries (@jphme) First AGIEval results fo...


Skunkworks AI ▷ #datasets (5 messages):

Links mentioned:


Skunkworks AI ▷ #off-topic (2 messages):

Links mentioned:


LLM Perf Enthusiasts AI ▷ #gpt4 (6 messages):


LLM Perf Enthusiasts AI ▷ #claude (2 messages):


AI21 Labs (Jamba) ▷ #jamba (2 messages):