Frozen AI News archive

Evals: The Next Generation

**Scale AI** highlighted issues with data contamination in benchmarks like **MMLU** and **GSM8K**, proposing a new benchmark where **Mistral** overfits and **Phi-3** performs well. **Reka** released the **VibeEval** benchmark for multimodal models addressing multiple choice benchmark limitations. **Sam Altman** of **OpenAI** discussed GPT-4 as "dumb" and hinted at **GPT-5** with AI agents as a major breakthrough. Researchers jailbroke **GPT-3.5** via fine-tuning. Global calls emerged to ban AI-powered weapons, with US officials urging human control over nuclear arms. Ukraine launched an AI consular avatar, while **Moderna** partnered with **OpenAI** for medical AI advancements. **Sanctuary AI** and **Microsoft** collaborate on AI for general-purpose robots. MIT introduced **Kolmogorov-Arnold networks** with improved neural network efficiency. **Meta AI** is training **Llama 3** models with over 400 billion parameters, featuring multimodality and longer context.

Canonical issue URL

The problem of data/benchmark contamination is often a passing joke but this year is reaching a breaking point with decreasing trust in the previous practice of self reported scores on well known academic benchmarks like MMLU and GSM8K. Scale AI released A Careful Examination of Large Language Model Performance on Grade School Arithmetic which proposed a new GSM8K-like benchmark that would be less contaminated, and plotted the deviations - Mistral seems to overfit notably on GSM8k, and Phi-3 does remarkably well:

image.png

Reka has also released a new VibeEval benchmark for multimodal models, their chosen specialty. They tackle the well known MMLU/MMMU issues with multiple choice benchmarks not being a good/stable measure for chat models.

Lastly we'll feature Jim Fan's thinking on the path forward for evals:

image.png


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Development and Capabilities

AI Regulation and Safety

AI Applications and Partnerships

AI Research and Advancements

Memes and Humor


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

LLMs in Space and Efficient Inference

Evaluating and Improving LLMs

Open Source Models and Frameworks

Emerging Models and Techniques

Industry Developments

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries

  1. Model Advancements and Fine-Tuning:

    • Increasing LoRA rank to 128 for Llama 3 to prioritize understanding over memorization, adding over 335M trainable parameters [Tweet]
    • Exploring multi-GPU support for model training with Unsloth, currently limited to single GPU [GitHub Wiki]
    • Releasing Llama-3 8B Instruct Gradient with RoPE theta adjustments for longer context handling [HuggingFace]
    • Introducing Hermes 2 Pro based on Llama-3 architecture, outperforming Llama-3 8B on benchmarks like AGIEval [HuggingFace]
  2. Hardware Optimization and Deployment:

    • Discussions on optimal GPU choices for LLMs, considering PCIe bandwidth, VRAM requirements (ideally 24GB+), and performance across multiple GPUs
    • Exploring local deployment options like RTX 4080 for smaller LLMs versus cloud solutions for privacy
    • Optimizing VRAM usage during training by techniques like merging datasets without increasing context length
    • Integrating DeepSpeed's ZeRO-3 with Flash Attention for efficient large model fine-tuning
  3. Multimodal AI and Computer Vision:

    • Introducing Motion-I2V for image-to-video generation with diffusion-based motion modeling [Paper]
    • Sharing resources on PyTorch Lightning integration with models like SegFormer, Detectron, YOLOv5/8 [Docs]
    • Accelerating diffusion models like Stable Diffusion XL by 3x using PyTorch 2 optimizations [Tutorial]
    • Unveiling Google's Med-Gemini multimodal models for medical applications [Video]
  4. Novel Neural Network Architectures:

    • Proposing Kolmogorov-Arnold Networks (KANs) as interpretable alternatives to MLPs [Paper]
    • Introducing Universal Physics Transformers for versatile simulations across datasets [Paper]
    • Exploring VisualFactChecker (VFC) for high-fidelity image/3D object captioning without training [Paper]
    • Sharing a binary vector representation approach for efficient unsupervised image patch encoding [Paper]
  5. Misc:


PART 1: High level Discord summaries

CUDA MODE Discord

CUDA Debugging Tips and Updates: Members exchanged insights on CUDA debugging, recommending resources such as a detailed Triton debugging lecture, and the importance of using the latest version of Triton, citing recent bug fixes in the interpreter.

CUDA Profiling Woes and Wisdom: Engineers grappled with inconsistent CUDA profiling results, suggesting the utilization of NVIDIA profiling tools like Nsight Compute/Systems over cudaEventRecord. A tinygrad patch for NVIDIA was shared, aiming to aid similar troubleshooting efforts.

Torch and PyTorch Prowess: Discussions mentioned the need for expertise in PyTorch internals, specifically ATen/linalg, while TorchInductor aficionados were pointed to a learning resource (though unspecified). A call went out to any PyTorch contributors for in-depth platform knowledge.

Advances in AI Model Training Constructs: Conversations in #llmdotc revealed a considerable volume of activity centered on model training. From FP32 master copy of params to CUDA Graphs, the talks included a range of technical challenges related to performance, precision, and complexity, coupled with links to various GitHub issues and pull requests for collaborative problem-solving.

Diving Deeper into Engineering Sparsity: Engineers mulled over the Effort Engine, debating its benchmark performances and the balance between speed and quality. Points of contemplation included parameter importance over precision, the quality trade-offs in weight pruning, and potential model improvements.

Forward-Thinking with AMD and Intel Tech: Enthusiasm was shown for AMD's HIP language with a tutorial playlist on the AMD ROCm platform, indicating a growing interest in diversified programming languages for GPUs. Additionally, a mention of Intel joining the PyTorch webpage suggested movement toward broader support across different architectures.


LM Studio Discord

CLI's New Frontier: The release of LM Studio 0.2.22 introduced a new command-line interface, lms, enabling functionalities such as loading/unloading LLMs and starting/stopping the local server, with development open for contributions on GitHub.

Tackling LLM Installation Chaos: Community discussions highlighted installation issues of LM Studio 0.2.22 Preview, which were surmounted by providing a corrected download link; meanwhile, users exchanged ideas on model performance improvements and quantization techniques, especially for the Llama 3 model.

Headless Operation Innovations: Members shared strategies for running LM Studio headlessly on systems without a graphical user interface, suggesting xvfb and other workarounds, creating a pathway for containerization possibilities like Docker.

ROCm and AMD Under the Lens: Conversations centered on the compatibility of different AMD GPUs with ROCm, alongside the challenges of ROCm's Linux support, highlighting the community's quest for efficient use of diverse hardware infrastructures.

Hardware Discourse Goes Deep: Discussions delved into the nitty-gritty of hardware choices, especially on suitable GPUs for running LLMs and the impact of PCIe 3.0 vs 4.0 on multi-GPU VRAM performance, culminating in a consensus that a minimum of 24GB VRAM is ideal for formidable models like Meta Llama 3 70B.


Stability.ai (Stable Diffusion) Discord


Unsloth AI (Daniel Han) Discord


Nous Research AI Discord


Perplexity AI Discord

Note: For the detailed and latest updates on API offerings and models like Sonar Large, check the official documentation.


Eleuther Discord


OpenRouter (Alex Atallah) Discord


OpenAI Discord


HuggingFace Discord


LlamaIndex Discord

LlamaIndex 0.3 Heralds Enhanced Interoperability: Version 0.3 of LlamaIndex.TS introduces Agent support for ReAct, Anthropic, and OpenAI, a generic AgentRunner class, standardized Web Streams, and a bolstered type system detailed in their release tweet. The update also outlines compatibility with React 19, Deno, and Node 22.

AI Engineers, RAG Tutorial Awaits: A new tutorial series on Retrieval-Augmented Generation (RAG) by @nerdai progresses from basics to managing long-context RAG, accompanied by a YouTube tutorial and a GitHub notebook.

Llamacpp Faces Parallel Dilemmas: In Llamacpp, concerns have been voiced about deadlocks while processing parallel queries, stemming from the lack of continuous batching support on a CPU server. Sequential request processing is seen as a potential workaround.

Word Loom Proposes Language Exchange Framework: The Word Loom specification is proposed for separating code from natural language, enhancing both composability and mechanical comparisons, with an aim to be globalization-friendly, as outlined in the Word Loom update proposition.

Strategies for Smarter AI Deployments: Discussions highlighted the sufficiency of the RTX 4080's 16 GB VRAM for smaller LLMs operations, while privacy concerns have some users shifting towards local computation stations over cloud alternatives like Google Colab for fine-tuning language models. Additionally, integrating external APIs with QueryPipeline and techniques for post-processing with a reranker to improve RAG application accuracy emerged as strategic considerations.


Modular (Mojo 🔥) Discord

Mojo's Anniversary Dominates Discussions: The Mojo Bot community commemorated its 1-year anniversary with speculations about a significant update release tomorrow. There were fond reflections on the progress Mojo made, particularly enhancements in traits, references, and lifetimes.

Modular Updates Celebrated: Community contributions have shaped the latest Mojo 24.3 release, leading to positive evaluations of its integration in platforms like Ubuntu 24.04. Concurrently, MAX 24.3 was announced, showcasing advancements in AI pipeline integration through the Engine Extensibility API, enhancing developer experiences in managing low-latency, high-throughput inferences as detailed in the MAX Graph APIs documentation.

CHERI's Potential Game-Changer for Security: The CHERI architecture is touted to significantly reduce vulnerability exploits by 70%, according to discussions referencing a YouTube video and the Colocation Tutorial. Talk of its adoption hinted at the possibility of transforming operating system development, empowering Unix-style software development, and potentially rendering conventional security methods obsolete.

Evolving Language Design and Performance: AI engineers continue to digest and deliberate on Mojo's language design objectives, aspiring to infer lifetimes and mutability akin to Hylo and debating the merit and safety of pointers over references. Community members leveraged Mojo's atomic operations for multi-core processing, achieving 100M records processing in 3.8 seconds.

Educational Content Spreads Mojo and MAX Awareness: Enthusiasm for learning and promotion of Mojo and MAX is evident with shared content like a video with Chris Lattner discussing Mojo, referenced as "Tomorrow's High Performance Python," and a PyCon Lithuania talk promoting Python's synergy with the MAX platform.


OpenInterpreter Discord

Bridging the Gap for AI Vtubing: Two AI Vtuber resources are now available, with one kit needing just a few credentials for setup on GitHub - nike-ChatVRM, as announced on Twitter. The other, providing an offline and uncensored experience, is shared along with a YouTube demo and source code on GitHub - VtuberAI.

Speed Boost for Whisper RKNN Users: A Git branch is now available that provides up to a 250% speed boost for Whisper RKNN on SBC with Rockchip RK3588, which can be accessed at GitHub - rbrisita/01 at rknn.

Ngrok Domain Customization Steps Outlined: Someone detailed a process for ngrok domain configuration, including editing tunnel.py and using a specific command line addition, with a helpful resource at ngrok Cloud Edge Domains.

Solving Independent Streaks in Ollama Bot: Trouble arose with Ollama, hinting at quirky autonomous behavior without waiting for user prompts, yet specific steps for resolution were not provided.

Eager for OpenInterpreter: There was speculation about the roll-out timeline for the OpenInterpreter app, the seamless inclusion of multimodal capabilities, and a sharing of community-driven assistance on various technical aspects. Solutions such as using the --os flag with GPT-4 for Windows OS mode compatibility, and a cooperative spirit were highlighted in the discussions.


Latent Space Discord


OpenAccess AI Collective (axolotl) Discord

Time to Mask Instruct Tags?: Engineers debated masking instruct tags during training to enhance ChatML performance, using a custom ChatML format, and considered the impact on model generation.

Llama-3 Leaps to 8B: Llama-3 8B Instruct Gradient is now available, featuring RoPE theta adjustments for improved context length handling, with discussions on its implementation and limitations at Llama-3 8B Gradient.

Axolotl Devs Patch Preprocessing Pain Points: A pull request was submitted to address a single-worker problem in the Orpo trainer and similarly in the TRL Trainer, allowing multithreading for speedier preprocessing, captured in PR #1583 on GitHub.

Python 3.10 Sets the Stage: A new baseline has been set within the community, where Python 3.10 is now the minimum version required for developing with Axolotl, enabling the use of latest language features.

Optimizing Training with ZeRO-3: Talks revolved around integrating DeepSpeed's ZeRO-3 and Flash Attention for finetuning to accelerate training, where ZeRO-3 optimizes memory without affecting quality, when appropriately deployed.


LAION Discord


AI Stack Devs (Yoko Li) Discord


LangChain AI Discord

Groq, No Wait Required: Direct sign-up to Groq's AI services is confirmed through a provided link to Groq's console, eliminating waitlist concerns for those eager to tap into Groq's capabilities.

AI's Script Deviation Head-Scratcher: Strategies to mitigate AI veering off script in human-AI interaction projects are sought after, highlighting the need for maintaining conversational flow without looping responses.

Adaptive RAG Gains Traction: A new Adaptive RAG technique, which selects optimal strategies based on query complexity, is discussed alongside a YouTube video explaining the approach.

LangChain Luminaries Launch Updates and Tools: An improved LangChain v0.1.17, Word Loom's open spec, deployment of Langserve on GCP, and Pydantic-powered tool definitions for GPT showcase the community's breadth of innovation with available resources on GitHub for Word Loom, a LangChain chatbot, and a Pydantic tools repository.

Feedback Loop Frustration in LangServe: A member's experience with LangServe's feedback feature highlights the importance of clear communication channels when submitting feedback, even after a successful submission response; changes may not be immediate or noticeable.


tinygrad (George Hotz) Discord

tinygrad Tackles Conda Conundrum: The tinygrad environment faced hitches on M1 Macs due to an AssertionError linked to an invalid Metal library, with potential fixes on the horizon, as well as a bounty posted for a solution to conda python issues after system updates, with progress reported recently.

From Podcast to Practice: One member's interest in tinygrad spiked after a Lex Fridman podcast, leading to recommendations to dive into the tinygrad documentation on GitHub for further understanding and comparing it with PyTorch.

Hardware Head-Scratcher for tinygrad Enthusiasts: A member deliberated over the choice between an AMD XT board and a new Mac M3 for their tinygrad development rig, highlighting the significance of choosing the right hardware for optimal development.

Resolving MNIST Mysteries with Source Intervention: An incorrect 100% MNIST accuracy alert prompted a member to ditch the pip version and successfully compile tinygrad from source, solving the version discrepancy and underscoring the approachability of tinygrad's build process.

CUDA Clarifications and Symbolic Scrutiny: Questions bubbled up about CUDA usage in scripts impacting performance, while another member pondered the differentiation between RedNode and OpNode, and the presence of blobfile was affirmed to be crucial for loading tokenizer BPE in tinygrad's LLaMA example code.


Mozilla AI Discord


Cohere Discord


Interconnects (Nathan Lambert) Discord


Alignment Lab AI Discord

Since there was only a single, non-technical message shared, which read "Hello," by user manojbh, there is no relevant technical discussion to summarize. Please provide messages that contain technical, detail-oriented content for a proper summary.


Datasette - LLM (@SimonW) Discord

Seeking a Language Model Janitor: Discussions highlighted the need for a Language Model capable of identifying and deleting numerous localmodels from a hard drive, underscoring a practical use case for AI in system maintenance and organization.


DiscoResearch Discord


AI21 Labs (Jamba) Discord

Jamba-Instruct Rolls Out: AI21 Labs announced the release of Jamba-Instruct, as per a tweet linked by a member. This could signal new developments in instruction-based AI models.


The Skunkworks AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

CUDA MODE ▷ #general (7 messages):

Link mentioned: CCCL - Google Drive: no description found


CUDA MODE ▷ #triton (11 messages🔥):

Link mentioned: Lecture 14: Practitioners Guide to Triton: https://github.com/cuda-mode/lectures/tree/main/lecture%2014


CUDA MODE ▷ #cuda (14 messages🔥):


CUDA MODE ▷ #torch (3 messages):


CUDA MODE ▷ #algorithms (11 messages🔥):


CUDA MODE ▷ #cool-links (2 messages):


CUDA MODE ▷ #beginner (2 messages):


CUDA MODE ▷ #youtube-recordings (1 messages):


CUDA MODE ▷ #torchao (1 messages):

Link mentioned: FP6 dtype! · Issue #208 · pytorch/ao: 🚀 The feature, motivation and pitch https://arxiv.org/abs/2401.14112 I think you guys are really going to like this. The deepspeed developers introduce FP6 datatype on cards without fp8 support, wh.....


CUDA MODE ▷ #off-topic (2 messages):

Link mentioned: Let's build the GPT Tokenizer: The Tokenizer is a necessary and pervasive component of Large Language Models (LLMs), where it translates between strings and tokens (text chunks). Tokenizer...


CUDA MODE ▷ #triton-puzzles (2 messages):


CUDA MODE ▷ #llmdotc (813 messages🔥🔥🔥):

<ul>
<li><strong>Master Params Mayhem</strong>: A recent merge enabling FP32 master copy of params by default disrupted expected model behavior, causing significant loss mismatches.</li>
<li><strong>Stochastic Rounding to the Rescue</strong>: Tests showed that incorporating stochastic rounding during parameter updates aligns results more closely with expected behavior.</li>
<li><strong>CUDA Concerns</strong>: Discussion raised around the substantial size and compilation time of cuDNN and possible optimizations for better usability within the llm.c project.</li>
<li><strong>CUDA Graphs Glow Dimly</strong>: CUDA Graphs, which improve kernel launch overhead, were briefly mentioned as a possible performance booster, but current GPU idle times imply limited benefits.</li>
<li><strong>Aiming for NASA Level C Code?🚀</strong>: Ideation around improving llm.c code to potentially meet safety-critical standards, with a side dream of LLMs in space and discussions on optimizing for more significant model sizes.</li>
</ul>

Links mentioned:


CUDA MODE ▷ #rocm (2 messages):

Link mentioned: AMD HIP Tutorial: In this series of videos, we will teach how to use the HIP programming language to program AMD GPUs running on the AMD ROCm platform. This set of videos is a...


CUDA MODE ▷ #oneapi (1 messages):

neurondeep: also added intel on pytorch webpage


LM Studio ▷ #💬-general (240 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (159 messages🔥🔥):

Links mentioned:


LM Studio ▷ #announcements (1 messages):

Link mentioned: GitHub - lmstudio-ai/lms: LM Studio in your terminal: LM Studio in your terminal. Contribute to lmstudio-ai/lms development by creating an account on GitHub.


LM Studio ▷ #🧠-feedback (4 messages):


LM Studio ▷ #⚙-configs-discussion (3 messages):


LM Studio ▷ #🎛-hardware-discussion (247 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (152 messages🔥🔥):

Links mentioned:


LM Studio ▷ #amd-rocm-tech-preview (25 messages🔥):

Links mentioned:


LM Studio ▷ #🛠-dev-chat (4 messages):

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (766 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (512 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (17 messages🔥):


Unsloth AI (Daniel Han) ▷ #help (139 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (35 messages🔥):


Unsloth AI (Daniel Han) ▷ #suggestions (11 messages🔥):

Links mentioned:


Nous Research AI ▷ #ctx-length-research (25 messages🔥):

Links mentioned:


Nous Research AI ▷ #off-topic (20 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (8 messages🔥):

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

<ul>
  <li><strong>Hermes 2 Pro Debuts on Llama-3 8B</strong>: Nous Research introduces <strong>Hermes 2 Pro</strong>, their first model based on Llama-3 architecture, available on HuggingFace. It outperforms Llama-3 8B Instruct on several benchmarks including AGIEval and TruthfulQA. <a href="https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B">Explore Hermes 2 Pro</a>.</li>
  <li><strong>New Capabilities & Structured Output</strong>: Hermes 2 Pro brings Function Calling and Structured Output capabilities, using dedicated tokens to simplify streaming function calls. The model also shows improvements in function calling evaluation and structured JSON output metrics.</li>
  <li><strong>Quantized Model Versions Are Available</strong>: For those interested in optimized models, GGUF quantized versions of Hermes 2 Pro can be accessed, providing a more efficient alternative. <a href="https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF">Check out the quantized version of Hermes 2 Pro</a>.</li>
  <li><strong>Team Effort Recognized</strong>: The development of Hermes Pro models is credited to the collaborative work of several contributors, alongside those customizing tools to support the models' unique prompt formatting needs.</li>
  <li><strong>Social Media Updates</strong>: You can follow along with Nous Research's updates and announcements regarding Hermes 2 Pro via their <a href="https://twitter.com/NousResearch/status/1785779313826308096">Twitter post</a>.</li>
</ul>

Links mentioned:


Nous Research AI ▷ #general (441 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (34 messages🔥):

Links mentioned:


Nous Research AI ▷ #bittensor-finetune-subnet (1 messages):


Nous Research AI ▷ #rag-dataset (16 messages🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (19 messages🔥):

Links mentioned:


Perplexity AI ▷ #general (303 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (17 messages🔥):


Perplexity AI ▷ #pplx-api (18 messages🔥):

Link mentioned: Supported Models: no description found


Eleuther ▷ #general (32 messages🔥):

Link mentioned: Efficient Representation of Natural Image Patches: Utilizing an abstract information processing model based on minimal yet realistic assumptions inspired by biological systems, we study how to achieve the early visual system's two ultimate objecti...


Eleuther ▷ #research (155 messages🔥🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (44 messages🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (1 messages):


OpenRouter (Alex Atallah) ▷ #announcements (3 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (2 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (186 messages🔥🔥):

Links mentioned:


OpenAI ▷ #ai-discussions (107 messages🔥🔥):

Link mentioned: GitHub - openai/simple-evals: Contribute to openai/simple-evals development by creating an account on GitHub.


OpenAI ▷ #gpt-4-discussions (9 messages🔥):


OpenAI ▷ #prompt-engineering (36 messages🔥):


OpenAI ▷ #api-discussions (36 messages🔥):


HuggingFace ▷ #general (153 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (5 messages):

Link mentioned: Med-Gemini: A High-Level Overview: A high-level overview on Med-Gemini, Google's "Family" (said in the voice of Vin Diesel) of Multimodal GenAI models for medicine. Med-Gemini has folks in the...


HuggingFace ▷ #cool-finds (5 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (8 messages🔥):

Links mentioned:


HuggingFace ▷ #computer-vision (6 messages):

Links mentioned:


HuggingFace ▷ #NLP (3 messages):

Link mentioned: Paper page - RARR: Researching and Revising What Language Models Say, Using Language Models: no description found


HuggingFace ▷ #diffusion-discussions (1 messages):

sayakpaul: Might be a better question for A1111 forums.


LlamaIndex ▷ #blog (4 messages):

Links mentioned:


LlamaIndex ▷ #general (112 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (6 messages):

Link mentioned: Word Loom proposed update: Word Loom proposed update. GitHub Gist: instantly share code, notes, and snippets.


Modular (Mojo 🔥) ▷ #general (8 messages🔥):


Modular (Mojo 🔥) ▷ #💬︱twitter (3 messages):


Modular (Mojo 🔥) ▷ #✍︱blog (2 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #announcements (1 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #ai (1 messages):


Modular (Mojo 🔥) ▷ #tech-news (2 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #🔥mojo (60 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-blogs-vids (4 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (2 messages):


Modular (Mojo 🔥) ▷ #nightly (2 messages):


OpenInterpreter ▷ #general (61 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (12 messages🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (2 messages):

Links mentioned:


Latent Space ▷ #ai-general-chat (33 messages🔥):

Links mentioned:


Latent Space ▷ #llm-paper-club-west (35 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (24 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (11 messages🔥):

Link mentioned: FIX: TRL trainer preprocessing step was running in one process by ali-mosavian · Pull Request #1583 · OpenAccess-AI-Collective/axolotl: Description We weren't passing dataset_num_proc to TRL training config, thus the initial data preprocessing steps in the TRL trainer was running in one process only. Motivation and Context Speeds ...


OpenAccess AI Collective (axolotl) ▷ #general-help (3 messages):


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (15 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (15 messages🔥):

Links mentioned:


LAION ▷ #general (27 messages🔥):

Links mentioned:


LAION ▷ #research (12 messages🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #app-showcase (28 messages🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #ai-town-discuss (3 messages):


AI Stack Devs (Yoko Li) ▷ #ai-town-dev (7 messages):

Link mentioned: WineHQ - Stellaris: no description found


LangChain AI ▷ #general (25 messages🔥):


LangChain AI ▷ #langserve (1 messages):


LangChain AI ▷ #share-your-work (6 messages):

Links mentioned:


LangChain AI ▷ #tutorials (1 messages):

Link mentioned: - YouTube: no description found


tinygrad (George Hotz) ▷ #general (20 messages🔥):


tinygrad (George Hotz) ▷ #learn-tinygrad (13 messages🔥):

Link mentioned: tinygrad: You like pytorch? You like micrograd? You love tinygrad! <3


Mozilla AI ▷ #llamafile (33 messages🔥):

Links mentioned:


Cohere ▷ #general (24 messages🔥):

Link mentioned: Preambles: no description found


Cohere ▷ #collab-opps (1 messages):


Interconnects (Nathan Lambert) ▷ #ml-questions (10 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (2 messages):

Link mentioned: Tweet from Teortaxes▶️ (@teortaxesTex): ...actually, why the hell am I assuming it's not their model, disseminated for collective pentesting - miqu-like oddly specific quant leak to preclude improvements - sudden 4chan link, throwaway...


Interconnects (Nathan Lambert) ▷ #random (11 messages🔥):


Interconnects (Nathan Lambert) ▷ #posts (2 messages):


Alignment Lab AI ▷ #general-chat (1 messages):

manojbh: Hello


Datasette - LLM (@SimonW) ▷ #llm (1 messages):


DiscoResearch ▷ #discolm_german (1 messages):

Link mentioned: LLaMA Pro: Progressive LLaMA with Block Expansion: Humans generally acquire new skills without compromising the old; however, the opposite holds for Large Language Models (LLMs), e.g., from LLaMA to CodeLLaMA. To this end, we propose a new post-pretra...


AI21 Labs (Jamba) ▷ #jamba (1 messages):

paulm24: Jamba-Instruct is out: https://twitter.com/AI21Labs/status/1786038528901542312