Frozen AI News archive

Quis promptum ipso promptiet?

**Anthropic** released upgrades to their Workbench Console, introducing new prompt engineering features like chain-of-thought reasoning and prompt generators that significantly reduce development time, exemplified by their customer **Zoominfo**. **OpenAI** teased a "magic" new development coming soon, speculated to be a new LLM replacing GPT-3.5 in the free tier or a search competitor. The open-source community highlighted **Llama 3 70B** as "game changing" with new quantized weights for **Llama 3 120B** and CUDA graph support for **llama.cpp** improving GPU performance. **Neuralink** demonstrated a thought-controlled mouse, sparking interest in modeling consciousness from brain signals. The **ICLR 2024** conference is being held in Asia for the first time, generating excitement.

Canonical issue URL

AI News for 5/9/2024-5/10/2024. We checked 7 subreddits, 373 Twitters and 28 Discords (419 channels, and 4923 messages) for you. Estimated reading time saved (at 200wpm): 556 minutes.

We have been fans of Anthropic's Workbench for a while, and today they released some upgrades helping people improve and templatize their prompts.

image.png

Pretty cool, not really the end of prompt engineer but nice to have. Let's be honest, it's been a really quiet week before the storm of both OpenAI's big demo day (potentially a voice assistant?) and Google I/O next week.


Table of Contents

[TOC]


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

OpenAI Announcements

Anthropic Developments

Llama and Open-Source Models

Neuralink Demo

ICLR Conference

Miscellaneous

Memes and Humor


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Progress and Capabilities

AI Ethics and Governance

AI Models and Architectures


AI Discord Recap

A summary of Summaries of Summaries

  1. Large Language Model (LLM) Advancements and Releases:

    • Meta's Llama 3 model is generating excitement, with an upcoming hackathon hosted by Meta offering a $10K+ prize pool. Discussions revolve around fine-tuning, evaluation, and the model's performance.
    • LLaVA-NeXT models promise enhanced multimodal capabilities for image and video understanding, with local testing encouraged.
    • The release of Gemma, boasting a 10M context window and requiring less than 32GB memory, sparks interest and skepticism regarding output quality.
    • Multimodal Model Developments: Several new multimodal AI models were announced, including Idefics2 with a fine-tuning demo (YouTube), LLaVA-NeXT (blog post) with expanded image and video understanding capabilities, and the Lumina-T2X family (Reddit post) for transforming noise into various modalities based on text prompts. The Scaling_on_scales (GitHub) approach challenged the necessity of larger vision models.
  2. Optimizing LLM Inference and Training:

    • Innovations like vAttention and QServe aim to improve GPU memory efficiency and quantization for LLM inference, enabling larger batch sizes and faster serving.
    • Consistency Large Language Models (CLLMs) introduce parallel decoding to reduce inference latency, mimicking human cognitive processes.
    • Discussions on optimizing CUDA kernels, Triton performance, and the trade-offs between determinism and speed in backward passes for LLM training.
    • Vrushank Desai's series explores optimizing inference latency for diffusion models by leveraging GPU architecture intricacies.
  3. AI Model Interpretability and Evaluation:

    • The Inspect AI framework from the UK AI Safety Institute offers components for evaluating LLMs, including prompt engineering, tool usage, and multi-turn dialog.
    • Eleuther AI discusses the CrossCare project, which analyzes disease prevalence bias across demographics in LLMs and pretraining data.
    • Debates around the impact of pretraining datasets on "zero-shot" generalization of multimodal models, as detailed in an arXiv paper.
    • The Mirage multi-level tensor algebra superoptimizer aims to optimize deep neural networks, though its benchmark claims face skepticism.
  4. Open-Source AI Tools and Libraries:

    • LlamaIndex announces local LLM integration, TypeScript agent building guides, and integration with Google Firestore, fostering open AI development.
    • OpenInterpreter enables AI task automation using GPT-4 and OpenCV, with new releases adding OS flag and Computer API support.
    • Hugging Face integrates B-LoRA training into advanced DreamBooth for implicit style-content separation using a single image.
    • Intel's ipex-llm accelerates local LLM inference and fine-tuning on Intel CPUs and GPUs, though it currently lacks LM Studio support.

PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord

Artisan Bot Immerses in Discord: Stability AI launched Stable Artisan, a Discord bot boasting Stable Diffusion 3 and Stable Video Diffusion features for content creation, bolstered by tools like Search and Replace, Background Removal, and Outpainting to revolutionize user interactions directly on Discord.

Open-Source or Not? The SD3 Debate Rages: Discord members heatedly debated the potential open-sourcing of Stable Diffusion 3 (SD3), exploring motives for the current API-restricted access and speculating on future outcome scenarios, including possible refinement before release.

Exploring the Stable Diffusion Universe: The community engaged with various Stable Diffusion model versions, including SDXL and ControlNets, evaluating their limitations and the substantial enhancements brought forth by community-developed models like Lora.

Aspiring for 360-Degree Creation: A user sparked discussion on crafting 360-degree images, sharing multiple resources and seeking guidance on methodologies, referencing platforms like Skybox AI and discussions on Reddit.

Tech Support to the Rescue in Real Time: Practical and succinct exchanges provided quick resolutions to common execution errors, such as "DLL load failed while importing bz2", emphasizing the Discord community's agility in offering peer-to-peer technical support.


Perplexity AI Discord

Perplexity Partners with SoundHound: Perplexity AI has entered a partnership with SoundHound, with the aim to integrate online large language models (LLMs) into voice assistants across various devices, enhancing real-time web search capabilities.

Perplexity Innovates Search and Citations: An update on Perplexity AI introduces incognito search, ensuring that user inquiries vanish after 24 hours, combined with enhanced citation previews to bolster user trust in information sources.

Pro Search Glitch and Opus Limitations Spark Debate: The engineering community is facing challenges with the Pro Search feature, which currently fails to deliver internet search results or source citations. Additionally, dissatisfaction surfaced regarding the daily 50-use limit for the Opus model on Perplexity AI, sparking discussions for potential alternatives and solutions.

API Conundrum for AI Engineers: Engineers have noted issues with API output consistency, where the same prompts yield different results compared to those on Perplexity Labs, despite using identical models. Queries have been raised regarding the cause of the discrepancies and requests for guidance on effective prompting for the latest models.

Engagement with Perplexity's Features and New Launches: Users are engaging with features such as making threads shareable and exploring various inquiries including the radioactivity of bananas and the nature of mathematical rings. Additionally, there's interest in Natron Energy's latest launch, reported through Perplexity's sharing platform.


Unsloth AI (Daniel Han) Discord

Unsloth Studio Stalls for Philanthropy: Unsloth Studio's release is postponed due to the team focusing on releasing phi and llama projects, with about half of the studio's project currently complete.

Optimizer Confusion Cleared: Users were uncertain about how to specify optimizers in Unsloth but referenced the Hugging Face documentation for clarification on valid strings for optimizers, including "adamw_8bit".

Training Trumps Inference: The Unsloth team has stated a preference for advancing training techniques rather than inference, where the competition is fierce. They've touted progress in accelerating training in their open-source contributions.

Long Context Model Skepticism: Discussions point to scepticism among users regarding the feasibility and evaluation of very long context models, such as a mentioned effort to tackle up to a 10M context length.

Dataset Cost-Benefit Debated: The community has exchanged differing views on the investment needed for high-quality datasets for model training, considering both instruct tuning and synthetic data creation.

Market-First Advice for Aspiring Bloggers: A member's idea for a multi-feature blogging platform prompted advice on conducting market research and ensuring a clear customer base to avoid a lack of product/market fit.

Ghost 3B Beta Tackles Time and Space: Early training of the Ghost 3B Beta model demonstrates its ability to explain Einstein's theory of relativity in lay terms across various languages, hinting at its potential for complex scientific communication.

Help Forums Foster Fine-Tuning Finesse: The Unsloth AI help channel is buzzing with tips for fine-tuning AI models on Google Colab, though multi-GPU support is a wanted yet unavailable feature. Solutions for CUDA memory errors and a nod towards YouTube fine-tuning tutorials are shared among users.

Customer Support AI at Your Service: ReplyCaddy, a tool based on a fine-tuned Twitter dataset and a tiny llama model for customer support, was showcased, with acknowledgments to Unsloth AI for fast inference assistance, found on hf.co.


LM Studio Discord

LM Studio Laments Library Limitations: While LM Studio excels with models like Llama 3 70B, users struggle to run models such as llama1.6 Mistral or Vicuña even on a 192GB Mac Studio, pointing to a mysterious RAM capacity issue despite ample system resources. There's also discomfort among users concerning the LM Studio installer on Windows since it doesn't offer installation directory selection.

AI Models Demand Hefty Hardware: Running large models necessitates substantial VRAM; members discussed VRAM being a bigger constraint than RAM. Intel's ipex-llm library was introduced to accelerate local LLM inference on Intel CPUs and GPUs Intel Analytics Github, but it's not yet compatible with LM Studio.

New Frontier of Multi-Device Collaboration: Engineers explored the challenges and potential for integrating AMD and Nvidia hardware, addressing the theoretical possibility versus the practical complexity. The fading projects like ZLUDA, aimed at broadening CUDA support for non-Nvidia hardware, were lamented ZLUDA Github.

Translation Model Exchange: For translation projects, Meta AI's NLLB-200, SeamlessM4T, and M2M-100 models came highly recommended, elevating the search for efficient multilingual capabilities.

CrewAI's Cryptic Cut-Off: When faced with truncated token outputs from CrewAI, users deduced that it wasn't quantization to blame. A mishap in the OpenAI API import amid conditional statements was the culprit, a snag now untangled, reaffirming the devil's in the details.


HuggingFace Discord

Graph Learning Enters LLM Territory: The Hugging Face Reading Group explored the integration of Graph Machine Learning with LLMs, fueled by Isamu Isozaki's insights, complete with a supportive write-up and a video.

Demystifying AI Creativity: B-LoRA's integration into advanced DreamBooth's LoRA training script promises new creative heights just by adding the flag --use_blora and training for a relatively short span, as per the diffusers GitHub script and findings in the research paper.

On the Hunt for Resources: AI enthusiasts sought guidance and shared resources across a variety of tasks, with a notable GitHub repository on creating PowerPoint slides using OpenAI's API and DALL-E available at Creating slides with Assistants API and DALL-E and the mention of Ankush Singal's Medium articles for table extraction tools.

Challenging NLP Channel Conversations: The NLP channel tackled diverse topics such as recommending models for specific languages—indicating a preference for sentence transformers and encoder models, instructing versions of Llama, and also referenced community involvement in interview preparations.

Hiccups and Fixes in Diffusion Discussions: The diffusion discussions detailed issues and potential solutions related to HuggingChat bot errors and color shifts in diffusion models, noting a possible fix for login issues by switching the login module from lixiwu to anton-l in order to troubleshoot a 401 status code error.


Modular (Mojo 🔥) Discord


Nous Research AI Discord


OpenAI Discord


Eleuther Discord


CUDA MODE Discord

AI Hype Train Hits Practical Station: The community is buzzing with discussions on the practical aspects of deep learning optimization, contrasting with the usual hype around AI capabilities. Specific areas of focus include saving and loading compiled models in PyTorch, acceleration of compiled artifacts, and the non-support of MPS backend in Torch Inductor as illustrated in a PR by msaroufim.

Memory Efficiency Breakthroughs: Innovations like vAttention and QServe are reshaping GPU memory efficiency and serving optimizations for large language models (LLMs), promising larger batch sizes without internal fragmentation and efficient new quantization algorithms.

Engineering Precision: CUDA vs Triton: Critical comparisons between CUDA and Triton for warp and thread management, including performance nuances and kernel-launch overheads, were dissected. A YouTube lecture on the topic was recommended, with discussions pointing out the pros and cons of using Triton, notably its attempt at minimizing Python-related overhead through potential C++ runtimes.

Optimization Odyssey: Links shared revealed a fascination with optimizing inference latency for models like Toyota's diffusion model, discussed in Vrushank Desai's series found here, and a "superoptimizer" explored in the Mirage paper for DNNs, raising eyebrows regarding benchmark claims and the lack of autotune.

CUDA Conundrums and Determinism Dilemmas: From troubleshooting CUDA's device-side asserts to setting the correct NVCC compiler flags, beginners are wrestling with the nuances of GPU computing. Meanwhile, seasoned developers are debating determinism in backward passes and the trade-offs with performance in LLM training, as discussed in the llmdotc channel.


Latent Space Discord

New Kid on the Block Outshines Olmo: A model from 01.ai is claimed to vastly outperform Olmo, stirring up interest and debate within the community about its potential and real-world performance.

Sloppy Business: Borrowing from Simon Willison's terminology, community members adopt "slop" to describe unwanted AI-generated content. Here's the buzz about AI etiquette.

LLM-UI Cleans Up Your Markup Act: llm-ui was introduced as a solution for refining Large Language Model (LLM) outputs by addressing problematic markdown, adding custom components, and enhancing pauses with a smoother output.

Meta Llama 3 Hackathon Gears Up: An upcoming hackathon focused on Llama 3 has been announced, with Meta at the helm and a $10K+ prize pool, looking to excite AI enthusiasts and developers. Details and RSVP here.

AI Guardrails and Token Talk: Discussions revolved around LLM guardrails featuring tools like Outlines.dev, and the concept of token restriction pregeneration, an approach ill-suited for API-controlled models like those from OpenAI.


LAION Discord


OpenInterpreter Discord

Groq API Joins OpenInterpreter's Toolset: The Groq API is now being used within OpenInterpreter, with the best practice being to use groq/ as the prefix in completion requests and define the GROQ_API_KEY. Python integration examples are available, aiding in rapid deployment of Groq models.

OpenInterpreter Empowers Automation with GPT-4: OpenInterpreter demonstrates successful task automation, specifically using GPT-4 alongside OpenCV/Pyautogui for GUI navigation tasks on Ubuntu systems.

Innovative OpenInterpreter and Hardware Mashups: Community members are creatively integrating OpenInterpreter with Billiant Labs Frame to craft unique applications such as AI glasses, as shown in this demo, and are exploring compatible hardware like the ESP32-S3-BOX 3 for the O1 Light.

Performance Variability in Local LLMs: While OpenInterpreter's tools are actively used, members have observed inconsistent performance in local LLMs for file system tasks, with Mixtral recognized for enhanced outcomes.

Updates and Advances in LLM Landscapes: The unveiling of LLaVA-NeXT models marks progress in local image and video understanding. Concurrently, OpenInterpreter's 0.2.5 release has brought in new features like the --os flag and a Computer API, detailed in the change log, improving inclusiveness and empowering developers with better tools.


LlamaIndex Discord

LLM Integration for All: LlamaIndex announced a new feature allowing local LLM integrations, supporting models like Mistral, Gemma, and others, and shared details on Twitter.

TypeScript and Local LLMs Unite: There's an open-source guide for building TypeScript agents that leverage local LLMs, like Mixtral, announced on Twitter.

Top-k RAG Approach Discouraged for 2024: A caution against using top-k RAG for future projects trended, hinting at emerging standards in the community. LlamaIndex tweeted this guidance here.

Graph Database Woes and Wonders: A user detailed their method of turning Gmail content into a Graph database via a custom retriever but is now looking for ways to improve efficiency and data feature extraction.

Interactive Troubleshooting: When facing a NotImplementedError with Mistral and HuggingFace, users were directed to a Colab notebook to facilitate the setup of a ReAct agent.


OpenAccess AI Collective (axolotl) Discord

Pretraining Predicaments and Fine-Tuning Frustrations: Engineers reported challenges and sought advice on pretraining optimization, dealing with a dual Epoch Enigma where one epoch unexpectedly saved a model twice, and facing a Pickle Pickle with PyTorch, which threw a TypeError when it couldn’t pickle a 'torch._C._distributed_c10d.ProcessGroup' object.

LoRA Snafu Sorted: A fix was proposed for a LoRA Configuration issue, advising to include 'embed_tokens' and 'lm_head' in the settings to address a ValueError; this snippet was shared for precise YAML configuration:

lora_modules_to_save:
  - lm_head
  - embed_tokens

Additionally, an engineer struggling with an AttributeError in transformers/trainer.py was counseled on debugging steps, including batch inspection and data structure logging.

Scaling Saga: For fine-tuning extended contexts in Llama 3 models, linear scaling was recommended, while dynamic scaling was suggested as the better option for scenarios outside fine-tuning.

Bot Troubles Teleported: A Telegram bot user highlighted a timeout error, suggesting networking or API rate-limiting issues could be in play, with the error message being: 'Connection aborted.', TimeoutError('The write operation timed out').

Axolotl Artefacts Ahead: Discussion on the Axolotl platform revealed capabilities to fine-tune Llama 3 models, confirming a possibility for handling a 262k sequence length and further curious explorations into fine-tuning a 32k dataset.


Interconnects (Nathan Lambert) Discord


LangChain AI Discord

Structuring AI with a Google Twist: Discussion unfolded around emulating Google's Gemini AI studio structured prompts within various LLMS, introducing function calling as a new approach to managing LangChain's model interactions.

Navigating LangGraph and Vector Databases: Users troubleshoot issues with ToolNode in LangGraph, with pointers to the LangGraph documentation for in-depth guidance; while others deliberated the complexities and costs of vector databases, with some favoring pgvector for its simplicity and open-source availability, and a comprehensive comparison guide recommended for those considering free local options.

Python vs POST Discrepancies in Chain Invocation: A peculiar case emerged where a member experienced different outcomes when employing a chain() through Python compared to utilizing an /invoke endpoint, indicating the chain began with an empty dictionary via the latter method, signaling potential variations in LangChain's initialization procedures.

AgencyRuntime Beckons Collaborators: A prominent article introduces AgencyRuntime, a social platform designed for crafting modular teams of generative AI agents, and extends an invitation to enhance its capabilities by incorporating further LangChain features.

Learning from LangChain Experts: Tutorial content released guides users on linking crewAI with the Binance Crypto Market and contrasts Function Calling agents and ReACt agents within LangChain, offering practical insights for those fine-tuning their AI applications.


OpenRouter (Alex Atallah) Discord


Cohere Discord


Datasette - LLM (@SimonW) Discord


Mozilla AI Discord


LLM Perf Enthusiasts AI Discord


tinygrad (George Hotz) Discord

Metal Build Conundrum: Despite extensive research including Metal-cpp API, MSL spec, Apple documentation, and a developer reference, a user is struggling to understand the libraryDataContents() function in a Metal build process.

Tensor Vision: An online visualizer has been developed by a user to help others comprehend tensor shapes and strides, potentially simplifying learning for AI engineers.

TinyGrad Performance Metrics: Clarification was provided that InterpretedFlopCounters in TinyGrad's ops.py are used as performance proxies through flop count insights.

Buffer Registration Clarified: Responding to an inquiry about self.register_buffer in TinyGrad, a user mentioned that initializing a Tensor with requires_grad=False is the alternative approach in TinyGrad.

Symbolic Range Challenge: There's a call for a more symbolic approach to functions and control flow within TinyGrad, hinting at an ambition to have a rendering system that comprehends expansions of general algebraic lambdas and control statements symbolically.


Alignment Lab AI Discord


Skunkworks AI Discord


AI Stack Devs (Yoko Li) Discord


The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ▷ #announcements (1 messages):

Link mentioned: Stable Artisan: Media Generation and Editing on Discord — Stability AI: One of the most frequent requests from the Stable Diffusion community is the ability to use our models directly on Discord. Today, we are excited to introduce Stable Artisan, a user-friendly bot for m...


Stability.ai (Stable Diffusion) ▷ #general-chat (877 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #announcements (2 messages):

Link mentioned: SoundHound AI and Perplexity Partner to Bring Online LLMs to Next Gen Voice Assistants Across Cars and IoT Devices: This marks a new chapter for generative AI, proving that the powerful technology can still deliver optimal results in the absence of cloud connectivity. SoundHound’s work with NVIDIA will allow it to ...


Perplexity AI ▷ #general (715 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (15 messages🔥):


Perplexity AI ▷ #pplx-api (9 messages🔥):


Unsloth AI (Daniel Han) ▷ #general (476 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (19 messages🔥):

Link mentioned: Reddit - Dive into anything: no description found


Unsloth AI (Daniel Han) ▷ #help (113 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (10 messages🔥):

Link mentioned: Reply Caddy - a Hugging Face Space by jed-tiotuico: no description found


LM Studio ▷ #💬-general (150 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (99 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (8 messages🔥):


LM Studio ▷ #📝-prompts-discussion-chat (1 messages):


LM Studio ▷ #🎛-hardware-discussion (46 messages🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (4 messages):


LM Studio ▷ #memgpt (6 messages):


LM Studio ▷ #amd-rocm-tech-preview (2 messages):


LM Studio ▷ #crew-ai (12 messages🔥):


LM Studio ▷ #🛠-dev-chat (4 messages):


HuggingFace ▷ #announcements (6 messages):

Links mentioned:


HuggingFace ▷ #general (247 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (2 messages):

Link mentioned: Enhancing Q-Learning with Large Language Model Heuristics: Q-learning excels in learning from feedback within sequential decision-making tasks but requires extensive sampling for significant improvements. Although reward shaping is a powerful technique for en...


HuggingFace ▷ #cool-finds (4 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (8 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (29 messages🔥):

Links mentioned:


HuggingFace ▷ #core-announcements (1 messages):

<ul>
  <li><strong>B-LoRA is now diffusing creativity</strong>: <a href="https://github.com/huggingface/diffusers/blob/main/examples/advanced_diffusion_training/train_dreambooth_lora_sdxl_advanced.py">B-LoRA training</a> is integrated into the advanced DreamBooth LoRA training script. Users simply need to add a <code>'--use_blora'</code> flag to their config and train for 1000 steps to harness its capabilities.</li>
  <li><strong>Understanding B-LoRA</strong>: The <a href="https://huggingface.co/papers/2403.14572">B-LoRA paper</a> highlights key insights, including the fact that two unet blocks are essential for encoding content and style, and illustrating how B-LoRA can achieve implicit style-content separation using just one image.</li>
</ul>

Links mentioned:


HuggingFace ▷ #computer-vision (12 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (10 messages🔥):


HuggingFace ▷ #diffusion-discussions (7 messages):


Modular (Mojo 🔥) ▷ #general (18 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (3 messages):


Modular (Mojo 🔥) ▷ #ai (1 messages):

pepper555: what project?


Modular (Mojo 🔥) ▷ #🔥mojo (135 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (5 messages):


Modular (Mojo 🔥) ▷ #📰︱newsletter (1 messages):

Zapier: Modverse Weekly - Issue 33 https://www.modular.com/newsletters/modverse-weekly-33


Modular (Mojo 🔥) ▷ #🏎engine (10 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #nightly (132 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #off-topic (6 messages):

Links mentioned:


Nous Research AI ▷ #interesting-links (1 messages):

deoxykev: https://hao-ai-lab.github.io/blogs/cllm/


Nous Research AI ▷ #general (174 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (6 messages):

Link mentioned: NousResearch/Meta-Llama-3-8B · Hugging Face: no description found


Nous Research AI ▷ #bittensor-finetune-subnet (1 messages):


Nous Research AI ▷ #world-sim (48 messages🔥):

Link mentioned: worldsim: no description found


OpenAI ▷ #annnouncements (1 messages):


OpenAI ▷ #ai-discussions (144 messages🔥🔥):


OpenAI ▷ #gpt-4-discussions (34 messages🔥):


OpenAI ▷ #prompt-engineering (2 messages):


OpenAI ▷ #api-discussions (2 messages):


Eleuther ▷ #general (29 messages🔥):

Links mentioned:


Eleuther ▷ #research (129 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (10 messages🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (3 messages):


Eleuther ▷ #lm-thunderdome (1 messages):

Link mentioned: Inspect: Open-source framework for large language model evaluations


Eleuther ▷ #gpt-neox-dev (8 messages🔥):

Links mentioned:


CUDA MODE ▷ #general (2 messages):


CUDA MODE ▷ #triton (19 messages🔥):

Links mentioned:


CUDA MODE ▷ #torch (19 messages🔥):

Links mentioned:


CUDA MODE ▷ #algorithms (3 messages):

Links mentioned:


CUDA MODE ▷ #cool-links (7 messages):

Link mentioned: Diffusion Inference Optimization: no description found


CUDA MODE ▷ #beginner (6 messages):


CUDA MODE ▷ #llmdotc (69 messages🔥🔥):

Link mentioned: Run CUDNN in LLM.c Walkthrough: All of the commands used in this video in sequence:ssh [email protected] clone https://github.com/karpathy/llm.c.gitsudo apt updatesudo apt install pyt...


Latent Space ▷ #ai-general-chat (25 messages🔥):

Links mentioned:


Latent Space ▷ #ai-in-action-club (71 messages🔥🔥):

Links mentioned:


LAION ▷ #general (8 messages🔥):

Link mentioned: SNAC with flattening & reconstruction: Speech only codec:https://colab.research.google.com/drive/11qUfQLdH8JBKwkZIJ3KWUsBKtZAiSnhm?usp=sharingGeneral purpose (32khz) Codec:https://colab.research.g...


LAION ▷ #research (79 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #general (64 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (21 messages🔥):

Link mentioned: LLaVA-NeXT: Stronger LLMs Supercharge Multimodal Capabilities in the Wild: LLaVA-NeXT: Stronger LLMs Supercharge Multimodal Capabilities in the Wild


LlamaIndex ▷ #blog (5 messages):


LlamaIndex ▷ #general (68 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):


OpenAccess AI Collective (axolotl) ▷ #general (6 messages):


OpenAccess AI Collective (axolotl) ▷ #general-help (9 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (21 messages🔥):

lora_modules_to_save:
  - lm_head
  - embed_tokens

to the configuration file as a solution.

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ideas-and-feedback (3 messages):


Interconnects (Nathan Lambert) ▷ #news (3 messages):


Interconnects (Nathan Lambert) ▷ #ml-questions (6 messages):

Link mentioned: EasyLM/EasyLM/models/llama/llama_train_rm.py at main · hamishivi/EasyLM: Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax. - hamishivi/EasyLM


Interconnects (Nathan Lambert) ▷ #ml-drama (6 messages):

Link mentioned: Leaked Deck Reveals OpenAI's Pitch on Publisher Partnerships: no description found


Interconnects (Nathan Lambert) ▷ #random (14 messages🔥):

Links mentioned:


LangChain AI ▷ #general (25 messages🔥):

Link mentioned: Picking a vector database: a comparison and guide for 2023: no description found


LangChain AI ▷ #langserve (1 messages):


LangChain AI ▷ #share-your-work (1 messages):

Links mentioned:


LangChain AI ▷ #tutorials (2 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (3 messages):

<ul>
  <li>
    <strong>Launch of Languify.ai:</strong> A new browser extension called
    <a href="https://www.languify.ai/">Languify.ai</a> was launched to help optimize website text to increase user engagement and sales. The extension utilizes Openrouter to interact with different models based on the user's prompts.
  </li>
  <li>
    <strong>AnythingLLM User Seeks Simplicity:</strong> A member expressed interest in the newly introduced Languify.ai as an alternative to AnythingLLM which they found to be overkill for their needs.
  </li>
  <li>
    <strong>Beta Testers Wanted for Rubik's AI:</strong> An invitation was extended for beta testing an advanced research assistant and search engine, offering a 2-month free premium trial of features like GPT-4 Turbo, Claude 3 Opus, Mistral Large, among others. Interested individuals are encouraged to provide feedback and can sign up through <a href="https://rubiks.ai/">Rubik's AI</a> with the promo code <code>RUBIX</code>.
  </li>
</ul>

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (25 messages🔥):

Links mentioned:


Cohere ▷ #general (18 messages🔥):

Links mentioned:


Cohere ▷ #announcements (9 messages🔥):

Link mentioned: Introducing Command R Fine-Tuning: Industry-Leading Performance at a Fraction of the Cost: Command R fine-tuning offers superior performance on enterprise use cases and costs up to 15x less than the largest models on the market.


Datasette - LLM (@SimonW) ▷ #ai (18 messages🔥):

Links mentioned:


Mozilla AI ▷ #llamafile (16 messages🔥):

Links mentioned:


LLM Perf Enthusiasts AI ▷ #general (3 messages):

Link mentioned: Tweet from Jan : Spreadsheets are the lifeblood of many biology labs, but extracting insights from messy data is a huge challenge. We wanted to see if AI could help us reliably pull data from any arbitrary spreadsheet...


LLM Perf Enthusiasts AI ▷ #gpt4 (10 messages🔥):


tinygrad (George Hotz) ▷ #general (1 messages):

helplesness: Hello


tinygrad (George Hotz) ▷ #learn-tinygrad (10 messages🔥):

Link mentioned: Shape & Stride Visualizer: no description found


Alignment Lab AI ▷ #general-chat (1 messages):


Skunkworks AI ▷ #off-topic (1 messages):

pradeep1148: https://www.youtube.com/watch?v=4MzCpZLEQJs


AI Stack Devs (Yoko Li) ▷ #events (1 messages):