Frozen AI News archive

Mamba-2: State Space Duality

**Mamba-2**, a new **state space model (SSM)**, outperforms previous models like Mamba and Transformer++ in **perplexity** and **wall-clock time**, featuring **8x larger states** and **50% faster training**. It introduces the concept of **state space duality (SSD)** connecting SSMs and linear attention. The **FineWeb-Edu dataset**, a high-quality subset of the **15 trillion token FineWeb dataset**, filtered using **llama-3-70b** for educational quality, enables better and faster LLM learning, potentially reducing tokens needed to surpass **GPT-3** performance. Additionally, perplexity-based data pruning using a **125M parameter model** improves downstream performance and reduces pretraining steps by up to **1.45x**. The **Video-MME benchmark** evaluates multi-modal LLMs on video analysis across multiple visual domains and video lengths.

Canonical issue URL

AI News for 5/31/2024-6/3/2024. We checked 7 subreddits, 384 Twitters and 29 Discords (400 channels, and 8575 messages) for you. Estimated reading time saved (at 200wpm): 877 minutes.

Over the weekend we got the FineWeb Technical Report (which we covered a month ago), and it turns out that it does improve upon CommonCrawl and RefinedWeb with better filtering and deduplication.

However we give the weekend W to the Mamba coauthors, who are somehow back again with Mamba-2, a core 30 lines of Pytorch which outperforms Mamba and Transformer++ in both perplexity and wall-clock time.

image.png

Tri recommends reading the blog first, developing Mamba-2 over 4 parts:

  1. The Model

As developed in our earlier works on structured SSMs, they seem to capture the essence of continuous, convolutional, and recurrent sequence models – all wrapped up in a simple and elegant model.

Despite the work that went into making Mamba fast, it’s still much less hardware-efficient than mechanisms such as attention.

  1. The Theory
  1. The Algorithm
  1. The Systems

{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

AI and Machine Learning Research

AI Ethics and Societal Impact

AI Applications and Demos

Miscellaneous

Humor and Memes


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Model Releases and Updates

AI Capabilities and Limitations

AI Development Tools and Techniques

AI Ethics and Societal Impact

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries

  1. LLM Advancements and Multimodal Applications:

    • Granite-8B-Code-Instruct from IBM enhances instruction-following for code tasks, surpassing major benchmarks. Stable Diffusion 3 Medium set to launch, promises better photorealism and typography, scheduled for June 12.

    • The AI engineering community discusses VRAM requirements for SD3, with predictions around 15GB while considering features like fp16 optimization for potential reductions. FlashAvatar promises 300FPS digital avatars using Nvidia RTX 3090, stoking interest in high-fidelity avatar creation.

  2. Fine-Tuning Techniques and Challenges:

    • Recommendations for overcoming tokenizer issues in half-precision training suggest tokenizer.padding_side = 'right' and using techniques from LoRA for enhanced fine-tuning. Axolotl users face issues with binary classification, suggesting Bert as an alternative.

    • Community insights highlight the effective use of Gradio's OAuth for private app access and the utility of share=True for quick app testing. Troubleshooting includes handling issues with inference setups in Kaggle and discrepancies with loss values in Axolotl, considering factors like input-output preprocessing.

  3. Open-Source Projects and Community Collaborations:

    • Manifold Research’s call for collaboration on multimodal transformers and control tasks aims to build a comprehensive open-source Generalist Model. StoryDiffusion and OpenDevin emerge as new open-source AI projects, sparking interest.

    • Efforts to integrate TorchAO with LM Evaluation Harness focus on adding APIs for quantization support. Community initiatives, such as adapting Axolotl for AMD compatibility, highlight ongoing efforts in refining AI tools and frameworks.

  4. AI Infrastructure and Security:

    • Hugging Face security incident prompts a recommendation for rotating tokens and switching to fine-grained access tokens, affecting users' infrastructures like HF Spaces. Discussions in OpenRouter reference database timeouts in Asia, leading to service updates and decommissioning certain models like Llava 13B and Hermes 2 Vision 7B.

    • ZeRO++ framework presents significant communication overhead reduction in large model training, aiding LLM implementations. The Paddler stateful load balancer enhances llama.cpp's efficiency, potentially streamlining model serving capabilities.

  5. AI Research and Ethical Discussions:

    • Yudkowsky’s controversial strategy against AI development sparks debate, with aggressive measures like airstrikes on data centers. LAION community reacts, discussing the balance between open collaboration and preventing misuse.

    • New Theories on Transformer Limitations: Empirical evidence suggests transformers struggle with composing functions on large domains, leading to new approaches in model design. Discussions on embedding efficiency continue, comparing context windows for performance across LLM implementations.


PART 1: High level Discord summaries

HuggingFace Discord


Unsloth AI (Daniel Han) Discord


Stability.ai (Stable Diffusion) Discord


Perplexity AI Discord


CUDA MODE Discord

Let's Chat Speculatively: Engineers shared insights into speculative decoding, with suggestions like adding gumbel noise and a deterministic argmax. Recorded sessions on the subject are expected to be uploaded after editing, and discussions highlighted the importance of ablation studies to comprehend sampling parameter impacts on acceptance rates.

CUDA to the Cloud: Rental of H100 GPUs was discussed for profiling purposes, recommending providers such as cloud-gpus.com and RunPod. The challenges in collecting profiling information without considerable hacking were also noted.

Work and Play: A working group for production kernels and another for PyTorch performance-related documentation were announced, inviting collaboration. Additionally, a beginner's tip was given to avoid the overuse of @everyone in the community to prevent unnecessary notifications.

Tech Talks on Radar: Upcoming talks and workshops include a session on Tensor Cores and high-performance scan algorithms. The community also anticipates hosting Prof Wen-mei Hwu for a public Q&A, and a session from AMD's Composable Kernel team.

Data Deep Dives and Development Discussions: Discussion in #llmdotc was rich with details like the successful upload of a 200GB dataset to Hugging Face and a proposal for LayerNorm computation optimization, alongside a significant codebase refactor for future-proofing and easier model architecture integration.

Of Precision and Quantization: The AutoFP8 GitHub repository was introduced, aiming at automatic conversion to FP8 for increased computational efficiency. Meanwhile, integrating TorchAO with the LM Evaluation Harness was debated, including API enhancements for improved quantization support.

Parsing the Job Market: Anyscale is seeking candidates with interests in speculative decoding and systems performance, while chunked prefill and continuous batching practices were underscored for operational efficiencies in predictions.

Broadcasting Knowledge: Recordings of talks on scan algorithms and speculative decoding are to be made available on the CUDA MODE YouTube Channel, providing resources for continuous learning in high-performance computing.

PyTorch Performance Parsing: A call to action was made for improving PyTorch's performance documentation during the upcoming June Docathon, with emphasis on current practices over deprecated concepts like torchscript and a push for clarifying custom kernel integrations.


LM Studio Discord

Note: Specific links to models, discussions, and GitHub repositories were provided in the respective channels and can be referred back to for further technical details and context.


Nous Research AI Discord


LLM Finetuning (Hamel + Dan) Discord


OpenAI Discord


Modular (Mojo 🔥) Discord


Eleuther Discord


OpenRouter (Alex Atallah) Discord

Database Woes in the East: OpenRouter users reported database timeouts in Asia, mainly in regions like Seoul, Mumbai, Tokyo, and Singapore. A fix was implemented which led to rolling back some latency improvements to address this issue.

OpenRouter Under Fire for API Glitches: Despite a patch, users continued to face 504 Gateway errors, with some temporarily bypassing the issue using EU VPNs. User suggestions included the addition of provider-specific uptime statistics for better service accountability.

Model Decommissioning and Recommendation: Due to low usage and high costs, OpenRouter is retiring models such as Llava 13B and Hermes 2 Vision 7B (alpha) and suggests switching to alternatives like FireLlava 13B and LLaVA v1.6 34B.

Seamless API Switcheroo: OpenRouter’s standardized API simplifies switching between models or providers, as seen in the Playground, without necessitating code alterations, acknowledging easier management for engineers.

Popularity Over Benchmarks: OpenRouter tends to rank language models based on real-world application, detailing model usage rather than traditional benchmarks for a pragmatic perspective available at OpenRouter Rankings.


LAION Discord


LlamaIndex Discord


Latent Space Discord

AI's Intrigue and Tumult in Latent Space: An AI Reverse Turing Test video surfaced, sparking interest by depicting advanced AIs attempting to discern a human among themselves. Meanwhile, accusations surfaced around llama3-V allegedly misappropriating MiniCPM-Llama3-V 2.5's academic work, as noted on GitHub.

The Future of Software and Elite Influence: Engineers digested the implications of "The End of Software," a provocative Google Doc, while also discussing Anthropic's Dario Amodei’s rise to Time's Top 100 after his decision to delay the chatbot Claude’s release. An O'Reilly article on operational aspects of LLM applications was also examined for insights on a year of building with these models.

AI Event Emerges as Industry Nexus: The recent announcement of the AI Engineering World Forum (AIEWF), detailed in a tweet, stoked anticipation with new speakers, an AI in Fortune 500 track, and official events covering diverse LLM topics and industry leadership.

Zoom to the Rescue for Tech Glitch: A Zoom meeting saved the day for members experiencing technical disruptions during a live video stream. They bridged to continued discussion by accessing the session through the shared Zoom link.


LangChain AI Discord

RAG Systems Embrace Historical Data: Community members discussed strategies for integrating historical data into RAG systems, recommending optimizations for handling CSV tables and scanned documents to enhance efficiency.

Game Chatbots Game Stronger: A debate on the structure of chatbots for game recommendations led to advice against splitting a LangGraph Chatbot agent into multiple agents, with a preference for a unified agent or pre-curated datasets for simplicity.

LangChain vs OpenAI Showdown: Conversations comparing LangChain with OpenAI agents pointed out LangChain's adaptability in orchestrating LLM calls, highlighting that use case requirements should dictate the choice between abstraction layers or direct OpenAI usage.

Conversational AI Subjects Trending in Media: Publications surfaced in the community include explorations of LLMs with Hugging Face and LangChain on Google Colab, and the rising importance of conversational agents in LangChain. Key resources include exploratory guide on Medium and a deep dive into conversational agents by Ankush k Singal.

JavaScript Meets LangServe Hurdle: A snippet shared the struggles within the JavaScript community when dealing with the RemoteRunnable class in LangServe, as evidenced by a TypeError related to message array processing.


tinygrad (George Hotz) Discord

Tinygrad Progress Towards Haskell Horizon: Discussions highlighted a member's interest in translating tinygrad into Haskell due to Python's limitations, while another suggested developing a new language specifically for tinygrad’s uop end.

Evolving Autotuning in AI: The community critiqued older autotuning methods like TVM, emphasizing the need for innovations that address shortcomings in block size and pipelining tuning to enhance model accuracy.

Rethinking exp2 with Taylor Series: Users, including georgehotz, examined the applicability of Taylor series to improve the exp2 function, discussing the potential benefits of CPU-like range reduction and reconstruction methods.

Anticipating tinygrad's Quantum Leap: George Hotz excitedly announced tinygrad 1.0's intentions to outstrip PyTorch in speed for training GPT-2 on NVIDIA and AMD, accompanied by a tweet highlighting upcoming features like FlashAttention, and proposing to ditch numpy/tqdm dependencies.

NVIDIA's Lackluster Showcase Draws Ire: Nvidia's CEO Jensen Huang's COMPUTEX 2024 keynote video raised expectations for revolutionary reveals but ultimately left at least one community member bitterly disappointed.


OpenAccess AI Collective (axolotl) Discord


Cohere Discord

Open Call for AI Collab: Manifold Research is on the hunt for collaborators to work on building an open-source "Generalist" model, inspired by GATO, targeting multimodal and control tasks across domains like vision, language, and more.

Cohere Community Troubleshoots: A broken dashboard link in the Cohere Chat API documentation was spotted and flagged, with community members stepping in to acknowledge and presumably kickstart a fix.

AI Model Aya 23 Gets the Thumbs Up: A user shares a successful testing of Cohire's Aya 23 model and hints at a desire to distribute their code for peer review.

Community Tag Upgrade Revealed: Discord's updated tagging mechanism sparks conversation and excitement in the community, with members sharing a link to the tag explanation.

Support Network Activated: For those experiencing disappearing chat histories or other issues, redirections to Cohere's support team at [email protected] or the server's designated support channel are provided.


OpenInterpreter Discord


Interconnects (Nathan Lambert) Discord


Mozilla AI Discord


DiscoResearch Discord


Datasette - LLM (@SimonW) Discord


AI21 Labs (Jamba) Discord


MLOps @Chipro Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI Stack Devs (Yoko Li) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

HuggingFace ▷ #announcements (1 messages):

Link mentioned: Spaces Overview: no description found


HuggingFace ▷ #general (974 messages🔥🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (28 messages🔥):

Link mentioned: GitHub - asukaminato0721/d2l.ai-rs: use candle to implement some of the d2l.ai: use candle to implement some of the d2l.ai. Contribute to asukaminato0721/d2l.ai-rs development by creating an account on GitHub.


HuggingFace ▷ #cool-finds (3 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (11 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (5 messages):

Link mentioned: TerDiT: Ternary Diffusion Models with Transformers: Recent developments in large-scale pre-trained text-to-image diffusion models have significantly improved the generation of high-fidelity images, particularly with the emergence of diffusion models ba...


HuggingFace ▷ #computer-vision (9 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (24 messages🔥):

Links mentioned:


HuggingFace ▷ #diffusion-discussions (1 messages):

Link mentioned: Diving into Diffusion Policy with LeRobot: In a recent blog post, we looked at the Action Chunking Transformer (ACT). At the heart of ACT lies an encoder-decoder transformer that when passed in * an image * the current state of the robot ...


HuggingFace ▷ #gradio-announcements (2 messages):

Link mentioned: Dynamic Apps With Render Decorator: A Step-by-Step Gradio Tutorial


Unsloth AI (Daniel Han) ▷ #general (919 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (13 messages🔥):


Unsloth AI (Daniel Han) ▷ #help (170 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (32 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #community-collaboration (4 messages):


Stability.ai (Stable Diffusion) ▷ #announcements (2 messages):

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (1009 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #general (796 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (33 messages🔥):


Perplexity AI ▷ #pplx-api (12 messages🔥):

Link mentioned: Supported Models: no description found


CUDA MODE ▷ #general (29 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (14 messages🔥):

Links mentioned:


CUDA MODE ▷ #torch (4 messages):


CUDA MODE ▷ #announcements (4 messages):


CUDA MODE ▷ #cool-links (2 messages):

Link mentioned: Mat’s Blog - CUDA MODE - Accelerate your code with massively parallel programming plus some other tricks: no description found


CUDA MODE ▷ #jobs (3 messages):

Links mentioned:


CUDA MODE ▷ #beginner (22 messages🔥):

Links mentioned:


CUDA MODE ▷ #pmpp-book (2 messages):

Link mentioned: Join our Cloud HD Video Meeting: Zoom is the leader in modern enterprise video communications, with an easy, reliable cloud platform for video and audio conferencing, chat, and webinars across mobile, desktop, and room systems. Zoom ...


CUDA MODE ▷ #youtube-recordings (12 messages🔥):

Links mentioned:


CUDA MODE ▷ #torchao (15 messages🔥):

Link mentioned: Build software better, together: GitHub is where people build software. More than 100 million people use GitHub to discover, fork, and contribute to over 420 million projects.


CUDA MODE ▷ #ring-attention (4 messages):


CUDA MODE ▷ #off-topic (36 messages🔥):


CUDA MODE ▷ #hqq (3 messages):

Link mentioned: Faster and Smaller Whisper: A Deep Dive into Quantization and Torch Compilation: A support blog for speeding up whisper by batch processing.


CUDA MODE ▷ #triton-viz (1 messages):

kerenzhou: It shows a single cta, right?


CUDA MODE ▷ #llmdotc (504 messages🔥🔥🔥):

Links mentioned:


CUDA MODE ▷ #bitnet (52 messages🔥):

Links mentioned:


CUDA MODE ▷ #pytorch-docathon (1 messages):

Link mentioned: Performance Tuning Guide is very out of date · Issue #2861 · pytorch/tutorials: 🚀 Descirbe the improvement or the new tutorial The first thing you see when you Google PyTorch performance is this. The recipe is well written but it's very much out of data today https://pytorch...


LM Studio ▷ #💬-general (338 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (86 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (22 messages🔥):


LM Studio ▷ #📝-prompts-discussion-chat (3 messages):


LM Studio ▷ #⚙-configs-discussion (26 messages🔥):


LM Studio ▷ #🎛-hardware-discussion (74 messages🔥🔥):

Links mentioned:


LM Studio ▷ #autogen (4 messages):


LM Studio ▷ #amd-rocm-tech-preview (1 messages):

zerocool9724: HIPSDK support is a hardware thing?


LM Studio ▷ #crew-ai (1 messages):

julio1307: Existe alguma alternativa "mais leve" em vez do LM Studio?


LM Studio ▷ #🛠-dev-chat (8 messages🔥):

Link mentioned: GitHub - JocysCom/VsAiCompanion: AI Companion that analyzes your project files and works alongside you to streamline development and boost productivity.: AI Companion that analyzes your project files and works alongside you to streamline development and boost productivity. - JocysCom/VsAiCompanion


Nous Research AI ▷ #ctx-length-research (1 messages):

manojbh: Do you have examples?


Nous Research AI ▷ #off-topic (57 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (10 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (250 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (53 messages🔥):


Nous Research AI ▷ #project-obsidian (1 messages):

manojbh: Are there benchmarks ?


Nous Research AI ▷ #rag-dataset (2 messages):


Nous Research AI ▷ #world-sim (21 messages🔥):

Link mentioned: GitHub - essserrr/crt-terminal: Retro styled terminal shell: Retro styled terminal shell. Contribute to essserrr/crt-terminal development by creating an account on GitHub.


LLM Finetuning (Hamel + Dan) ▷ #general (81 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #workshop-1 (9 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #🟩-modal (19 messages🔥):

Link mentioned: Pet Art Dreambooth with Hugging Face and Gradio: This example finetunes the Stable Diffusion XL model on images of a pet (by default, a puppy named Qwerty) using a technique called textual inversion from the “Dreambooth” paper. Effectively, it teach...


LLM Finetuning (Hamel + Dan) ▷ #learning-resources (9 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #hugging-face (1 messages):

The provided message history contains only a single question about credits:


LLM Finetuning (Hamel + Dan) ▷ #replicate (3 messages):


LLM Finetuning (Hamel + Dan) ▷ #berryman_prompt_workshop (1 messages):

computer_internet_man: all the old skills work, hoocoodanode


LLM Finetuning (Hamel + Dan) ▷ #whitaker_napkin_math (31 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #workshop-2 (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #workshop-3 (1 messages):

nik_hil__: I'm with u 👀


LLM Finetuning (Hamel + Dan) ▷ #abhishek_autotrain_llms (3 messages):

Link mentioned: GGUF My Repo - a Hugging Face Space by ggml-org: no description found


LLM Finetuning (Hamel + Dan) ▷ #clavie_beyond_ragbasics (3 messages):


LLM Finetuning (Hamel + Dan) ▷ #jason_improving_rag (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #jeremy_python_llms (3 messages):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #gradio (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #axolotl (34 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #zach-accelerate (23 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #wing-axolotl (23 messages🔥):

Link mentioned: minimal-trainer-zoo/sequence_classification.py at main · muellerzr/minimal-trainer-zoo: Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines - muellerzr/minimal-trainer-zoo


LLM Finetuning (Hamel + Dan) ▷ #freddy-gradio (2 messages):

Link mentioned: Sharing Your App: A Step-by-Step Gradio Tutorial


LLM Finetuning (Hamel + Dan) ▷ #charles-modal (43 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #langchain-langsmith (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #allaire_inspect_ai (3 messages):


LLM Finetuning (Hamel + Dan) ▷ #credits-questions (42 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #fireworks (6 messages):


LLM Finetuning (Hamel + Dan) ▷ #emmanuel_finetuning_dead (3 messages):

No links or blogposts were discussed in these messages.


LLM Finetuning (Hamel + Dan) ▷ #braintrust (5 messages):


LLM Finetuning (Hamel + Dan) ▷ #west-coast-usa (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #east-coast-usa (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #europe-tz (6 messages):


LLM Finetuning (Hamel + Dan) ▷ #announcements (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #predibase (7 messages):


LLM Finetuning (Hamel + Dan) ▷ #career-questions-and-stories (17 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #openai (1 messages):

peterg0093: The question is, will we get GPT-5 before we get OpenAI credits?


OpenAI ▷ #ai-discussions (315 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (55 messages🔥🔥):


OpenAI ▷ #prompt-engineering (7 messages):


OpenAI ▷ #api-discussions (7 messages):


Modular (Mojo 🔥) ▷ #general (198 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (1 messages):

ModularBot: From Modular: https://twitter.com/Modular/status/1797699002353488183


Modular (Mojo 🔥) ▷ #📺︱youtube (1 messages):

Link mentioned: Getting started with MAX release and nightly builds: In this video, we'll guide you through the entire process of installing and configuring both the MAX release and nightly builds on your system. You'll learn ...


Modular (Mojo 🔥) ▷ #🔥mojo (79 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (4 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #🏎engine (2 messages):


Modular (Mojo 🔥) ▷ #nightly (30 messages🔥):

Links mentioned:


Eleuther ▷ #general (28 messages🔥):

Links mentioned:


Eleuther ▷ #research (125 messages🔥🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (2 messages):

Links mentioned:


Eleuther ▷ #lm-thunderdome (7 messages):


Eleuther ▷ #multimodal-general (3 messages):

Link mentioned: Laying the Foundations for Vision and Multimodal Mechanistic Interpretability & Open Problems — AI Alignment Forum: Behold the dogit lens. Patch-level logit attribution is an emergent segmentation map. Join our Discord here. …


OpenRouter (Alex Atallah) ▷ #announcements (13 messages🔥):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (112 messages🔥🔥):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #소개 (1 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #일반 (1 messages):

lemmyle: 첫 번째


OpenRouter (Alex Atallah) ▷ #紹介 (1 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #一般 (1 messages):

lemmyle: 初め


LAION ▷ #general (98 messages🔥🔥):

Links mentioned:


LAION ▷ #research (5 messages):

Links mentioned:


LlamaIndex ▷ #announcements (1 messages):

Link mentioned: LlamaIndex Webinar: The Future of Web Agents with MultiOn 🤖 · Zoom · Luma: We are excited to chat about the Agentification of the Internet with Web Agents, with Div Garg from MultiOn! Context: We are transitioning into a world where…


LlamaIndex ▷ #blog (6 messages):


LlamaIndex ▷ #general (80 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):

crypto_carter: anyone working on combining semantic layers with SQL Retrievers?


Latent Space ▷ #ai-general-chat (51 messages🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

Link mentioned: Tweet from swyx 🇸🇬 (@swyx): Announcing second wave of speakers + Updates! @aidotengineer Changelog: ➕ Official Hackathon + Preparty Jun 24th ➕ see @HF0Residency announcement today 👀 ➕ hosting @ashtom as our Closing Keynote! ➕ ...


Latent Space ▷ #ai-in-action-club (33 messages🔥):

Link mentioned: Join our Cloud HD Video Meeting: Zoom is the leader in modern enterprise video communications, with an easy, reliable cloud platform for video and audio conferencing, chat, and webinars across mobile, desktop, and room systems. Zoom ...


LangChain AI ▷ #general (41 messages🔥):

Link mentioned: GitHub - MOUNAJEDK/GameSeeker-VideoGamesRecommendationChatbot at langgraph-logic-implementation: A chatbot specialized in offering personalized video games recommendations based on user's preferences. - GitHub - MOUNAJEDK/GameSeeker-VideoGamesRecommendationChatbot at langgraph-logic-imple...


LangChain AI ▷ #langserve (1 messages):


LangChain AI ▷ #langchain-templates (4 messages):


LangChain AI ▷ #share-your-work (13 messages🔥):

Links mentioned:


LangChain AI ▷ #tutorials (4 messages):

Links mentioned:


tinygrad (George Hotz) ▷ #general (42 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (8 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #general (29 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (8 messages🔥):

Link mentioned: Dmg02/names_textcat · Datasets at Hugging Face: no description found


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (7 messages):

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


Cohere ▷ #general (33 messages🔥):

Link mentioned: Using the Chat API: no description found


Cohere ▷ #project-sharing (6 messages):


OpenInterpreter ▷ #general (21 messages🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (11 messages🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (1 messages):

cyanidebyte: https://github.com/v2rockets/Loyal-Elephie


Interconnects (Nathan Lambert) ▷ #news (16 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (9 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (7 messages):


Interconnects (Nathan Lambert) ▷ #memes (1 messages):

Link mentioned: Tweet from Elon Musk (@elonmusk): no description found


Mozilla AI ▷ #announcements (1 messages):


Mozilla AI ▷ #llamafile (17 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (6 messages):

Links mentioned:


Datasette - LLM (@SimonW) ▷ #llm (5 messages):

Links mentioned:


AI21 Labs (Jamba) ▷ #jamba (5 messages):


MLOps @Chipro ▷ #events (2 messages):

Link mentioned: Alliance Medical Innovation Challenge: Empowering global health through AI-based solutions: solving the problems of the future




{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}