Frozen AI News archive

ReALM: Reference Resolution As Language Modeling

**Apple** is advancing in AI with a new approach called **ReALM: Reference Resolution As Language Modeling**, which improves understanding of ambiguous references using three contexts and finetunes a smaller **FLAN-T5** model that outperforms **GPT-4** on this task. In Reddit AI news, an open-source coding agent **SWE-agent** achieves **12.29%** on the SWE-bench benchmark, and **RAGFlow** introduces a customizable retrieval-augmented generation engine. A new quantization method, **QuaRot**, enables efficient 4-bit inference. AI applications include a t-shirt design generator, **podgenai** for GPT-4 based podcast generation, and an open-source model from **HuggingFace** that runs without a GPU. Industry discussions focus on the impact of large language models on the AI field and efforts to decentralize AI development. **Takuto Takizawa** joins **Stability AI Japan** as Head of Sales & Partnerships.

Canonical issue URL

Apple is finally waking up to AI in a big way ahead of WWDC. We featured MM1 a couple weeks ago and now a different team is presenting ReALM: Reference Resolution As Language Modeling. Reference resolution in their terminology refers to understanding what ambiguous references like "they" or "that" or "the bottom one" or "this number present onscreen" refer to, based on 3 contexts - 1) what's on screen, 2) entities relevant to the conversation, and 3) background entities. They enable all sorts of assistant-like usecases:

image.png

Which is a challenging task given it basically has to read your mind.

The authors use a mix of labeled and synthetic data to finetune a much smaller FLAN-T5 model that beats GPT4 at this task: image.png

No model release, no demo. But it's nice to see how they are approaching this problem, and the datasets and models are small enough to be replicable for anyone determined enough.

The AI content creator industrial complex has gone bonkers over it, of course. There only a few more months' worth of headlines to make about things beating GPT4 before this is itself beaten to death.


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence. Comment crawling still not implemented but coming soon.

AI Research and Development

AI Applications and Tools

AI Industry and Trends

Stable Diffusion Discussion

AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

Anthropic Research on Jailbreaking LLMs

Adversarial Validation Technique for Identifying Distribution Shifts

Impact of Taiwan Earthquake on Semiconductor Supply

AI Advancements and Developments

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries

  1. Advancements in Memory-Efficient LLM Training:

    • A new attention mechanism called DISTFLASHATTN claims to reduce quadratic peak memory usage to linear for training long-context LLMs, enabling up to 8x longer sequences. However, the paper lacks pseudocode for the backward pass, raising concerns about reproducibility.
    • Discussions around CUDA optimization techniques like DISTFLASHATTN and its potential to revolutionize LLM training through memory efficiency and speed improvements over existing solutions like Ring Self-Attention.
  2. AI Model Evaluations and Benchmarking:

    • The SWE-agent open-source system claims comparable accuracy to Devin on the SWE-bench for autonomously solving GitHub issues.
    • Varying performance of models like GPT-4, Claude, and Opus on tasks like solving historical prompts, math riddles, and code generation, highlighting the need for comprehensive evaluations.
    • Platforms like Chaiverse.com for rapid feedback on RP-LLM models and LMSys Chatbot Arena Leaderboard for model benchmarking.
  3. Prompt Engineering and Multimodal AI:

    • Discussions on prompt engineering techniques for tasks like translation while preserving markdown, generating manager prompts, and improving multimodal QA using Chain of Thought.
    • The potential of DSPy for prompt optimization compared to other frameworks like LangChain and LlamaIndex.
    • Explorations into multimodal AI like using Stable Diffusion for depth mapping from stereo images and the launch of Stable Audio 2.0 for high-quality music generation.
  4. Open-Source AI Developments and Deployments:

  5. Misc Themes:


PART 1: High level Discord summaries

LAION Discord


Stability.ai (Stable Diffusion) Discord

Forge Ahead with Stable Diffusion: Users report that Forge, a user interface for Stable Diffusion, delivers superior performance especially on RTX 3060 and RTX 4080 graphics cards. DreamShaper Lightning (SDXL) models come recommended for efficiency and speed in image generation.

Anticipation High for SD3: The Stable Diffusion community is actively awaiting the release of Stable Diffusion 3.0, projected to launch in the next 3-5 weeks, with improvements to text rendering expected, though perfect spelling may remain elusive.

Creative AI Unleashed, But Not 'Unleash': Members are experimenting with Stable Diffusion to generate art for projects like tabletop RPGs and are considering storytelling through AI-generated visual narratives, possibly in comic or movie formats.

Tech Tips for Troubled Times: Discussions centered on addressing issues such as slow image generation and unwanted text appearance, with participants suggesting optimizations, and mentioning GitHub links as starting points for troubleshooting.

Features Forecast: There's evident excitement about upcoming features like sparse control net, SegMOE, and audiosparx models, with the community sharing resources and anticipating new possibilities for AI-generated content.


Unsloth AI (Daniel Han) Discord

Cortana 1.0 Chat Model Sparks Curiosity: Engineers discussed creating an AI prompt model named Cortana 1.0, based on the Halo series AI, emphasizing creating effective chat modes and prompt structures for streamlined interaction.

Unsloth Enterprise Capability Clarified: It was clarified that Unsloth Enterprise does indeed support full model training with a speed enhancement of 2-5x over FA2, rather than the expected 30-40x.

AI Optimization Exchange: A set of lively discussions covered diverse optimization topics, including advances in Unsloth AI with a mention of Daniel Han's Tweet, GitHub resources for accelerating AI inference like ipex-llm, and troubleshooting with AI models, notably the compatibility of SFTTrainer with Gemma models.

Innovative Approach to Asteroid Mining: The Open Asteroid Impact project captured interest with a novel concept of bringing asteroids to Earth to harness resources more effectively.

Groundwork for Full Stack Prospects: Solicitations for a skilled full stack developer within the community were made, and users were encouraged to DM if they could recommend or offer assistance.


Perplexity AI Discord

Reading Between the PDF Lines: Engineers discussed AI models such as Claude and Haiku for interpreting PDFs, with a focus on context windows and Perplexity's Pro features, especially the "Writing" focus and enabling "Pro" for accuracy. Some users favored Sonar for faster responses.

Ad-talk Sparks User Spat: The possibility of Perplexity introducing ads sparked debate, following statements by Perplexity's Chief Business Officer on integrating sponsored suggestions. Concerns were raised about the potential impact on the user experience for Pro subscribers, citing a Verge article on the subject.

PDF Roadblocks and Image Generation: While addressing technical issues, users clarified that Perplexity's mobile apps lack image generation support—an inconvenience tempered by the website’s desktop-like functionality on mobile devices for image generation. Separate discussions pointed to users wanting to lift the 25MB PDF limit for increased efficiency.

Engineers Exchange 'Supply Links': Referral programs and discounts became a hot topic, with mentions of savings through supplied links.

API Woes and Workarounds: Within the Perplexity API realm, users grappled with the lack of team support and payment issues for API credits, while also sharing frustrations over rate limits and receiving outdated responses from the sonar-medium-online model. The advice ranged from accurate request logging to refining system prompts for up-to-date news.

Curiosity Drives Deep Dives:


Latent Space Discord

Open Source AI Matches Devin: The SWE-agent presented as an open-source alternative to Devin has shown comparable performance on the SWE-bench, prompting discussions on its potential integrations and applications.

Apple's AI Research Readiness: A new paper by Apple showcases ReALM, hinting at AI advancements that could eclipse GPT-4's capabilities, closely integrated with the upcoming iOS 18 for improved Siri interactions.

Conundrum with Claude: Users are experimenting with Claude Opus but finding it challenged by complex tasks, leading to recommendations of the Prompt Engineering Interactive Tutorial for enhanced interactions with the model.

Supercharged Sound with Stable Audio 2.0: StabilityAI has introduced Stable Audio 2.0, pushing the boundaries of AI-generated music with its ability to produce full-length, high-quality tracks.

DALL-E Gets an Edit Button: ChatGPT Plus now includes features that allow users to edit DALL-E generated images and edit conversation prompts, bringing new dimensions of customization and control, detailed on OpenAI's help page.

DSPy Framework Discussion Heats Up: The LLM Paper Club scrutinized the DSPy framework's functionality and its advantage in prompt optimization over other frameworks, sparking ideas about its application in diverse projects such as voice API logging apps and a platform for summarizing academic papers.


Nous Research AI Discord


LM Studio Discord


OpenAI Discord


tinygrad (George Hotz) Discord

Saying Goodbye to a Linux GPU Pioneer: John Bridgman's retirement from AMD sparked discussions on his contributions to Linux drivers, with George Hotz commenting on the state of AMD's management and future directions. Hotz called for anonymous tips from AMD employees for a possible blog expose, amidst community concerns over AMD's follow-through on driver issues and open-source promises as highlighted in debates and a Phoronix article.

Linux Kernel and NVIDIA's Open Move: The discourse extended to implications of varying kernel versions, particularly around Intel's Xe and i915 drivers, and the transition preferences amongst Linux distributions, with a nod towards moving from Ubuntu 22.04 LTS to 24.04 LTS. Additionally, George Hotz referenced his contribution towards an open NVIDIA driver initiative, stirring conversations about the state of open GPU drivers compared to proprietary ones.

Tinygrad's Path to V1.0 Involves the Community: Exploration of tinygrad's beam search heuristic and CommandQueue functionality highlighted George Hotz's emphasis on the need for improved documentation to aid users in learning and contributing, including a proposed tutorial inspired by "Write Yourself a Scheme in 48 Hours". This goes hand-in-hand with community contributions, like this command queue tutorial, to polish tinygrad.

Active Member Engagement Strengthens Tinygrad: The community's initiative in creating learning materials received kudos, with members offering resources and stepping up to live stream their hands-on experiences with tinygrad, fostering a collaborative learning environment. This aligns with the collective goal to reach tinygrad version 1.0, cementing the platform's position as a tool for education and innovation.

Rethinking Memory Use in AI Models: A technical debate ensued on memory optimization during the forward pass of models, particularly regarding the use of activation functions with inverses, leveraging the inverse function rule. This represents the community's engagement in not only tooling but also foundational principles to refine processing efficiency in AI computations.


OpenInterpreter Discord

OpenInterpreter Dives into App Development: Development is progressing on an Open Interpreter iPhone app with about 40% completion, driven by community collaboration on GitHub, inspired by Jordan Singer's Twitter concept.

Making Tech More Accessible: There's a push in the Open Interpreter community to introduce a Conversational UI layer to aid seniors and the disabled, aiming to significantly streamline their interaction with technology.

Security Measures in a Digital Age: Members are warned to steer clear of potentially hazardous posts from a seemingly Open Interpreter X account suspected of being compromised, in efforts to avert crypto wallet intrusions.

Out-of-the-Box Porting Initiatives: OpenInterpreter is blurring platform lines with a new repo for Android's Termux installation, work on a M5 Cardputer port, and a discussion for implementing local STT solutions amid cost concerns with GPT-4.

Anticipation for AI Insights: The community shares a zest for in-depth understanding of LLMs, potentially indicating high interest in gaining advanced technical knowledge about AI systems.


Eleuther Discord


HuggingFace Discord

Boost Privacy in Repos: Hugging Face now enables enterprise organizations to set repository visibility to public or private by default, enhancing privacy control. Their tweet has more details.

Publish with a Command: Quarto users can deploy sites on Hugging Face using use quarto publish hugging-face, as shared in recent Twitter and LinkedIn posts.

Gradio's New Sleek Features: Gradio introduces automatic deletion of state variables and lazy example caching in the latest 4.25.0 release, detailed in their changelog.

Exploring the CLI Frontier: A shared YouTube video explains how to use Linux commands, containers, Rust, and Groq in the command line interface for developers.

Pushing LLMs to Operative Zen: A user inquires about fine-tuning language models on PDFs with constrained computational resources, with a focus on inference using open-source models. Meanwhile, a discussion unfolds about modifying special tokens in a tokenizer when fine-tuning an LLM.


LangChain AI Discord

Persistent Context Quest in Chat History: Engineers discussed maintaining persistent context in chats, especially when interfacing with databases of 'question : answer' pairs, but did not converge on a specific solution. Reference was made to LangChain issues and documentation for potential ways forward.

Video Tutorial For LangServe Playground: An informative video tutorial introducing the Chat Playground feature in LangServe was shared, aimed at easing the initial setup and showcasing its integration with Langsmith.

Voice Commands the Future: Launch of several AI voice apps such as CallStar AI and AllMind AI was announced, suggesting a trend towards voice as the interface for AI interactions. Links were provided for community support on platforms like Product Hunt and Hacker News.

AI Engineering Troubles and Tutorials: A CI issue was reported on a langchain-ai/langserve pull request; and guidance was sought for a NotFoundError when employing LangChain's ChatOpenAI and ChatPromptTemplate. Meanwhile, novices were directed to a comprehensive LangChain Quick Start Guide.

Galactic API Services Offered and Prompting Proficiency Test: GalaxyAI provided free access to premium AI models, emphasizing API compatibility with Langchain, although the service link was missing. Another initiative, GitGud LangChain, challenged proficient prompters to test a new code transformation tool to uphold code quality.


Modular (Mojo 🔥) Discord

Mojo Mingles with Memory Safety: The integration of Mojo language into ROS 2 suggests potential benefits for robotics development, enhanced by Mojo's memory safety practices. C++ and Rust comparison shows the growing interest in performance and safety in robotics environments.

Docker Builds Set Sails: Upcoming Modular 24.3 will include a fix aimed at improving the efficiency of automated docker builds, which has been well-received by the community.

Logger's Leap to Flexibility: The logger library in Mojo has been updated to accept arbitrary arguments and keyword arguments, allowing for more dynamic logging that accommodates versatile information alongside messages.

Mojo Dicts Demand More Speed: Community engagement on the One Billion Row Challenge revealed that the performance of Dict in Mojo needs enhancement, with efforts and discussions ongoing about implementing a custom, potentially SIMD-based, Dict that could keep pace with solutions like swiss tables.

The Collective Drive for Mojo's Nightly Improvements: Members expressed a desire for clearer pathways to contribution and troubleshooting for Mojo's stdlib development with discussions on GitHub clarifying challenges such as parsing errors and behavior of Optional types, indicative of active collaboration to refine Mojo's offerings.


OpenRouter (Alex Atallah) Discord


LlamaIndex Discord

RankZephyr Eclipses the Competition: The integration of RankZephyr into advanced Retrieval-Augmented Generation systems is suggested to enhance reranking, with the RankLLM collection recognized for its fine-tuning capabilities.

Enhancing Research Agility with AI Copilots: A webinar summary reveals key strategies in building an AI Browser Copilot, focusing on a prompt engineering pipeline, KNN few-shot examples, and vector retrieval, with more insights available on LlamaIndex’s Twitter.

Timely Data Retrieval Innovations: KDB.AI is said to improve Retrieval-Augmented Generation by incorporating time-sensitive queries for hybrid searching, facilitating a more nuanced search capability critical for contexts like financial reporting, as illustrated in a code snippet.

Intelligent Library Redefines Knowledge Management: A new LLM-powered digital library for professionals and teams is touted to revolutionize knowledge organization with features allowing creation, organization, and annotation in an advanced digital environment, as announced in a LlamaIndex tweet.

Community Dialogues Raise Technical Questions: Discussions in the community include challenges with indexing large PDFs, issues with qDrant not releasing a lock post IngestionPipeline, limitations of the HuggingFace API, model integration using the Ollama class, and documentation gaps in recursive query engines with RAG.


OpenAccess AI Collective (axolotl) Discord

Axolotl Docs Get a Fresh Coat: The Axolotl documentation received an aesthetic update, but a glaring omission of the Table of Contents was swiftly corrected as shown in this GitHub commit, although further cleanup is needed for consistency between headings and the Table of Contents.

Deployment Woes and Wins for Serverless vLLMs: Experiences with Runpod and serverless vLLMs were shared, highlighting challenges along with a resource on how to deploy large language model endpoints.

Data Aggregation Headaches: Efforts to unify several datasets, comprising hundreds of gigabytes, face complications including file alignment. Presently, TSV files and pickle-formatted index data are used for quick seeking amid discussions on more efficient solutions.

Casual AI Model Smackdown: A light-hearted debate compared the preferences of AI models such as 'qwen mow' vs 'jamba', with the community joking about the need for additional data and resources.

Call for High-Def Data: A community member seeks resources to obtain a collection of 4K and 8K images, indicating a project or research that demands high-resolution image data.


Mozilla AI Discord


Interconnects (Nathan Lambert) Discord


CUDA MODE Discord


AI21 Labs (Jamba) Discord

Token Efficiency Talk: A user highlighted a paper's finding that throughput efficiency increases with per token measurement, calculated by the ratio of end-to-end throughput (both encoding and decoding) over the total number of tokens.

Speed Debate Heats Up: There's a divide on how the addition of tokens affects generation speed — while encoding can be done in parallel, the inherent sequential nature of decoding suggests each new token would add to the processing time.

Focus on Encoding Performance: Clarification in the discussion pointed to a graph that plotted the speed of generating a fixed 512 tokens, implying that observed speed improvements in the plot should be attributed to faster encoding rather than decoding.

Decoding: The Sequential Slowdown Dilemma: Queries arose about the possibility of increasing the speed of decoding despite its sequential dependency, which theoretically mandates a waiting period for each token's predecessor.


Skunkworks AI Discord


Alignment Lab AI Discord

It seems there is not enough context to generate a summary. Please provide more information or discussions across the channels within the Discord guild to output a meaningful summary.


Datasette - LLM (@SimonW) Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

LAION ▷ #general (699 messages🔥🔥🔥):

Links mentioned:


LAION ▷ #research (11 messages🔥):

Links mentioned:


LAION ▷ #learning-ml (1 messages):

Link mentioned: Meetup #3 LangChain and LLM: Using LangSmith to go from prototype to production, mer. 17 avr. 2024, 18:30 | Meetup: Nous avons le plaisir d'accueillir Harrison Chase, le Co-Founder et CEO de LangChain, pour notre troisième Meetup LangChain and LLM France ! Ne loupez pas cette occasion u


Stability.ai (Stable Diffusion) ▷ #general-chat (568 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (241 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (12 messages🔥):

Link mentioned: Open Asteroid Impact: no description found


Unsloth AI (Daniel Han) ▷ #help (278 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #general (469 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (23 messages🔥):


Perplexity AI ▷ #pplx-api (24 messages🔥):


Latent Space ▷ #ai-general-chat (76 messages🔥🔥):

Links mentioned:


Latent Space ▷ #llm-paper-club-west (356 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #off-topic (4 messages):

Link mentioned: Tweet from John Yang (@jyangballin): SWE-agent is our new system for autonomously solving issues in GitHub repos. It gets similar accuracy to Devin on SWE-bench, takes 93 seconds on avg + it's open source! We designed a new agent-co...


Nous Research AI ▷ #interesting-links (17 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (137 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (34 messages🔥):

Links mentioned:


Nous Research AI ▷ #project-obsidian (3 messages):


Nous Research AI ▷ #bittensor-finetune-subnet (2 messages):


Nous Research AI ▷ #rag-dataset (7 messages):

Link mentioned: glaiveai/rag_sample · Datasets at Hugging Face: no description found


Nous Research AI ▷ #world-sim (88 messages🔥🔥):

Links mentioned:


LM Studio ▷ #💬-general (170 messages🔥🔥):

Link mentioned: AnythingLLM | The ultimate AI business intelligence tool: AnythingLLM is the ultimate enterprise-ready business intelligence tool made for your organization. With unlimited control for your LLM, multi-user support, internal and external facing tooling, and 1...


LM Studio ▷ #🤖-models-discussion-chat (13 messages🔥):

Link mentioned: MyAnimeList.net - Anime and Manga Database and Community : Welcome to MyAnimeList, the world's most active online anime and manga community and database. Join the online community, create your anime and manga list, read reviews, explore the forums, follo...


LM Studio ▷ #🧠-feedback (3 messages):


LM Studio ▷ #🎛-hardware-discussion (69 messages🔥🔥):

Links mentioned:


LM Studio ▷ #autogen (3 messages):


LM Studio ▷ #crew-ai (3 messages):


OpenAI ▷ #annnouncements (1 messages):


OpenAI ▷ #ai-discussions (173 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (57 messages🔥🔥):


OpenAI ▷ #prompt-engineering (11 messages🔥):


OpenAI ▷ #api-discussions (11 messages🔥):


tinygrad (George Hotz) ▷ #general (148 messages🔥🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (28 messages🔥):

Links mentioned:


OpenInterpreter ▷ #general (93 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (66 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (2 messages):


Eleuther ▷ #general (67 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (53 messages🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (4 messages):

Link mentioned: Neel Nanda MATS Stream - Admissions Procedure + FAQ: no description found


Eleuther ▷ #lm-thunderdome (24 messages🔥):

Link mentioned: Build software better, together: GitHub is where people build software. More than 100 million people use GitHub to discover, fork, and contribute to over 420 million projects.


Eleuther ▷ #gpt-neox-dev (2 messages):

Link mentioned: Quickstart — PyTorch 2.2 documentation: no description found


HuggingFace ▷ #announcements (3 messages):


HuggingFace ▷ #general (70 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (2 messages):

Link mentioned: Super User Do- Tinkering with Linux commands, Containers, Rust, and Groq: A brief intro for basic commands to navigate your computer from what's called the "command line interface" or "CLI". How to update, upgrade, move in and out ...


HuggingFace ▷ #cool-finds (5 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (14 messages🔥):

Links mentioned:


HuggingFace ▷ #computer-vision (3 messages):


HuggingFace ▷ #NLP (13 messages🔥):

Link mentioned: transformers/src/transformers/cache_utils.py at c9f6e5e35156e068b227dd9b15521767f6afd4d2 · huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. - huggingface/transformers


HuggingFace ▷ #diffusion-discussions (8 messages🔥):


HuggingFace ▷ #gradio-announcements (1 messages):


LangChain AI ▷ #general (104 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #langserve (2 messages):

Links mentioned:


LangChain AI ▷ #share-your-work (7 messages):

Links mentioned:


LangChain AI ▷ #tutorials (1 messages):

Link mentioned: Quickstart | 🦜️🔗 Langchain: In this quickstart we'll show you how to:


Modular (Mojo 🔥) ▷ #general (38 messages🔥):

Link mentioned: Mojo🔥 roadmap & sharp edges | Modular Docs: A summary of our Mojo plans, including upcoming features and things we need to fix.


Modular (Mojo 🔥) ▷ #💬︱twitter (3 messages):


Modular (Mojo 🔥) ▷ #✍︱blog (1 messages):

Link mentioned: Modular: What’s new in Mojo 24.2: Mojo Nightly, Enhanced Python Interop, OSS stdlib and more: We are building a next-generation AI developer platform for the world. Check out our latest post: What’s new in Mojo 24.2: Mojo Nightly, Enhanced Python Interop, OSS stdlib and more


Modular (Mojo 🔥) ▷ #ai (4 messages):

Link mentioned: GitHub - ros2-rust/ros2_rust: Rust bindings for ROS 2: Rust bindings for ROS 2 . Contribute to ros2-rust/ros2_rust development by creating an account on GitHub.


Modular (Mojo 🔥) ▷ #tech-news (2 messages):


Modular (Mojo 🔥) ▷ #🔥mojo (30 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (1 messages):


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (7 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #📰︱newsletter (2 messages):

Link mentioned: Modverse Weekly - Issue 28: Welcome to issue 28 of the Modverse Newsletter covering Featured Stories, the Max Platform, Mojo, & Community Activity.


Modular (Mojo 🔥) ▷ #nightly (13 messages🔥):

Link mentioned: mojo/stdlib/src/collections/optional.mojo at nightly · modularml/mojo: The Mojo Programming Language. Contribute to modularml/mojo development by creating an account on GitHub.


OpenRouter (Alex Atallah) ▷ #general (91 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #blog (4 messages):

Links mentioned:


LlamaIndex ▷ #general (45 messages🔥):

Link mentioned: Ollama - LlamaIndex: no description found


LlamaIndex ▷ #ai-discussion (7 messages):

Link mentioned: How to build a RAG app using Gemini Pro, LlamaIndex (v0.10+), and Pinecone: Let's talk about building a simple RAG app using LlamaIndex (v0.10+) Pinecone, and Google's Gemini Pro model. A step-by-step tutorial if you're just getting ...


OpenAccess AI Collective (axolotl) ▷ #general (48 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (5 messages):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (2 messages):


Mozilla AI ▷ #llamafile (36 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (1 messages):


Interconnects (Nathan Lambert) ▷ #random (29 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #sp2024-history-of-open-alignment (1 messages):


CUDA MODE ▷ #general (1 messages):

iron_bound: https://github.com/intel-analytics/ipex-llm


CUDA MODE ▷ #algorithms (4 messages):

Link mentioned: DISTFLASHATTN: Distributed Memory-efficient Attention for Long-context LLMs Training: FlashAttention (Dao, 2023) effectively reduces the quadratic peak memory usage to linear in training transformer-based large language models (LLMs) on a single GPU. In this paper, we introduce DISTFLA...


CUDA MODE ▷ #beginner (2 messages):

Link mentioned: CUDA MODE: A CUDA reading group and community https://discord.gg/cudamode Supplementary content here https://github.com/cuda-mode Created by Mark Saroufim and Andreas Köpf


CUDA MODE ▷ #ring-attention (2 messages):

Link mentioned: DISTFLASHATTN: Distributed Memory-efficient Attention for Long-context LLMs Training: FlashAttention (Dao, 2023) effectively reduces the quadratic peak memory usage to linear in training transformer-based large language models (LLMs) on a single GPU. In this paper, we introduce DISTFLA...


AI21 Labs (Jamba) ▷ #jamba (6 messages):


Skunkworks AI ▷ #general (1 messages):


Skunkworks AI ▷ #finetuning (1 messages):


Alignment Lab AI ▷ #general-chat (1 messages):

jinastico: <@748528982034612226>


Datasette - LLM (@SimonW) ▷ #llm (1 messages):