Frozen AI News archive

DBRX: Best open model (just not most efficient)

**Databricks Mosaic** has released a new open-source model called **DBRX** that outperforms **Grok**, **Mixtral**, and **Llama2** on evaluations while being about **2x more efficient** than Llama2 and Grok. The model was trained on **12 trillion tokens** using **3,000 H100 GPUs** over 2 months, with an estimated compute cost of **$10 million**. It uses OpenAI's **100k tiktoken tokenizer** and shows strong zero-shot code generation performance, even beating **GPT-4** on the Humaneval benchmark. DBRX also upstreamed work to **MegaBlocks** open source. Despite its scale and efficiency, DBRX's performance on MMLU is only slightly better than Mixtral, raising questions about its scaling efficiency. The focus of DBRX is on enabling users to train models efficiently, with MoE training being about **2x more FLOP-efficient** than dense models, achieving similar quality with nearly **4x less compute** than previous MPT models. This release is part of the ongoing competition for open-source AI leadership, including models like **Dolly**, **MPT**, and **Mistral**. *"If it activates 36B params, the model's perf should be equivalent to a 72B dense model or even 80B,"* says Qwen's tech lead.

Canonical issue URL

There's a LOT to like about Databricks Mosaic's new model (Corporate blog, team blog, free HF space demo, GitHub, HN, Jon Frankle tweet, Vitaliy Chiley tweet, Wired puff piece, terrible horrible no good very bad Techcrunch take, Daniel Han arch review, Qwen lead response):

All is truly great but you also have to be really good at reading between the lines to find what we're not saying above...

…or just read the right discords:

https://assets.buttondown.email/images/c9622806-2c2c-42b6-8629-b0178e19ff27.png?w=960&fit=max

In other words, a new MoE model trained on >12x the data and +50% the experts (and having +70% the param count per expert - 12 choose 4 of 12B experts vs 8 choose 2 of 7B experts) of Mixtral is somehow only 1% better than Mixtral on MMLU (however it is indeed great on coding). Weird, no? As Qwen's tech lead says:

"If it acativates 36B params, the model's perf should be equivalent to a 72B dense model or even 80B. In consideration of training on 12T tokens, I think it has the potential to be much better. 78 or higher for MMLU is what I expect."

Like Dolly and MPT before it, the main focus is more that "you can train models with us" than it is about really going after Mistral's open source crown:

"Our customers will find that training MoEs is also about 2x more FLOP-efficient than training dense models for the same final model quality. End-to-end, our overall recipe for DBRX (including the pretraining data, model architecture, and optimization strategy) can match the quality of our previous-generation MPT models with nearly 4x less compute."

Mosaic is already talking up the recent Lilac acquisition as a part of the story:

image.png


Table of Contents

[TOC]


REDDIT

AI Models and Benchmarks

AI Applications and Use Cases

AI Development and Optimization

AI Hardware and Infrastructure

AI News and Discussions


PART X: AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs

Model Releases & Updates

Frameworks & Tools

Research & Techniques

Discussions & Perspectives

Applications & Use Cases

Startups & Funding

Humor & Memes


PART 0: Summary of Summaries of Summaries


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord

Resolution Matters: Discussions highlighted that Stable Diffusion 1.5 (SD 1.5) functions optimally at base resolutions of 512x512. The community expects that Stable Diffusion 3 (SD3) will enhance the token limits and incorporate built-in expressions and actions.

VRAM Requirements for Stability: The AI engineers speculated on the capability of upcoming models like SD3 to operate efficiently on machines with 8GB or 12GB of VRAM. The benefits and potential drawbacks of transformers (xformers) were a heated topic.

Revving Up for Release: There is strong anticipation for the release of SD3 within the community, although no specific release date has been shared.

Game On With AI: Engineers exchanged ideas about using AI to create 2D game assets, suggesting the conversion of 3D model renderings into 2D pixel art. Recommendations favored Linux distributions such as Manjaro and Garuda for optimal performance on AMD GPUs.

Training Time Talk: A precise estimate is that it should take about an hour to train the lora on Stable Diffusion XL (SDXL) with high-end GPUs like the RTX 3090, given proper configurations.


Nous Research AI Discord

LLMs Face Memory Games and Falter: LLMs like Mistral 7B and Mixtral are finding it challenging to perform in-context recall tasks, which involve splitting and repeating sentences while maintaining their original context positions, even at token counts as low as 2500 or 5000. A benchmark to evaluate in-context recall called the ai8hyf split and recall test has been made available on GitHub, provoking conversations on the necessity for exact string matching and recall in sizable contexts.

Mixed Views on DBRX and Other Open Models: The community's hands-on experience with DBRX has been less than impressive, with feedback pointing to possible improvements via better fine-tuning or system prompt changes. Comparisons among various open models including Mixtral, Grok-1, Lemur-70B, and Nous Pro brought to light Mixtral's commendable performance, while some larger models did not see expected gains, spawning conversations about the MoE models' memory intensive nature and their trade-offs.

Innovations with Voice and Vision: The integration of voice chat using Deepgram & Mistral AI technology is showcased through a shared YouTube video, while ASRock’s Intel Arc A770 Graphics card is highlighted for its favorable specs over alternatives like the RTX 4070. Moreover, Databricks' release of open-license MoE LLM called DBRX Instruct offers a new player in the specialized domain of few-turn interactions, accessible via Hugging Face.

AI Conversations Take Whimsical Turns: World simulations involve AI displaying a penchant for characters like Sherlock Holmes and offbeat self-portrayals as trees and otherworldly beings, offering both amusement and unique roleplaying data. Meanwhile, issues with mobile responsiveness are being flagged, particularly on Samsung devices within the WorldSim framework.

RAG-ing Discussion and Collaborative Hermes: The community is actively discussing the critical role of retrieval in Retrieval Augmented Generation (RAG), alongside inventive approaches like Retrieval Augmented Thoughts (RAT) that couple RAG with Chain of Thought (CoT) prompting. A concerted effort is underway to advance Hermes, emphasizing datasets and techniques to enhance capabilities, documented in a collaborative Google Doc and noting the community's eagerness to contribute.


Unsloth AI (Daniel Han) Discord

F1 Score Custom Callback Is Here: A user's question about tracking F1 score values post-training has led to a consensus: you can indeed implement a custom callback to achieve this. Regardless of using Trainer or SFTTrainer, the outcome should be consistent.

Gemma & TinyLlama Get Continuous Attention: A community member focuses on continuous integration and iteration with models like gemma2b and tinyllama, targeting excellence.

Efficient Vector Database Enables Larger Embedding Handling: Cohere-ai released BinaryVectorDB, capable of efficiently managing hundreds of millions of embeddings, visible at BinaryVectorDB Repository.

Quantization and LISA Outshine in Model Training and Inference: Discussion spotlighted embedding quantization for efficient retrieval and the new Layerwise Importance Sampled AdamW (LISA), which outperforms LoRA with low memory consumption, detailed at LISA Paper on arXiv.

Localizing Large Language Models Yields Translation Treasure: Community focus turned to creating localized LLMs with the discussion about expanding LLMs to Korean through a method from Yanolja, plus Japanese web novels translations being aligned with English at ParallelFiction-Ja_En-100k.


Perplexity AI Discord


OpenInterpreter Discord


LM Studio Discord


Latent Space Discord


HuggingFace Discord

Chat Assistants Augment with Web Savvy: Hugging Face introduced chat assistants capable of conversing with information sourced from the web, subsequently pointed out by Victor Mustar on Twitter.

Sentence Transformers Amps Up: Release of Sentence Transformers v2.6.0 upgrades performance with features like embedding quantization and the GISTEmbedLoss; the announcement was made by Tom Aarsen via Twitter.

Hugging Face Toolkits Level Up: A slew of updates across a range of Hugging Face libraries, including Gradio and transformers.js, have brought new functionalities to the table, with more information detailed in Omar Sanseviero's tweet.

Rocking the 4D with Gaussian Splatting: A 4D Gaussian splatting demo on Hugging Face Space wowed users with its capability to explore scenes in new dimensions, showcased here.

Looking Ahead in NLP: An AI learning recruit eagerly sought a roadmap for NLP studies in 2024, focusing on recommended resources for a solid foundation in the field.

A Dive into Diffusion Discussions: Visionary approaches to training and image manipulation were brainstormed, with the sdxs model achieving impressive speeds, ControlNet offering outpainting guidance, and the discussion moving to Hugging Face Channels such as Diffusers' GitHub and Twitter for community engagement.

Apple Silicon Gets GPT's Attention: MacOS devices with Apple Silicon gain GPU acceleration alternatives with MPS backend support now integrated into Hugging Face's crucial training scripts.

Navigating the NLP Expanse: From seeking advice in [NLP] about a comprehensive roadmap for learning NLP in 2024 to discussions of new models and features in [i-made-this], the community is all about pushing the boundaries of what's possible with AI.

Vision Quest for Error-Detection: [computer-vision] members dug into models for detecting text errors in images, CT image preprocessing norms, fine-tuning specifics for SAM, and the challenges faced with image summarization for technical drawings highlighted with a mention of the Llava-next model.


LlamaIndex Discord


OpenAI Discord

Sora's Surreal Impressions Garner Praise: Influential visual artists such as Paul Trillo have lauded Sora for its ingenuity in creating novel and whimsical concepts; however, efforts to gain whitelist access to Sora for further experimentation have hit a brick wall, as the application pathway has been shuttered.

ChatGPT Flexes its Code Muscles: Exchanges within the community reveal a preference for Claude 3's coding prowess over GPT-4, suggesting that Claude may offer superior intelligence in coding tasks. Meanwhile, engineers also shared best practices to prevent ChatGPT from returning incomplete stub code, recommending explicit instructions to elicit full code outputs without placeholders.

AI Engineers Crave Enhanced PDF Parsing: Conversations around PDF data extraction have pinpointed the challenges of using models like gpt-3.5-turbo-16k. Strategies such as processing PDFs in smaller chunks and utilizing embeddings to preserve context across pages were discussed as potential solutions.

Undisclosed AI Chatbot Requirements Stir Curiosity: Speculation around the hardware specifications necessary to run a 60b parameter AI chatbot has surfaced, with mentions of using DeepSeekCoder's 67b model, despite limitations in locally running OpenAI models.

API Integration Woes Kindles Community Advice: When a fellow engineer struggled with the openai.beta.threads.runs.create method for custom assistant applications, advice flowed, highlighting the variance in responses between assistant APIs and potential need for tweaking the prompts or parameters for consistent results.


Eleuther Discord

AI Tokens: To Be Big or Not to Be: The community engaged in a heated debate about whether larger tokenizers are more efficient, balancing the cost-benefit for end-users against potential challenges in capturing word relationships. While some advocated for their efficiency, others questioned the impact on model performance, with relevant discussions sparked by sources like Aman Sanger's tweet.

Cheeky DBRX Outshines GPT-4?: DBRX, the new MoE LLM by MosaicML and Databricks with 132B parameters, has been launched, inciting discussions about its architecture and performance benchmarks, possibly outperforming GPT-4. Intrigued engineers can dive into the specifics on Databricks' blog.

Alternatives for Evaluating Autoregressive Models on Squad: Suggestions range from using alternative candidate evaluation methods to constrained beam search, highlighting complications from tokenizer nuances. Additionally, papers on Retrieval Augmented FineTuning (RAFT) were shared, challenging traditions in "open-book" information retrieval tasks. The RAFT concept can be explored further here.

Seeking Unity in AI Software: An industry collaboration titled The Unified Acceleration Foundation (UXL) is in motion to create an open-source rival to Nvidia's CUDA, powering a movement for diversity in AI software.

muP's Secret Sauce in AI Models: Amidst whispers in the community, muP remains unpublicized as a tuning parameter for large models, while Grok-1's GitHub repo shows its implementation, fueling speculation on normalization techniques and their impacts on AI modeling. For a peek at the code, visit Grok-1 GitHub.


LAION Discord

Bold Leaps in AI Safety and Efficiency: Discussions highlighted concerns about AI models generating inappropriate content with unconditional prompts, alongside an in-depth article that examined the impact of language models on AI conference peer reviews. Technical debates orbited around strategies to mitigate catastrophic forgetting during finetuning, as exemplified by models like fluffyrock, and a YouTube tutorial focusing on continual learning was referenced.

Delving Into Job Markets and Satirical Skepticism: A job opening at a startup focused on diffusion models and fast inference was shared, with details available on Notion, while the complexity of claims about self-aware AI, specifically regarding Claude3, sparked humor in light of proof-reading applications, with related OpenAI chats (one, two) shared for context.

AI Ethics in the Limelight: A Twitter post showcasing potentially misleading data representation led to a broader conversation on ethical visualization practices, criticizing how axes manipulation can distort performance perception, as seen in the offending tweet.

Impressive Speeds with SDXS Models: SDXS models have accelerated diffusion model performance to impressive frame rates, achieving up to 100 FPS and 30 FPS on the SDXS-512 and SDXS-1024 models, respectively — a noteworthy jump on a single GPU.

Innovation in Multilingual Models and Dimensionality Reduction: The debut of Aurora-M, a multilingual LLM, brazens the landscape with continual pretraining goals and red teaming prospects, whereas new research points to layer-pruning with minimal performance loss in LLMs that use open-weight pretrained models. A novel image decomposition method, B-LoRA, achieves high-fidelity style-content separation, while scripts for automating image captioning with CogVLM and Dolphin 2.6 Mistral 7b - DPO show promise in processing vast image datasets and are available on GitHub.


CUDA MODE Discord

FSDP Shines in New Runs: Recent training runs with adamw_torch and fsdp on a 16k context show promising loss improvements, detailed on Weights & Biases. A PyTorch FSDP tutorial was recommended alongside a GitHub issue on loss instability to those compiling resources on Fully Sharded Data Parallel (FSDP) training.

ImportError Issues in Triton Ecosystem: Discord users faced ImportError complications involving libc.so.6 and triton_viz. Cloning the Triton repo and installing from source was suggested, while Triton's official wheel pipeline's failure was noted, requiring custom solutions until fixed.

CUDA and PyTorch Data Wrangling: A Discord member presented difficulties encountered when handling uint16 and half data types in CUDA and PyTorch. They reported linker errors and utilized reinterpret_cast to circumvent the issue, advocating for compile-time errors in PyTorch to mitigate runtime surprises.

Tackling MSVC and PyTorch C++ Binding Bugs: Users grappled with issues binding C++ to PyTorch on Windows due to platform constraints and compatibility hitches like the mismatch between CUDA and PyTorch versions. The successful approach involved matching CUDA 11.8 with PyTorch's version, resolving the ImportError.

SSD Bandwidth and IO Bound Operations: A Discord engineer pointed out that SSD IO bandwidth limits heavily influence operation performance, even with optimizations like rapids and pandas. This illuminates a perpetual challenge in achieving minimal Speed of Light (SOL) times on IO-bound processes in compute environments.


OpenAccess AI Collective (axolotl) Discord

Haiku's Potential Belies Its Size: Engineers are intrigued by Haiku's canniness despite having just 20 billion parameters, suggesting that data quality might be more significant than sheer size in LLMs.

Axolotl Users Encounter Docker Difficulties: One user faced trouble with the Axolotl Docker template on Runpod, which sparked a recommendation to change the volume to /root/workspace and reclone Axolotl as a possible fix.

Databricks Enters the MoE Fray: Databricks' DBRX Base, a MoE architecture-based LLM, emerges as a model to watch, with pondering around its training methodologies and how it stacks up against peers like Starling-LM-7B-alpha, which has shown superior benchmarking results and is available at Hugging Face.

Hugging Face Faces Pricey Critique and VLLM Lack: Some members voice dissatisfaction with Hugging Face, calling it "overpriced" and noting the absence of very large language models on the platform.

Philosophical AI Goes Beyond Technical Yardstick: In the community showcase, members lauded the advent of Olier, an AI finetuned on Indian philosophy texts, marking achievements in using structured datasets for deep subject matter understanding and advancing the dialogue capabilities of specialized AIs.


Modular (Mojo 🔥) Discord

Mojo Learning and Debugging Discourse: A mojolings tutorial is available on GitHub, helping newcomers to grasp Mojo concepts. Participants have shared tips for debugging Mojo in VSCode, including a workaround for breakpoint issues.

Rust and Mojo's Borrow Checker Brainstorm: Conversations circled around the complexities of Rust's borrow checker and anticipations for Mojo's upcoming borrow checker with "easier semantics." There's curiosity about linked lists and how they will integrate with Mojo, hinting at potential innovation in borrow checking with Mojo's model.

Modular on Social Media Splash: Modular tweeted updates which can be found here and here.

Deployment Made Simpler with AWS Integration: A blog walkthrough covers deploying models on Amazon SageMaker, notably MAX optimized model endpoints, including steps from model download to deployment on EC2 c6i.4xlarge instances – simplify the process here.

TensorSpec Troubles and Community Code Contribution: A member sought clarification on TensorSpec inconsistencies noted in the Getting Started Guide vs. the Python API reference. Community contributions include momograd, a Mojo implementation of micrograd, open for feedback.


LangChain AI Discord


tinygrad (George Hotz) Discord


Interconnects (Nathan Lambert) Discord

DBRX Makes a 132B Parameter Entrance: MosaicML and Databricks introduced DBRX, a large language model with 132B parameters and a 32k context length, available commercially via Hugging Face. While it's not open-weight, the promise of new SOTA benchmarks stirs up the community, alongside discussions of a constrictive license preventing use in improving other models.

Mosaic's Law Forecasts Costly Reductions: A community member highlighted Mosaic's Law, which predicts the cost of models with certain capabilities will reduce to one-fourth annually due to advancements in hardware, software, and algorithms. Meanwhile, a notable DBRX license term sparked debate by forbidding the use of DBRX to enhance any models outside its ecosystem.

GPT-4 Clinches SOTA Evaluation Crown: Conversations swirled around GPT-4's superior performance, its adoption as an evaluation tool over other models, and an innovative way to fund these experiments using an AI2 credit card. The cost efficiency and practicality of using GPT-4 are changing the game for researchers and engineers.

Fireside Chat Reveals Mistral's Heat: Interactions in the community unveiled a lighthearted fascination with Mistral's leadership, culminating in a YouTube Fireside Chat with CEO Arthur Mensch discussing open source, LLMs, and agent frameworks.

Reinforcement Gradient of Debate: AI engineers dissected the practicality of a binary classifier in a Reinforcement Learning with Human Feedback (RLHF) setting, raising concerns about effectiveness and learning without partial credits. The discussions cast doubt on whether a high-accuracy reward model alone could tune a successful language model and underlined the struggle of learning from sparse rewards without recognizing incremental progress.


OpenRouter (Alex Atallah) Discord


DiscoResearch Discord

Prompt Localization Matters: A discussion highlighted the potential degradation of German language model performance when fine-tuning with English prompts, suggesting language-specific prompt designs to prevent prompt bleed. The German translation for "prompt" includes Anweisung, Aufforderung, and Abfrage.

DBRX Instruct Revealed: Databricks introduced DBRX Instruct, a 132 billion-parameter open-source MoE model trained on 12 trillion tokens of English text, promising innovations in model architecture as detailed in their technical blog post. The model is available for trials in a Hugging Face Space.

Educational Resources for LLM Training?: A member sought knowledge on training large language models (LLMs) from scratch, sparking a conversation on available resources for this intricate process.

RankLLM Approach for German: There's a growing interest in adapting the RankLLM method, a specialized technique for zero-shot reranking, for German LLMs. A detailed examination of this topic is available in a thorough article.

Enhancing German Data Sets: Talk centered around dataset enhancement for German models, including a shared difficulty due to dataset size when fine-tuning Mistral. A community call for collaboration to improve German datasets was made, with a strategy to merge datasets to achieve a substantial 10,000 samples.


Alignment Lab AI Discord


Datasette - LLM (@SimonW) Discord

LLM Plugin: Handle with Care: A new LLM command-line plugin, llm-cmd, introduced by Simon Willison, allows for dynamic generation and execution of terminal commands, but users are cautioned due to its potential risks.

Show and Don't Tell: The usage example for llm-cmd included showing the first three lines of every file within a directory to demonstrate its practical utilities.

Plugin Performance Issues Spark Investigation: Users reported that the llm-cmd experienced indefinite hangs upon execution, prompting discussions on basic diagnostic approaches, while usual queries remained functional.

Honing In on the Culprit: Detailed troubleshooting revealed the input() function and readline.set_startup_hook() to be problematic in llm_cmd.py, specifically failing to insert text in the shell as anticipated in the LLM environment.

Clarity in Communication is Key: Discussions highlighted that clarity is essential, particularly when referencing llm, to avoid multiple interpretations which might confuse the user base.


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ▷ #general-chat (834 messages🔥🔥🔥):

(Note: The provided summary includes conversations only up to the cut-off message, which asked if image creation was still possible on the server. No further context was provided.)

Links mentioned:


Nous Research AI ▷ #ctx-length-research (10 messages🔥):

Links mentioned:


Nous Research AI ▷ #off-topic (40 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (9 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (345 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (44 messages🔥):

Link mentioned: Tweet from Cody Blakeney (@code_star): It’s finally here 🎉🥳 In case you missed us, MosaicML/ Databricks is back at it, with a new best in class open weight LLM named DBRX. An MoE with 132B total parameters and 32B active 32k context len...


Nous Research AI ▷ #rag-dataset (15 messages🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (249 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (302 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (4 messages):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (134 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (3 messages):

Link mentioned: Open LLM Leaderboard - a Hugging Face Space by HuggingFaceH4: no description found


Unsloth AI (Daniel Han) ▷ #suggestions (61 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #general (409 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (18 messages🔥):


Perplexity AI ▷ #pplx-api (18 messages🔥):


OpenInterpreter ▷ #general (188 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (140 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (4 messages):

Link mentioned: Pollen-Vision: Unified interface for Zero-Shot vision models in robotics: no description found


LM Studio ▷ #💬-general (193 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (29 messages🔥):

Links mentioned:


LM Studio ▷ #🎛-hardware-discussion (56 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (30 messages🔥):

Links mentioned:


LM Studio ▷ #langchain (1 messages):

Given the limited context and content provided, there's no substantive summary to be made from the message presented. The user's message appears to be a request for help or insight regarding an unspecified issue they've encountered, mentioning the use of various tutorials without success. No further information, discussion points, or specific topics were provided in the excerpt to create summary bullet points. If more messages or context were available, it could lead to a more comprehensive summary.


LM Studio ▷ #amd-rocm-tech-preview (5 messages):


LM Studio ▷ #crew-ai (7 messages):


Latent Space ▷ #ai-general-chat (100 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (3 messages):


Latent Space ▷ #llm-paper-club-west (183 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #announcements (10 messages🔥):

Links mentioned:


HuggingFace ▷ #general (162 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (2 messages):

Link mentioned: Groking Groq: A Deep Dive on Deep Learning: To "Grok" is to learn something deeply- as if you're drinking it in. AI has a way of requiring that you Grok a number of seemingly unrelated topics; making i...


HuggingFace ▷ #cool-finds (5 messages):

Link mentioned: Assessing GPT-4 for cell type annotation in single-cell RNA-seq analysis - Nature Methods: This study evaluates the performance of GPT-4 in single-cell type annotation.


HuggingFace ▷ #i-made-this (26 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (6 messages):

Links mentioned:


HuggingFace ▷ #core-announcements (1 messages):

Link mentioned: Build software better, together: GitHub is where people build software. More than 100 million people use GitHub to discover, fork, and contribute to over 420 million projects.


HuggingFace ▷ #computer-vision (22 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (1 messages):


HuggingFace ▷ #diffusion-discussions (19 messages🔥):

Links mentioned:


LlamaIndex ▷ #blog (3 messages):

Link mentioned: LLM Meetup with Predibase, LlamaIndex, Guardrails and Tryolabs | San Francisco · Luma: LLMOps: From Prototype To Production | Developer Meetup Join Predibase, LlamaIndex, Guardrails AI, and Tryolabs for an evening of food, drinks, and discussions on all things LLMOps while...


LlamaIndex ▷ #general (221 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (3 messages):


OpenAI ▷ #ai-discussions (90 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (14 messages🔥):


OpenAI ▷ #prompt-engineering (39 messages🔥):


OpenAI ▷ #api-discussions (39 messages🔥):


Eleuther ▷ #general (143 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (25 messages🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (13 messages🔥):

Link mentioned: grok-1/run.py at main · xai-org/grok-1: Grok open release. Contribute to xai-org/grok-1 development by creating an account on GitHub.


LAION ▷ #general (103 messages🔥🔥):

Links mentioned:


LAION ▷ #research (7 messages):

Links mentioned:


CUDA MODE ▷ #general (13 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (1 messages):

Link mentioned: Tweet from GitHub - FixTweet/FxTwitter: Fix broken Twitter/X embeds! Use multiple images, videos, polls, translations and more on Discord, Telegram and others: Fix broken Twitter/X embeds! Use multiple images, videos, polls, translations and more on Discord, Telegram and others - FixTweet/FxTwitter


CUDA MODE ▷ #cuda (1 messages):

Link mentioned: Untitled formOpenCV dnn cuda interface survey : OpenCV dnn cuda interface survey


CUDA MODE ▷ #torch (2 messages):


CUDA MODE ▷ #jobs (5 messages):

Link mentioned: CAREERS AT NVIDIA: no description found


CUDA MODE ▷ #beginner (17 messages🔥):

Link mentioned: Issues · pytorch/pytorch: Tensors and Dynamic neural networks in Python with strong GPU acceleration - Issues · pytorch/pytorch


CUDA MODE ▷ #pmpp-book (8 messages🔥):


CUDA MODE ▷ #torchao (1 messages):

marksaroufim: new RFC https://github.com/pytorch-labs/ao/issues/86


CUDA MODE ▷ #ring-attention (17 messages🔥):

Links mentioned:


CUDA MODE ▷ #gtc-meetup (1 messages):

vim410: oops. i missed this! i was at GTC and now i am back to middle of nowhere


CUDA MODE ▷ #triton-puzzles (26 messages🔥):

Link mentioned: GitHub - Deep-Learning-Profiling-Tools/triton-viz: Contribute to Deep-Learning-Profiling-Tools/triton-viz development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #general (60 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (14 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (8 messages🔥):

Link mentioned: transformers/examples/pytorch/language-modeling/run_clm.py at f01e1609bf4dba146d1347c1368c8c49df8636f6 · huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. - huggingface/transformers


OpenAccess AI Collective (axolotl) ▷ #bots (1 messages):

anothermetic: <@1163482975883772027> you work yet?


OpenAccess AI Collective (axolotl) ▷ #community-showcase (6 messages):

Link mentioned: Introducing Olier – an Integral Yoga AI initiative – La Grace: no description found


OpenAccess AI Collective (axolotl) ▷ #deployment-help (3 messages):


Modular (Mojo 🔥) ▷ #general (40 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (2 messages):


Modular (Mojo 🔥) ▷ #✍︱blog (1 messages):

Link mentioned: Modular: Deploying MAX on Amazon SageMaker: We are building a next-generation AI developer platform for the world. Check out our latest post: Deploying MAX on Amazon SageMaker


Modular (Mojo 🔥) ▷ #tech-news (4 messages):


Modular (Mojo 🔥) ▷ #🔥mojo (5 messages):


Modular (Mojo 🔥) ▷ #community-projects (1 messages):

Link mentioned: GitHub - dorjeduck/momograd: A Learning Journey: Micrograd in Mojo 🔥: A Learning Journey: Micrograd in Mojo 🔥 . Contribute to dorjeduck/momograd development by creating an account on GitHub.


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (1 messages):


Modular (Mojo 🔥) ▷ #🏎engine (1 messages):

Links mentioned:


LangChain AI ▷ #general (44 messages🔥):

Links mentioned:


LangChain AI ▷ #langserve (1 messages):


LangChain AI ▷ #share-your-work (4 messages):

Links mentioned:


LangChain AI ▷ #tutorials (3 messages):

Links mentioned:


tinygrad (George Hotz) ▷ #general (18 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (31 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (10 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (4 messages):

Link mentioned: Fireside Chat w/ Mistral CEO, Arthur Mensch: Join us to hear from Arthur Mensch, Co-founder & CEO of Mistral, in conversation w/ Elad Gil.​​Topics covered will include:​Open source & LLMs​Agents and mul...


Interconnects (Nathan Lambert) ▷ #rlhf (11 messages🔥):


Interconnects (Nathan Lambert) ▷ #reads (13 messages🔥):


OpenRouter (Alex Atallah) ▷ #app-showcase (5 messages):

Link mentioned: GitHub - mintsuku/sora: Sora is a Discord bot that integrates with the Open Router API to facilitate conversation in Discord servers.: Sora is a Discord bot that integrates with the Open Router API to facilitate conversation in Discord servers. - mintsuku/sora


OpenRouter (Alex Atallah) ▷ #general (30 messages🔥):

Links mentioned:


DiscoResearch ▷ #general (9 messages🔥):

Links mentioned:


DiscoResearch ▷ #embedding_dev (5 messages):

Link mentioned: Strategies for Effective and Efficient Text Ranking Using Large Language Models: The previous article did a deep dive into the prompting-based pointwise, pairwise, and listwise techniques that directly use LLMs to perform reranking. In this article, we will take a closer look at s...


DiscoResearch ▷ #discolm_german (18 messages🔥):


Alignment Lab AI ▷ #ai-and-ml-discussion (3 messages):


Alignment Lab AI ▷ #general-chat (8 messages🔥):


Alignment Lab AI ▷ #open-orca-community-chat (1 messages):

twistedshadows.: <@949913143277146154>


Datasette - LLM (@SimonW) ▷ #llm (12 messages🔥):

Link mentioned: llm cmd undo last git commit—a new plugin for LLM: I just released a neat new plugin for my LLM command-line tool: llm-cmd. It lets you run a command to to generate a further terminal command, review and edit that …


Skunkworks AI ▷ #off-topic (1 messages):

pradeep1148: https://www.youtube.com/watch?v=Kan7GofHSwg