Frozen AI News archive

$100k to predict LMSYS human preferences in a Kaggle contest

**Llama 3 models** are making breakthroughs with Groq's 70B model achieving record low costs per million tokens. A new **Kaggle competition** offers a $100,000 prize to develop models predicting human preferences from a dataset of over 55,000 user-LLM conversations. Open source evaluator LLMs like **Prometheus 2** outperform proprietary models such as **GPT-4** and **Claude 3 Opus** in judgment tasks. New datasets like **WildChat1M** provide over 1 million ChatGPT interaction logs with diverse and toxic examples. Techniques like **LoRA fine-tuning** show significant performance gains, and **NVIDIA's NeMo-Aligner** toolkit enables scalable LLM alignment across hundreds of GPUs. Factuality-aware alignment methods are proposed to reduce hallucinations in LLM outputs.

Canonical issue URL

It's been a quiet week for AI news. This is a fun new Kaggle challenge:

You'll work with a dataset from the Chatbot Arena, containing conversations and user preferences across various LLMs. By developing a model that accurately predicts human preferences, you'll contribute to improving chatbot performance and alignment with user expectations. The training dataset includes over 55,000 real-world user and LLM conversations and user preferences, with personally identifiable information removed. Your solution submission will be tested on a hidden test set of 25,000 samples.

The competition will run until August 5th, with a total prize of $100,000, featuring a $25,000 prize for 1st place, 20,000 prizes for 2nd through 4th places, and a 15,000 prize for 5th place.


Table of Contents

[TOC]


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

LLM Model Releases and Benchmarks

Datasets and Benchmarking

Techniques for Efficient LLM Training and Inference

Multimodal and Long-Range LLMs

Emerging Architectures and Training Paradigms

Miscellaneous


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Model Releases and Updates

AI Applications and Demos

AI Societal Impact and Concerns

AI Research and Benchmarking


AI Discord Recap

A summary of Summaries of Summaries

1. Large Language Model (LLM) Advancements and Challenges

2. AI Model Fine-tuning and Optimization Strategies

3. Open Source AI Frameworks and Libraries

4. AI Hardware Acceleration and Optimization

5. Misc


PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord

New Frontier in LLM Finetuning: Community members discussed near-full finetuning possibilities with Unsloth, exploring the potential of setting all parameters except layernorms to trainable. While Unsloth is focused on addressing llama.cpp and GGUF conversions, particularly the quantization and loading checkpoint shards challenge, sentiment analysis enthusiasts received tips on formatting vast databases for LLM compatibility.

Experimental Full Finetuning Tactics and Dataset Structuring: Unofficial strategies to enable full finetuning on Unsloth were shared, demonstrating improved losses relative to standard Hugging Face implementations. Discussions also delved into ideal dataset structuring for optimization, suggesting strategies for handling multiple "rejected" responses.

Phi 3 Executes in Browser, But Llama 3 Discord Absent: A tweet here demonstrated running Phi 3 in a web browser, while a member clarified that no dedicated Discord channel exists for Llama 3. Meanwhile, incorporating new roles in Llama 3 sparked debate, with type=code being a suggested alternative for tool_call.

Adapting Llama 3 With Self-Discovery and Triton's TK-GEMM: One ingenious user applied techniques from the Self-Discovery paper to enhance the reasoning capabilities of ChatGPT. Moreover, a PyTorch blog post highlighted Triton's FP8 GEMM to accelerate Llama 3 on NVIDIA H100 GPUs, promising optimization insights.

Quantization Quandary and Finetuning Finesse: Issues emerged when converting Llama 3 to GGUF, impacting fine-tuning data integrity, and similar problems arose when melding Lora with GGUF models. However, a pathway to understanding finetuning and model management is becoming clearer, with established community members suggesting the use of Unsloth's Colab notebooks for guidance.


Stability.ai (Stable Diffusion) Discord


CUDA MODE Discord

Gradient Adornments in Conversations: Discord members discussed advanced gradient techniques within PyTorch, where create_graph=True is employed for finer gradient details and Hessian-vector products. Techniques to estimate the Hessian's diagonal were mentioned, leveraging randomness for the estimations.

Triton Trials and Triumphs: Engineers faced challenges with IncompatibleTypeErrorImpl in Triton, but found solace in a tl.cast function fix after stumbling upon a gather function issue. Kernel debugging with PyTorch in PyCharm also proved problematic, even when setting TRITON_INTERPRET to "1".

Patching it Up with tinygrad: Members shared a multi-GPU support patch for tinygrad, endorsing Nvidia's open drivers. A GitHub conundrum surfaced about the right way to install custom PyTorch and CUDA extensions, seeking clarity through examples in the PyTorch AO library's setup process.

Catalyzing Community Contributions: The Effort project on GitHub received accolades for its impactful structure, while GreenBitAI’s toolkit was introduced as an ML framework enhancing PyTorch. It includes innovative gradient calculation methods and a potentially useful gemv kernel for inference spotlighted in bitblas.

torch woes and wins: PyTorch developers debated build strategies and optimizations, from build times for linear algebra components to kernel performance. The idea of padding vocabulary size to fairly compete in performance benchmarks was deliberated, revealing the nuanced considerations needed for equitable measures.

A Taste of LLM Innards: The llm.c project reached new efficiencies with 167K tokens/second using CUDA optimization techniques. Key discussions on CUDA streams, fused classifiers, and the strategic use of atom variables with scratch buffers highlighted the dense technical camaraderie.

Open Source Intel: It was briefly mentioned that Intel is now added to the PyTorch website, indicating a potential integration or support update.


LM Studio Discord

CLI Joins the LM Studio Toolbox: LM Studio has launched its new CLI tool, lms, designed to simplify the management of local LLMs, including loading and unloading models and starting or stopping servers. The CLI tool is available for the latest LM Studio 0.2.22 and beyond, and users are encouraged to contribute to its open source GitHub repository.

Llama's Conversion Complication: Collaboration in the LM Studio guild led to the successful resolution of several integration issues with llama.cpp, utilizing scripts such as convert-hf-to-gguf. Some users faced FileNotFoundError that was fixed by redownloading necessary files via huggingface-cli, with the community assisting in addressing conversion execution problems.

Model Performance and Oddities: Discussion in the models channel revealed endeavors to enhance story writing with Goliath 120B Longlora models and experiments to assess recall capabilities of models like LLAMA 3 on extensive texts. A curiosity emerged about ChatQA 1.5 showcasing unexpected response templates, whereas a bug in the latest LM Studio 0.2.22 prompted a new update for corrected behavior.

ROCm's Growing Pains and Triumphs: Members explored the capabilities of the latest LM Studio 0.2.22 ROCm Preview, with some testing the upper limits of RAM and context sizes and others addressing issues with embedding models. The introduction of lms CLI for AMD ROCm's preview and Linux support triggered spirited discussions about the tool's potential, bolstered by efforts in headless mode execution and dockerization.

Server-Client Connect Unlocked: Tips and fixes for configurations were shred, including a handy way to repopulate default configs, resolving access to LM Studio through WSL by using correct IP addresses, and enabling seamless communication between Windows and WSL environments for the app without additional complexity.


Perplexity AI Discord


Nous Research AI Discord

Hermes 2 Pro Hops into the Fray: The recently released Hermes 2 Pro integrated with LLaMA weights is making waves with its advanced QA, Function Calling, and JSON Mode capabilities. It’s garnering attention for exceptional inference speeds on mobile devices and has support material on GitHub and Hugging Face.

ChatML Equation S-Bahn: Tweaks to enable ChatML like using token replacement strategies and altering EOS symbols are being dissected by members, though details on the modifications are sparse.

World-sim Codex: A lively discussion around world-sim pointed out recent updates and shifts, such as the introduction of the Iron Age, and shared resources on consciousness and AI with links to YouTube talks.

Dataset Seekers Untie: Members queried about free generic datasets suitable for finetuning LLMs prior to initiating mining sequences, prompting shared interest but limited response in channels marked #bittensor-finetune-subnet and #rag-dataset.

LLama Crafting Corner: Troubleshooting around llamacpp led to suggestions of using ollama to sidestep handling C directly and to employ techniques like quantization and pruning for ideal CPU-run LLM scenarios. The conversations also explored the intriguing concept of moral non-commutativity in retrocausality and the psychological impacts therein.


Modular (Mojo 🔥) Discord

Bringing Mojo to the Command Line: The prism CLI toolkit for Mojo has been augmented with new features such as persistent flags, hooks, and flag groups. Updates are showcased on the project's GitHub page.

Test Driven Mojo Development: mojo-pytest, the plugin for testing in Mojo, now supports the new version 24.3. An issue to improve debuggability is tracked at Issue #9 on GitHub.

NuMojo Outpaces Rivals: The NuMojo project, aiming to enhance Mojo's standard library tensor functionality, has been updated for Mojo version 24.3 and shown to perform better than NumPy and Numba in benchmarks. Check out NuMojo's progress on GitHub.

Adventures in Learning Mojo: For those curious to integrate Mojo into workflows, a new "Let's mojo build -D your own -D version=1 app" tutorial is available. It's designed to illustrate Mojo's capabilities through a series of workflows and can be found on GitHub.

Nightly Releases Keeping Mojo Fresh: Mojo's development strides forward with more frequent nightly releases—eventually daily—aligning with infrastructure improvements. Nightly changelogs, like the introduction of __source_location() and improved docstring flexibility, can be perused at the Modular Docs Changelog.

Maxing Out on MAX Extensibility: MAX 24.3 introduces the brand new MAX Engine Extensibility API which aims to perfect PyTorch, ONNX, and Mojo model integrations. Detailed information on performance and hardware optimization is provided in the MAX Graph APIs.


OpenAI Discord

AI Job Market Roulette: The community engaged in a humorous debate about the fleeting nature of high-paying jobs in AI, with quips about the potential profitability of unconventional career paths like AI CEO or even a dentist.

Speculation Station for GPT-5 Ticket Prices: There's chatter on the potential pricing strategy for GPT-5, with the group divided on whether OpenAI would opt for regional pricing models or stick with a single price point for all.

Deja Vu for GPT-3 Devotees and Chat Rooms: Members expressed nostalgia over GPT-3 and Codex, despite the buzz around GPT-4, and raised questions about the absence of voice chat rooms for real-time discussion, citing moderation concerns.

Response Time Riddle with GPT-4: Talks about GPT-4's response times being slower than GPT-3.5, with mentions of gpt4 turbo facing significant latency, indicating that engineers are keeping a close eye on performance metrics.

Cutting Through the Clutter in AI Research: Discussions emphasized the distinction between publicly available research papers and the unrealistic expectation of OpenAI releasing fully trained proprietary models, due to their computational demands and proprietary elements.


HuggingFace Discord

Code Whispering with Moondream and FluentlyXL: Community contributions showcase Moondream 2 for batch processing and FluentlyXL v4, as well as Portuguese translations of HF's Audio course and a new MPI Codes repository for MPI development. An intelligence boost for LangChain and FinBERT's financial sentiment tuning were also discussed.

Babel Fish's Extended Family: The multilingual sphere expands with BLOOM supporting 55 languages and research on improving LLMs, exemplified by a curated list and the RARR approach for automatic attributions in text generation. Members are also keen on deploying models with Ray and assessing quality metrics for refined prompts.

Diffusion Model Mixology: In diffusion discussions, the community explores techniques for merging pipelines and partial diffusion methods, with a notable partial diffusion pull request for SD 1.5 found on GitHub. Overall, the topic of efficient and innovative model merging strategies garners attention.

Model Fine-Tuning Finesse: Best practices for fine-tuning models, like only adjusting classifier weights and customizing training loops, are debated, with a detailed guide on HuggingFace's Transformers and Keras. Members also discuss visual confirmations of models like Fluently-XL-v4 outperforming others on Instagram.

Seeking AI Mentors and Conversationalists: The community expresses a need for parquet converter-bots and more structured ways for members to provide peer support, like a possible #cv-study-group, while sharing knowledge and links for upskilling, such as a YouTube video on fine-tuning AI models and an exploration of graph ML's impact on LLMs.


LlamaIndex Discord


Latent Space Discord


Eleuther Discord

LLMs Translating Before Answering: Engineers debate Large Language Models (LLMs) processing multilingual inputs by potentially converting them to English first, referencing "Understanding Language Models by Fine-grained Language Identification". An important nuance for those looking to optimize multilingual LLM systems.

Lost Research Directions Evoke Nostalgia: A reflective exchange on understudied ML fields, such as adversarial robustness and domain-specific modeling, lamented due to the industry's overshadowing allure. Notably poignant for the career paths of researchers in the field.

Leakage Looms Over Benchmarks: Concerns in benchmark dataset leakage for LLMs stir conversation, emphasizing the challenges in gauging leaks and rectifying them. Two papers, one on leakage detection and another proposing new methods like fresh benchmark questions, fuel the discussion.

English as a Pivot in LLMs Proves Generative: llama models' findings suggest English as a pivot language is a sound strategy, potentially boosting those working on cross-model generalizability. Such replication adds weight to the approach for those developing multilingual LLMs.

Language Models Dream of Chess Mastery: A study involving a transformer trained solely on chess games achieves high performance, sans heuristics, as cited in a DeepMind paper. Demonstrates the scope of scale training for AI engineers interested in out-of-box model applications.

Grandmaster-Level Chess Without Search: A study using a transformer model trained on a dataset of 10 million chess games was brought up, demonstrating the model's high performance in chess without domain-specific enhancements or explicit search algorithms. The DeepMind paper indicates that training models at scale can lead to competitive levels of play without the approaches traditional chess engines use.


OpenAccess AI Collective (axolotl) Discord


OpenInterpreter Discord

Documentation Dilemma Resolved: Access to instructions for Ollama, Jan.ai, and Llamafile is improved with a direct link to the Open Interpreter local installation guide, emphasizing dolphin-mixtral configurations to streamline the setup process.

Performance Enhancements for Whisper RKNN: A notable 250% performance surge is achieved for Whisper RKNN on Rockchip RK3588 SBCs as shared in the rbrisita's GitHub branch, and there's an anticipation of upcoming LLM RKNN feature integrations.

AI Vtubing Enters Open Source Arena: The AI Vtuber community benefits from a pair of new resources: an AI Vtuber starter kit on GitHub, and an offline-ready, API-free Vtuber repository, with a live proof-of-concept showcased on YouTube.

Interactivity Extended to Mobile: Insight into hosting Open Interpreter on servers for broader access and setting up mobile-friendly, local models was shared, linking to specific Android device setup and running Open Interpreter locally.

Sound Choices in Speaker Selection: A discerning approach is underway to select the optimal speaker for an unnamed electronics project, promising future insights based on the integration and validation results.


OpenRouter (Alex Atallah) Discord

OpenRouter Battles Traffic Surge: OpenRouter grappled with higher-than-normal errors due to a traffic spike, with scaling efforts in progress to mitigate intermittent connectivity issues.

Money Moves: A proposal to integrate WeChat Pay and Alipay via Stripe was discussed, with the community aware of it requiring additional paperwork; meanwhile, suggestions to develop an app for smoother transactions using Google payment services were also floated.

Model Size Matters: The AI community showed keen interest in next-generation language models like LLaMA-3, with anticipation for potential releases by entities like Soliloquy, while recognizing the limitations tied to proprietary models.

Fine-Tuning Finesse: Engineers debated the risk of model dumbing post-fine-tuning without instruct datasets, agreeing that blending old and new data might safeguard against catastrophic forgetting.

Gemini Pro Troubleshooting: Technical solutions were shared for problems encountered with Gemini Pro messages, such as starting prompts with an "assistant" role to facilitate better interactions.


AI Stack Devs (Yoko Li) Discord

StoryDiffusion Crafted by Angry Penguin: StoryDiffusion sparks interest, engaging members with AI storytelling potential, following a link shared by angry.penguin.

AI Town Troubles and Tools: Disruptions from empty messages and strings of numbers in ai-town-discuss highlight tokenizer concerns; meanwhile, resources like @TheoMediaAI's AI simulation exploration and @cocktailpeanut's sqlite replay web app for AI Town catch attention.

Node Woes in Backend Development: Incorrect Node version causes stumbling blocks in local deployment of convex-local-backend; workaround involves switching to Node v18. A community-sourced issue was logged regarding a TypeError with .ts extension during setup.

Raspberry Pi Channel Piqued Interest: An expression of deep contemplation and a member's acknowledgment reveal that the ai-raspberry-pi channel meets certain members' specialized interests in AI development on small-scale hardware.

Cocktail Peanut Receives Undefined Kudos: A mysterious member praises cocktail peanut amid discussions but leaves the community guessing the work or breakthrough being referenced.


LAION Discord


LangChain AI Discord

Hackathon Alert: Build AI Products in 54 Hours for Cash: The BeeLoud hackathon, scheduled for May 10-12, invites participants to create AI innovations within 54 hours, with a prize pool of up to $25,000. For more details, see Build - BeeLoud.

LangChain and RAG Empower Email Crafting: LangChain's LangGraph Agents now leverage Retrieval-Augmented Generation (RAG) to enhance AI-assisted email drafting, promising both efficiency and quality improvements, as detailed in a Medium article.

Java Devs, Meet LangChain: A newly available langchain4j Java port of LangChain has been announced, broadening the scope for integrating AI applications across different platforms and languages. Interested engineers can explore langchain4j on GitHub.

Dragonfly Boosts LangChain's Performance: By integrating the Dragonfly in-memory data store with LangChain, developers can expect improved chatbot performance and context management which is explained with examples in their latest blog post.

Langserve Decoded: The langserve feedback endpoint clarification was provided, where an "OK" response merely indicates that feedback has been successfully submitted, but might still be rejected if the server deems it unauthenticated or invalid.


Interconnects (Nathan Lambert) Discord


Cohere Discord

PDF Search System Unearthed: A member proposed a search system for large PDF documents, discussing strategies including document summarization via LLMs, embedding generation for semantic search, and LLMs-based key information indexing.

Llama Tokenization Mysteries Revealed: Queries arose regarding the necessity of a beginning-of-string (<BOS_TOKEN>) when using the llama-cpp-python library with Command R+, with observations of its automatic inclusion during tokenization.

RAG Access with Cohere Confirmed: A user's question about the feasibility of using a free Cohere API key for RAG was answered, confirmation was given of its availability, albeit with rate limitations.

C4AI Command R+ Gets Quantized: Technical conversation unfolded around the C4AI Command R+ model, with a focus on its quantized variant, and varying system requirements for local implementation.

Code Interpreter SDK Takes the Stage: An announcement regarding the launch of the Code Interpreter SDK surfaced, alongside a discussion about its distinction in the context of pre-existing technologies.


Mozilla AI Discord


tinygrad (George Hotz) Discord


AI21 Labs (Jamba) Discord

Jamba-Instruct Is Live: AI21 Labs has launched Jamba-Instruct, a sophisticated instruction-tuned hybrid SSM-Transformer model, designed to enhance commercial application performance. The company highlights the model's capabilities in a recent Twitter announcement and a detailed blog post.

AI21 Labs Welcomes Feedback for Jamba-Instruct: AI21 Labs is inviting industry feedback for Jamba-Instruct and indicates their openness to discuss custom requirements, including context windows exceeding the initial 256K limit.

Reading Up on Jamba-Instruct: Engineers interested in the Jamba-Instruct model can gain a deeper understanding by reading the official blog post, which talks about its deployment for reliable commercial use and quality benchmarks.

Higher Context Windows on the Horizon: An AI21 Labs staff member has expressed their interest in exploring significantly larger context windows for Jamba-Instruct and has invited users to collaborate on this potential expansion to meet specific use scenarios.


Alignment Lab AI Discord


DiscoResearch Discord


Skunkworks AI Discord


Datasette - LLM (@SimonW) Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Unsloth AI (Daniel Han) ▷ #general (734 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (20 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (580 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (162 messages🔥🔥):

Link mentioned: Dog Awkward GIF - Dog Awkward Awkward dog - Discover & Share GIFs: Click to view the GIF


Stability.ai (Stable Diffusion) ▷ #general-chat (753 messages🔥🔥🔥):

Links mentioned:


CUDA MODE ▷ #general (3 messages):


CUDA MODE ▷ #triton (2 messages):

Link mentioned: [Frontend] Add tl.cast function. by jlebar · Pull Request #3813 · openai/triton: This resolves an inconsistency in Triton, that every other function on Tensors has an associated free function -- i.e. you can do x.foo and tl.foo(x).


CUDA MODE ▷ #cuda (6 messages):

Links mentioned:


CUDA MODE ▷ #torch (43 messages🔥):

Links mentioned:


CUDA MODE ▷ #algorithms (5 messages):


CUDA MODE ▷ #cool-links (4 messages):


CUDA MODE ▷ #torchao (2 messages):

Links mentioned:


CUDA MODE ▷ #off-topic (9 messages🔥):

Link mentioned: Let's build the GPT Tokenizer: The Tokenizer is a necessary and pervasive component of Large Language Models (LLMs), where it translates between strings and tokens (text chunks). Tokenizer...


CUDA MODE ▷ #triton-puzzles (1 messages):

srush1301: Hmm, yeah this description is wrong. I will update with a clearer version


CUDA MODE ▷ #hqq (4 messages):

Links mentioned:


CUDA MODE ▷ #llmdotc (644 messages🔥🔥🔥):

Links mentioned:


CUDA MODE ▷ #oneapi (1 messages):

neurondeep: also added intel on pytorch webpage


LM Studio ▷ #💬-general (350 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (159 messages🔥🔥):

Links mentioned:


LM Studio ▷ #announcements (2 messages):

Links mentioned:


LM Studio ▷ #⚙-configs-discussion (8 messages🔥):


LM Studio ▷ #🎛-hardware-discussion (4 messages):


LM Studio ▷ #🧪-beta-releases-chat (18 messages🔥):

Links mentioned:


LM Studio ▷ #amd-rocm-tech-preview (32 messages🔥):

Links mentioned:


LM Studio ▷ #model-announcements (1 messages):


LM Studio ▷ #🛠-dev-chat (69 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):


Perplexity AI ▷ #general (308 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (22 messages🔥):


Perplexity AI ▷ #pplx-api (41 messages🔥):

Links mentioned:


Nous Research AI ▷ #off-topic (15 messages🔥):


Nous Research AI ▷ #interesting-links (19 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (104 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (45 messages🔥):

Links mentioned:


Nous Research AI ▷ #bittensor-finetune-subnet (1 messages):


Nous Research AI ▷ #rag-dataset (1 messages):

felixultimaforeverromanempire: anyone know fo good free generic data sets?


Nous Research AI ▷ #world-sim (86 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #general (99 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (3 messages):


Modular (Mojo 🔥) ▷ #✍︱blog (2 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #announcements (1 messages):


Modular (Mojo 🔥) ▷ #ai (4 messages):


Modular (Mojo 🔥) ▷ #tech-news (2 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #🔥mojo (137 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (3 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-blogs-vids (3 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (1 messages):

soracc: Good idea


Modular (Mojo 🔥) ▷ #📰︱newsletter (1 messages):

Zapier: Modverse Weekly - Issue 32 https://www.modular.com/newsletters/modverse-weekly-32


Modular (Mojo 🔥) ▷ #nightly (10 messages🔥):

Links mentioned:


OpenAI ▷ #ai-discussions (224 messages🔥🔥):


OpenAI ▷ #gpt-4-discussions (23 messages🔥):


OpenAI ▷ #prompt-engineering (3 messages):


OpenAI ▷ #api-discussions (3 messages):


HuggingFace ▷ #announcements (2 messages):

Links mentioned:


HuggingFace ▷ #general (163 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (9 messages🔥):

Links mentioned:


HuggingFace ▷ #cool-finds (7 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (9 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (6 messages):

Links mentioned:


HuggingFace ▷ #computer-vision (7 messages):

Link mentioned: Welcome to the Community Computer Vision Course - Hugging Face Community Computer Vision Course: no description found


HuggingFace ▷ #NLP (5 messages):

Link mentioned: Paper page - RARR: Researching and Revising What Language Models Say, Using Language Models: no description found


HuggingFace ▷ #diffusion-discussions (12 messages🔥):

Link mentioned: Comparing huggingface:main...bghira:partial-diffusion-2 · huggingface/diffusers: 🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX. - Comparing huggingface:main...bghira:partial-diffusion-2 · huggingface/diffusers


LlamaIndex ▷ #blog (5 messages):

Link mentioned: Introspective Agents: Performing Tasks With Reflection - LlamaIndex: no description found


LlamaIndex ▷ #general (140 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):


Latent Space ▷ #ai-general-chat (25 messages🔥):

Links mentioned:


Latent Space ▷ #llm-paper-club-west (33 messages🔥):

Link mentioned: Notion – The all-in-one workspace for your notes, tasks, wikis, and databases.: A new tool that blends your everyday work apps into one. It's the all-in-one workspace for you and your team


Latent Space ▷ #ai-in-action-club (65 messages🔥🔥):

Links mentioned:


Eleuther ▷ #general (49 messages🔥):

Links mentioned:


Eleuther ▷ #research (54 messages🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (1 messages):

Link mentioned: Papers with Code - MATH Benchmark (Math Word Problem Solving): The current state-of-the-art on MATH is GPT-4-code model (CSV, w/ code, SC, k=16). See a full comparison of 109 papers with code.


Eleuther ▷ #interpretability-general (8 messages🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (2 messages):

Link mentioned: Paper page - Prometheus 2: An Open Source Language Model Specialized in Evaluating Other Language Models: no description found


OpenAccess AI Collective (axolotl) ▷ #general (40 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (19 messages🔥):

Link mentioned: FIX: TRL trainer preprocessing step was running in one process by ali-mosavian · Pull Request #1583 · OpenAccess-AI-Collective/axolotl: Description We weren't passing dataset_num_proc to TRL training config, thus the initial data preprocessing steps in the TRL trainer was running in one process only. Motivation and Context Speeds ...


OpenAccess AI Collective (axolotl) ▷ #general-help (7 messages):

Link mentioned: llama.cpp/scripts/convert-gg.sh at master · ggerganov/llama.cpp: LLM inference in C/C++. Contribute to ggerganov/llama.cpp development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (15 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (32 messages🔥):

Links mentioned:


OpenInterpreter ▷ #general (63 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (10 messages🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (2 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #announcements (1 messages):


OpenRouter (Alex Atallah) ▷ #general (72 messages🔥🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #app-showcase (1 messages):

angry.penguin: https://huggingface.co/spaces/YupengZhou/StoryDiffusion


AI Stack Devs (Yoko Li) ▷ #ai-town-discuss (15 messages🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #ai-town-dev (26 messages🔥):

Link mentioned: TypeError [ERR_UNKNOWN_FILE_EXTENSION]: Unknown file extension ".ts" for /app/npm-packages/convex/src/cli/index.ts · Issue #1 · get-convex/convex-backend: I ran the steps in the prerequisites then got this when running just run-local-backend Error: Failed to run convex deploy: TypeError [ERR_UNKNOWN_FILE_EXTENSION]: Unknown file extension ".ts&quot...


AI Stack Devs (Yoko Li) ▷ #ai-raspberry-pi (2 messages):


LAION ▷ #general (26 messages🔥):

Link mentioned: GitHub - wesbz/SoundStream: This repository is an implementation of this article: https://arxiv.org/pdf/2107.03312.pdf: This repository is an implementation of this article: https://arxiv.org/pdf/2107.03312.pdf - wesbz/SoundStream


LAION ▷ #research (7 messages):


LangChain AI ▷ #general (26 messages🔥):

Links mentioned:


LangChain AI ▷ #langserve (1 messages):


LangChain AI ▷ #share-your-work (3 messages):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (13 messages🔥):

Link mentioned: Proximal Policy Optimization — Spinning Up documentation: no description found


Interconnects (Nathan Lambert) ▷ #ml-drama (4 messages):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (5 messages):

Link mentioned: Tweet from lmsys.org (@lmsysorg): Exciting news -- we're thrilled to announce that LMSYS + @kaggle are launching a human preference prediction competition with $100,000 in prizes! Your challenge is to predict which responses user...


Interconnects (Nathan Lambert) ▷ #rl (8 messages🔥):


Cohere ▷ #general (21 messages🔥):

Links mentioned:


Mozilla AI ▷ #llamafile (19 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #general (4 messages):


tinygrad (George Hotz) ▷ #learn-tinygrad (13 messages🔥):


AI21 Labs (Jamba) ▷ #announcements (1 messages):

Link mentioned: Built for the Enterprise: Introducing AI21’s Jamba-Instruct Model: An instruction-tuned version of our hybrid SSM-Transformer Jamba model, Jamba-Instruct is built for reliable commercial use, with best-in-class quality and performance.


AI21 Labs (Jamba) ▷ #jamba (4 messages):


Alignment Lab AI ▷ #general-chat (2 messages):


DiscoResearch ▷ #general (2 messages):

Links mentioned:


Skunkworks AI ▷ #off-topic (1 messages):


Datasette - LLM (@SimonW) ▷ #llm (1 messages):