Frozen AI News archive

AdamW -> AaronD?

**Aaron Defazio** is gaining attention for proposing a potential tuning-free replacement of the long-standing **Adam optimizer**, showing promising experimental results across classic machine learning benchmarks like ImageNet ResNet-50 and CIFAR-10/100. On Reddit, **Claude 3 Opus** has surpassed all **OpenAI** models on the LMSys leaderboard, while a user pretrained a **LLaMA-based 300M** model outperforming **bert-large** on language modeling tasks with a modest budget. The new **MambaMixer** architecture demonstrates promising results in vision and time series forecasting. In image generation, **Stable Diffusion 1.5** with LoRAs achieves realistic outputs, and the **WDXL** release showcases impressive capabilities. AI applications include an AI-generated Nike spec ad and a chatbot built with OpenAI models that may resist prompt injections. OpenAI is reportedly planning a ban wave targeting policy violators and jailbreak users. *"The high alpha seems to come from Aaron Defazio,"* highlighting his impactful work in optimizer research.

Canonical issue URL

It's a quiet Easter weekend and April Fools' is making it harder than normal to sift signal from noise (our contribution here). We do recommend sifting through Sequoia Ascent's playlist, if you're not close to each speaker's work (for example Andrew Ng mostly repeated the writeup we covered last week), which is now fully released.

Over in Twitter land, the high alpha seems to come from Aaron Defazio, which several of our AI High Signal follows highlighted as the "new LK-99" for engaging, "impossible" work in public. What's at stake: a potential tuning-free replacement of the very long lived Adam optimizer, and experimental results are currently showing learning at a Pareto frontier in a single run for basically every classic machine learning benchmark (ImageNet ResNet-50, CIFAR-10/100, MLCommons AlgoPerf):

image.png

He's writing the paper now, and many "better optimizers" have come and gone, but he is well aware of the literature and going for it. We'll see soon enough in a matter of months.


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence. Comment crawling still not implemented but coming soon.

AI Models and Performance

Stable Diffusion and Image Generation

AI Applications and Demos

AI Ethics and Policies

Memes and Humor

AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

AI Capabilities and Limitations

AI Development and Deployment

AI Applications and Use Cases

AI Ethics and Safety

Memes and Humor


AI Discords

A summary of Summaries of Summaries


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord


Perplexity AI Discord

Claude's Classroom Conundrum: The Claude 3 Opus model exhibits inconsistent performance with questions related to identifying remaining books in a room, with some models not producing correct answers despite several adjustments to prompts.

AI Model Melee: Engineers discussed AI model benchmarks, focusing on the comparative performance of Haiku, Gemini 1.5 Pro, and Claude Opus. The conversations highlighted differing strengths and functionalities but did not lean towards consensus on a superior model.

Pondering Partnerships and API Puzzles: For partnership interests with Perplexity, engineers are instructed to email [email protected], and seeking details about the API's source citation feature can be directed to Perplexity's Typeform. Additionally, "pplx-70b-online" model support is deprecated, and the alias concerns are culminating in suggesting an update to Perplexity's Supported Models documentation.

Credit Where Credit's Due: Reports of issues with credit purchases on Perplexity surfaced, hinting at potential complications with transaction systems or third-party security features like those implemented by Stripe. Member discussion advised for situational troubleshooting and inquired about further inspection.

Search Spectacles and Query Quirks: Engineers displayed a broad array of interests from Bohmian mechanics to Hyperloop through shared queries on Perplexity AI, but user-contributed informational threads lacked documentation support for their extendibility and shareability.


Unsloth AI (Daniel Han) Discord

Snapdragon Makes Waves: Qualcomm's Snapdragon Elite X Arm chip has impressed engineers with its 45 TOPs performance, leading to discussions about its cost-efficiency and comparisons with other chips like the Tesla T4's 65 TFLOPs of float16. The excitement was fueled by a YouTube video detailing the chip's benchmarks.

Model Training Optimized with Unsloth: Fine-tuning Mistral models with Unsloth AI can encounter dependency issues, but the Unsloth GitHub repository offers a Docker solution and a manual GGUF guide. Moreover, discussions suggest single GPU training is possible by setting os.environ["CUDA_VISIBLE_DEVICES"], although multi-GPU support is a potential future development.

AI Hardware Announcements Catch Attention: Intellifusion's new AI processor could be a game-changer for inference operations due to its cost-effectiveness, raising curiosity about its potential in training scenarios. Details can be found on Tom's Hardware.

Fine-Tuning Techniques Under Scrutiny: Engineers debate fine-tuning methods like QLora 4bit versus SFT/pretraining, discussing how the quantization process might affect performance. There's also talk about the paradox of dataset size in model training, where quality, not just quantity, determines the effectiveness.

ORPO Integration Sparks Commendation: The Unsloth + ORPO (Orthogonal Projection for Language Models Alignment) combination has been implemented effectively in LLaMA Factory, according to a paper on arXiv. The AI community shared success stories and optimizations, acknowledging particular efficacy in training with limited data samples.


Nous Research AI Discord

StyleGAN Gets a Fashion Makeover: When training StyleGAN2-ada with various fashion images, users inquired about the need for script modifications but did not mention outcomes or specify details on solutions.

Learners Take Flight with ML/AI Courses: For those charting a course into machine learning, particularly from other fields like aerospace, the community recommended starting with the foundational fastai courses, and moving toward specialized courses like the Hugging Face NLP course for a deep dive into language models and transformers.

Microsoft's Ternary LLM Paper Replicated: Results from a Microsoft paper on ternary Large Language Models, especially concerning the 3 billion parameter models at 100 billion operations, have been successfully replicated, as evidenced by the model bitnet_b1_58-3B on Hugging Face.

Nous Research Amplifies LLM Discussion with a Tweet: Nous Research fueled the conversation around LLMs with a twitter post, though the content of the announcement was not detailed in the messages.

Privacy Detection Dilemma: Hermes mistral 7b's difficulties in anonymizing PII sparked debate on how to enhance the model's capabilities. There was a mention of upcoming data integrations by NousResearch and models that may aid in improvement, such as open-llama-3b-v2-pii-transform.

Opinions Split on RAG Configurations: The community discussed the merits of using a single large RAG versus multiple specialized RAGs. While specific approaches or results were not mentioned, the conversation touched on the importance of metadata and the idea of integrating RAG with other tools to bolster functionality.

OpenSim Engages Philosophical and Practical Domains: Users debated the economic aspects of token output costs in LLM apps, explored the concept of "Hyperstition" within AI interactivity, and expressed desire for new features in WorldSim, like saving chat sessions with URLs for sharing.


LM Studio Discord

JSON Outputs Draw Developer Attention: AI engineers show interest in LMStudio's JSON output format for the development of practical applications. Seamless integration with langchain has been reported, making the process incredibly efficient.

Plugin Possibilities Percolate in LM Studio: The community calls for plugin support within LM Studio for expandability, while feature requests such as a Unified Settings Menu and Keyboard Shortcuts indicate a desire for a more customizable and efficient user interface.

Apple Silicon Users Adapt and Overcome: LLM users report challenges when running models on Apple Silicon M1 Macs, offering shared solutions like shutting down other apps to free up memory and exploring LoRA adaptation interfaces.

GPUs Under the Microscope after LM Studio Update: Post-update GPU issues with LM Studio, including disappearing GPU Acceleration options and unrecognized VRAM, catalyzes conversations around navigating hardware compatibility, multi-GPU setups, and memory usage.

Remote GPU Support Requested for Power Users: AI Engineers express interest in remote GPU support for LM Studio, noting parallels to services allowing remote gaming, and ask for open-source initiatives considering the community's emphasis on privacy and security.


OpenAI Discord

Voice Tech Marches On: OpenAI's Voice Engine can now generate natural speech just from text and a 15-second voice sample, though they're proceeding with caution to mitigate misuse risks. Simultaneously, OpenAI removed the signup barrier for ChatGPT, allowing instant AI engagement worldwide.

Prompt Engineering Reveals Tech Quirks: Some members experience difficulties when transferring LaTeX equations from ChatGPT to Microsoft Word, whilst others discussed nuanced AI approaches like meta-prompting and observed unusual behaviors in roleplaying scenarios with the gpt-4-0125-preview model.

VoiceCraft's New Frontier: VoiceCraft's GitHub repo and its accompanying demo highlight its speech editing and text-to-speech prowess, igniting discussions around the ethics of voice cloning and potential for misuse.

Choosing the Right AI Tools for Business Insights: In the tech community, there's uncertainty about whether to use the completion API or the assistant API for tasks like summarizing business data and generating quizzes, with ChatGPT format controls suggested as a deciding factor (API context management).

Model Mix-Up Clarified: Discussions clarified that ChatGPT is not an AI model itself, but an application that uses GPT models. Additionally, debates blossomed around the usage and limitations of Custom GPT and how developers might interface with GPT API directly for projects like automated video content management.


Eleuther Discord


LAION Discord

DBRX Base Hits Home Run: A non-gated re-upload of the DBRX Base model, notable for its mixture-of-experts architecture, reiterates the community's push for open weights and ungatekeeped access. The original models can be explored on Hugging Face.

Euler Method Proves Its Worth: Anecdotal evidence suggests that using the euler ancestral method optimizes results on terminus, backed by amusing examples of precise Chinese translations.

AI's Music Maestros Dissect Suno: Discussing AI music generation tools, particularly Suno's v2 vs v3, the community shared concerns about noise in voice generation and the potential leap v4 could bring.

Voice Synthesis Under the Microscope: Voices in the guild raised concerns about OpenAI's Voice Engine potentially eclipsing Voicecraft, while pondering on the strategic play involved and the potential repercussions on the US Elections.

Stochastic Rounding as a Training Booster: Engineers are looking into stochastic rounding techniques for training AI, presenting nestordemeure/stochastorch as a promising Pytorch implementation to try out.

Transforming Diffusion with Transformers: Conversations trend towards replacing UNETs with transformers in diffusion, with a key research paper guiding the way.

Decoding UNET Mysteries: A member breaks down UNETs as a tool for downsampling and then reconstructing images, which could help with discarding superfluous details in models.

Qwen1.5-MoE-A2.7B Raises Expectations: A buzz surrounds Qwen1.5-MoE-A2.7B, a model challenging larger counterparts with just 2.7 billion activated parameters, detailed across various platforms like GitHub, Hugging Face, and Demo.

V-JEPA Sets the Stage for Video Lava: The community examines V-JEPA's potential in enhancing video Lava, with GitHub resources at hand (V-JEPA GitHub) to broaden the data prep and training terrain.

Diffusion and Embedding Win Big With New Techniques: A paper discussing a new diffusion loss function offers a glimmer of hope against data corruption (paper link), while Gecko's approach in text embedding might be a game changer in accelerating training (Gecko paper link).


HuggingFace Discord

Blazing 1-Bit Model Weights Introduced: Hugging Face released 1.38 bit quantized model weights for large language models (LLMs), signaling strides towards more efficient AI models. Interested engineers can scrutinize the model here.

PAG Refines Samples Without Sacrificing Diversity: The utility of Perturbed-Attention Guidance (PAG) was showcased, which unlike Classifier-Free Guidance (CFG), doesn't reduce diversity when improving sample quality. The usage ratio of CFG 4.5 and PAG between 3.0 to 7.0 was recommended for enhanced results, based on research.

Real-Time Diffusion Now a Reality: The use of 1 step diffusion enabling 30fps generation at 800x800 resolution has been achieved using sdxl-turbo. For those intrigued by the seamless transitions, a Twitter thread with video snippets showcases the evolution of real-time video generation.

In Search of Tokenizer-Compatible Models: An inquiry was made about how to identify suitable assistant models for model.generate by tokenizer, with discussions pointing to the Hugging Face Hub API for potential solutions. Additionally, approaches to extracting domain-specific entities were explored, recommending leveraging pre-trained models or considering independent training for 20k documents.

Melding AI into Musical Alchemy: Discussions included the challenge of AI-generated music, blending artists' voices to create harmonies like those of Little Mix, highlighted by the intricacy of key adjustments. Other technical endeavors shared in the community involved the creation of Terraform provider for Hugging Face Spaces and the introduction of OneMix, a Remix-based SaaS boilerplate.


OpenInterpreter Discord

Getting Chatty with Open Interpreter: A video titled "Open Interpreter Advanced Experimentation - Part 2" reveals new experiments with the OpenInterpreter, demonstrating the platform's growing capabilities for technical innovation.

AI as a Sidekick: The Fabric project on GitHub, an open-source initiative, offers a modular framework designed to augment human skills with AI, utilizing a community-driven collection of AI prompts adaptable for various challenges.

Audio Issues Crackdown: In the OpenInterpreter community, an audio playback problem on MacOS involving ffmpeg was teased out, and solutions involving multiple commands were proposed to mitigate the trouble experienced after a response was generated.

Windows Walkthrough Update: The onboarding experience for Windows users working with the OpenInterpreter 01 client has seen enhancements with new pull requests (#192, #203) aimed at resolving compatibility challenges and improving the setup documentation.

Fine-Tuning for O1 Light Fabricators: Makers of the O1 Light are advised to upscale 3D printing files to 119.67% for fitting the components properly, signaling a community-driven focus on custom hardware optimization.


tinygrad (George Hotz) Discord

Intel Arc Meets Optimized Performance: Efforts to optimize transformers for Intel Arc GPUs identified the underperformance of IPEx library, as it wasn't employing fp16 effectively. Solutions involving PyTorch JIT yielded significant performance improvements for stable diffusion tasks.

Open Call: AMD GEMM Optimization Wanted: A $200 bounty is up for grabs for writing optimized GEMM code for AMD 7900XTX GPUs with instructions including HIP C++ integration. However, the endeavor is hampered by script issues involving missing modules and library paths.

Amendments Afoot in Tinygrad: Discussions are ongoing within the Tinygrad repository, pinpointing issues with failing tests and missing functionalities. One suggestion involves examining the shapetracker and uopt optimization to enable contributions even from non-GPU laptop setups.

AMD's Driver Saga: Conversations centered on AMD driver instability, calling for an open-source approach for firmware and suggesting various GPU reset methods like BACO and PSP mode2. A GitHub discussion thread expressed frustration over full reset limitations and ineffective communication channels with AMD.

Fusion and Views in Shape Manipulation: The technicalities of kernel fusion and shape manipulation in Tinygrad were broached, with a shared link on notes providing possible optimizations. An issue regarding memory layout complexities and uneven stride presentation was pinpointed and addressed in a recent pull request.


LlamaIndex Discord

Phorm.ai Teams Up with LlamaIndex: Phorm.ai integration provides TypeScript and Python support within LlamaIndex Discord, enabling queries and answers through "@-mention" within specific channels.

Learn RAFT, Don't Be Daft: A LlamaIndex webinar with RAFT co-authors, Tianjun Zhang and Shishir Patil, promises insights into domain-specific LLM fine-tuning, set for Thursday, 9am PT with sign-ups at lu.ma.

RAG Revolution Deep Dives: Guides and tutorials detail new strategies for enhancing Retrieval Augmented Generation, including self-reflective systems, integration with LlamaParse, and the importance of re-ranking, discussed across various platforms such as Twitter and YouTube.

LLM Research Made Accessible: A GitHub repository by shure-dev aims to consolidate impactful research papers on Large Language Models, serving as a comprehensive resource for AI enthusiasts.

Tackling LlamaIndex Document Dilemmas: Community members address complex issues, from managing oversized data chunks with SemanticSplitterNodeParser to improving outdated documentation, sharing best practices and solutions such as a helpful Colab tutorial.


OpenRouter (Alex Atallah) Discord

Novus Chat Jets onto OpenRouter: Novus Chat, a fresh platform integrating OpenRouter models, is creating buzz with free access to lowcost models and an invitation extended to AI enthusiasts to join its development discussions.

Ranking Reveal Creates Model Buzz: OpenRouter has introduced App Rankings for Models, allowing a glance at the top public apps that utilize specific models, with the Apps tab for each model revealing token stats; see Claude 3 Opus App Rankings as an example.

OpenRouter Sparks Chatbot API Conversation: Technical exchanges within the community are intensely focused on utilizing OpenRouter's APIs, embracing strategies for enhancing context retention and error handling while comparing functionalities between Assistant Message and Chat Completion approaches.

ClaudeAI Beta: Now Self-Moderating: OpenRouter's beta offering of Anthropic's Claude 3 Opus introduces a self-moderated version aiming to mitigate false positives, promising nuanced performance in sensitive contexts, as detailed in Anthropic's announcements.

Downtime Drama and Resolution: Recent Midnight Rose and Pysfighter2 models faced temporary downtime which was promptly resolved, whereas Coinbase payment issues were also flagged with assurance of a fix in progress, maintaining active wallet connections.


Latent Space Discord

Bold Climb Beyond the Binary: Discussions on 1-bit LLMs, referred to as "1.58 bits per parameter" due to ternary quantization, revealed skepticism about marketing hype vs technical precision. Community engagement included sharing of relevant papers and anecdotal reproductions of key findings.

Cross-Continental Voice Model Win: Voicecraft's new open-source speech model has outperformed ElevenLabs, with members sharing GitHub weights and positive experiences.

Bye-Bye, Boss: Stability AI's CEO stepping down made waves, with the community dissecting interviews such as Diamandis’s YouTube piece and speculating about company futures and the tech executive landscape.

Local LLMs Conquer Complexity: Discussions in the AI-In-Action club took a deep dive into the efficiency of local LLM function calling, with contrasting opinions on which methods lead the pack, outlines vs instructor, and exploration of mechanisms like regular expressions in text generation.

Anticipation for AI Agendas: Upcoming sessions about UI/UX patterns and RAG architectures stirred up interest, backed by a community-driven schedule. Sharing of resources and facilitation plans spotlighted the proactive preparation for future tech talks.


CUDA MODE Discord


OpenAccess AI Collective (axolotl) Discord


Modular (Mojo 🔥) Discord

Open-Sourcing Mojo: A Community Effort: The excitement about Modular's open-sourced Mojo standard library is palpable; however, there are frustrations due to limitations on non-internal/commercial applications and the lack of essential features like string sorting. Installation challenges on Linux Mint and desires for better profiling tools were also voiced, with official support confirmed for Ubuntu, MacOS, and WSL2 and guides provided for setup and local stdlib building.

Mojo's Threading Quest and Docs Expansion: Technical discussions on Mojo's multithreading capabilities highlighted the use of OpenMP for multi-core CPU enhancements and debates about external_call() functionality improvements. MLIR's syntax documentation is being improved to be more user-friendly, and there's a call for more detailed contributions.

Library and Language Enhancements: Several Mojo libraries have been updated to version 24.2, while the anticipation for a more evolved Reference component and better C/C++ interop in Mojo is strong. A new logging library, Stump, is introduced for the community to test.

Tackling Code Challenges: Performance and benchmarking channels discussed the one billion row challenge, noting the absence of certain standard library features and the need for improved memory allocation understanding. Meanwhile, the matmul.mojo example raised concerns over rounding errors and data type inconsistencies.

MAX Makes Moves into Triton: MAX Serving successfully operates as a backend for the Triton Inference Server, and the team is eager to support users in their migration efforts, emphasizing an easy transition and promising enhanced pipeline optimization.


Interconnects (Nathan Lambert) Discord

Benchmarks Set Stage for AI Bravado: The lm-sys released an advanced Arena-Hard benchmark aiming to better evaluate language models through intricate user queries. Debates arose around the potential biases in judging, especially exemplifying GPT-4's self-preference and its significant performance over Claude on Arena-Hard.

Token Talk Takes Theoretical Turn: Conversations pivoted to evaluating the informational content of tokens, with mutual information cited as a possible measure. Discussions framed this analysis against repeng strategies and Typicality methods, the latter detailed in an information theory-based paper.

Innovation Amidst The Hiring Game: Discussions revealed Stability AI actively recruiting top researchers, while Nathan Lambert described Synth Labs' non-traditional startup strategy, introducing ground-breaking papers preeminent to their product launches.

1-Bit Wonders: NousResearch validated Bitnet's claims through a 1B model trained on the Dolma dataset, released on Hugging Face, igniting discussions on the novelty and technicalities of 1-bit training.

sDPO Steps Up in RL: Shared insights unveiled stepwise DPO (sDPO) through a new paper, a technique that could democratize performance gains in model training, aligning models closely with human preferences without heavy financial backing.

Preserving Alignment Almanac: Nathan Lambert announced an initiative to document and discuss the evolution of open alignment techniques post-ChatGPT. Contributions such as an overview of various replicating models and considerations on preference optimization methods glean insight into the historical growth of the field, documented in Lambert's Notion Notes.


AI21 Labs (Jamba) Discord


LangChain AI Discord


Mozilla AI Discord


Datasette - LLM (@SimonW) Discord


DiscoResearch Discord


Skunkworks AI Discord

AI21's Jamba Jumps into the Fray: The Jamba model by AI21 has been shared within the Skunkworks AI community, touting enhancements in SSM-Transformer design for large language models.

Databricks' DBRX LLM Claims the Crown: Databricks' general-purpose large language model DBRX supposedly establishes new highs on multiple benchmarks, according to a shared video in the Skunkworks AI community.

Tackling Catastrophic Forgetting in Class-Incremental Learning: Research indicates that adapter tuning might be the key to combating catastrophic forgetting in CIL, employing feature sampling and prototype semantic shift analysis. The study is accessible through this arXiv link.

Closing the Gap Between Open-source and Commercial LLMs: A novel paper discusses methodologies aimed at empowering open-source LLMs to close the performance gap with their commercial counterparts, with strategies focused on 7B and 13B LLM enhancements. The paper's detailed insights are available here.


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ▷ #general-chat (980 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #general (915 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (36 messages🔥):


Perplexity AI ▷ #pplx-api (41 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (549 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (24 messages🔥):

Link mentioned: Now we know the SCORE | X Elite: Qualcomm's new Snapdragon X Elite benchmarks are out! Dive into the evolving ARM-based processor landscape, the promising performance of the Snapdragon X Eli...


Unsloth AI (Daniel Han) ▷ #help (461 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (7 messages):

Link mentioned: no title found: no description found


Unsloth AI (Daniel Han) ▷ #suggestions (6 messages):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #unsloth (1 messages):

Link mentioned: Support Unsloth AI on Ko-fi! ❤️. ko-fi.com/unsloth: Support Unsloth AI On Ko-fi. Ko-fi lets you support the people and causes you love with small donations


Nous Research AI ▷ #off-topic (19 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (8 messages🔥):

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

teknium: https://twitter.com/NousResearch/status/1773923241268003052


Nous Research AI ▷ #general (233 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (64 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #project-obsidian (1 messages):

Link mentioned: Sayali9141/traffic_signal_images · Datasets at Hugging Face: no description found


Nous Research AI ▷ #rag-dataset (46 messages🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (176 messages🔥🔥):

Links mentioned:


LM Studio ▷ #💬-general (285 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (50 messages🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (14 messages🔥):

Links mentioned:


LM Studio ▷ #🎛-hardware-discussion (90 messages🔥🔥):

Links mentioned:


LM Studio ▷ #langchain (4 messages):


LM Studio ▷ #amd-rocm-tech-preview (90 messages🔥🔥):


LM Studio ▷ #crew-ai (1 messages):


OpenAI ▷ #annnouncements (2 messages):

Links mentioned:


OpenAI ▷ #ai-discussions (98 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (31 messages🔥):


OpenAI ▷ #prompt-engineering (167 messages🔥🔥):

Link mentioned: Terms of use: no description found


OpenAI ▷ #api-discussions (167 messages🔥🔥):

Link mentioned: Terms of use: no description found


OpenAI ▷ #api-projects (3 messages):


Eleuther ▷ #general (260 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (169 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (4 messages):


Eleuther ▷ #interpretability-general (3 messages):

Links mentioned:


Eleuther ▷ #lm-thunderdome (12 messages🔥):

Links mentioned:


Eleuther ▷ #gpt-neox-dev (3 messages):


LAION ▷ #general (366 messages🔥🔥):

Links mentioned:


LAION ▷ #research (42 messages🔥):

Links mentioned:


HuggingFace ▷ #general (225 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (7 messages):


HuggingFace ▷ #cool-finds (12 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (41 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (40 messages🔥):

Links mentioned:


HuggingFace ▷ #computer-vision (19 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (9 messages🔥):

Link mentioned: Evaluate Retrieval Augmented Generation (RAG) Systems: Retrieval Augmented Generation is a powerful framework which improves the quality of responses that you get from LLMs. But if you want to create RAG systems ...


HuggingFace ▷ #diffusion-discussions (8 messages🔥):


OpenInterpreter ▷ #general (136 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (146 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (2 messages):

Links mentioned:


tinygrad (George Hotz) ▷ #general (251 messages🔥🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (16 messages🔥):

Links mentioned:


LlamaIndex ▷ #announcements (2 messages):

Links mentioned:


LlamaIndex ▷ #blog (10 messages🔥):

Link mentioned: no title found: no description found


LlamaIndex ▷ #general (218 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (4 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (229 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-general-chat (118 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-in-action-club (106 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #general (21 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (41 messages🔥):

Links mentioned:


CUDA MODE ▷ #cuda (14 messages🔥):


CUDA MODE ▷ #torch (3 messages):

Links mentioned:


CUDA MODE ▷ #announcements (1 messages):


CUDA MODE ▷ #algorithms (3 messages):

Link mentioned: Tweet from Aaron Defazio (@aaron_defazio): Update: Hold onto your hats, more results coming in! My new optimization approach demolishes a tuned Adam on DLRM.


CUDA MODE ▷ #beginner (9 messages🔥):

Link mentioned: Lightning AI | Turn ideas into AI, Lightning fast: The all-in-one platform for AI development. Code together. Prototype. Train. Scale. Serve. From your browser - with zero setup. From the creators of PyTorch Lightning.


CUDA MODE ▷ #pmpp-book (4 messages):


CUDA MODE ▷ #youtube-recordings (3 messages):

Link mentioned: Lecture 12: Flash Attention: no description found


CUDA MODE ▷ #torchao (1 messages):

Link mentioned: GaLore and fused kernel prototypes by jeromeku · Pull Request #95 · pytorch-labs/ao: Prototype Kernels and Utils Currently: GaLore Initial implementation of fused kernels for GaLore memory efficient training. TODO: triton Composable triton kernels for quantized training and ...


CUDA MODE ▷ #ring-attention (90 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #off-topic (16 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton-puzzles (10 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #general (130 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (45 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (27 messages🔥):

Link mentioned: GitHub - monk1337/auto-ollama: run ollama & gguf easily with a single command: run ollama & gguf easily with a single command. Contribute to monk1337/auto-ollama development by creating an account on GitHub.


Modular (Mojo 🔥) ▷ #general (30 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (4 messages):


Modular (Mojo 🔥) ▷ #📺︱youtube (1 messages):

Link mentioned: Modular Community Livestream - New in MAX 24.2: MAX 24.2 is now available! Join us on our upcoming livestream as we discuss everything new in MAX - open sourcing Mojo standard library, MAX Engine support f...


Modular (Mojo 🔥) ▷ #🔥mojo (76 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (11 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-blogs-vids (1 messages):

Link mentioned: Use locally built standard library in Mojo: Mojo standard library (stdlib) was open-sourced yesterday. It is exciting that the community can now contribute directly to the codebase. After spending some time with the stdlib repository, I want to...


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (3 messages):

Link mentioned: Issues · modularml/mojo: The Mojo Programming Language. Contribute to modularml/mojo development by creating an account on GitHub.


Modular (Mojo 🔥) ▷ #⚡serving (3 messages):

Link mentioned: Get started with MAX Serving | Modular Docs: A walkthrough showing how to try MAX Serving on your local system.


Interconnects (Nathan Lambert) ▷ #news (49 messages🔥):

Link mentioned: GitHub - lm-sys/arena-hard: Arena-Hard benchmark: Arena-Hard benchmark. Contribute to lm-sys/arena-hard development by creating an account on GitHub.


Interconnects (Nathan Lambert) ▷ #ml-questions (7 messages):

Link mentioned: Locally Typical Sampling: Today's probabilistic language generators fall short when it comes to producing coherent and fluent text despite the fact that the underlying models perform well under standard metrics, e.g., perp...


Interconnects (Nathan Lambert) ▷ #ml-drama (12 messages🔥):


Interconnects (Nathan Lambert) ▷ #random (2 messages):

Link mentioned: Tweet from Nous Research (@NousResearch): We are releasing our first step in validating and independently confirming the claims of the Bitnet paper, a 1B model trained on the first 60B tokens of the Dolma dataset. Comparisons made on the @we...


Interconnects (Nathan Lambert) ▷ #rl (2 messages):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #rlhf (8 messages🔥):

Link mentioned: sDPO: Don't Use Your Data All at Once: As development of large language models (LLM) progresses, aligning them with human preferences has become increasingly important. We propose stepwise DPO (sDPO), an extension of the recently populariz...


Interconnects (Nathan Lambert) ▷ #sp2024-history-of-open-alignment (24 messages🔥):

Links mentioned:


AI21 Labs (Jamba) ▷ #jamba (16 messages🔥):


AI21 Labs (Jamba) ▷ #general-chat (51 messages🔥):

Link mentioned: no title found: no description found


LangChain AI ▷ #general (34 messages🔥):

Links mentioned:


LangChain AI ▷ #langchain-templates (1 messages):

blackice9833: free nudes ♥️ https://discord.gg/bestnudes @everyone @here


LangChain AI ▷ #share-your-work (14 messages🔥):

Links mentioned:


LangChain AI ▷ #tutorials (2 messages):

Link mentioned: Langchain + Qdrant Local | Server (Docker) | Cloud | Groq | Tutorial: Do you want to learn a production grade vector database for your Langchain applications? Let's delve into the world of vector databases with Qdrant. Qdrant i...


Mozilla AI ▷ #llamafile (24 messages🔥):

Link mentioned: Release llamafile v0.7 · Mozilla-Ocho/llamafile: llamafile lets you distribute and run LLMs with a single file This release improves the performance and accuracy of both CPU and GPU computations in addition to security. tinyBLAS now gives outpu...


Datasette - LLM (@SimonW) ▷ #llm (9 messages🔥):


DiscoResearch ▷ #general (7 messages):

Links mentioned:


DiscoResearch ▷ #discolm_german (2 messages):

Link mentioned: Compare Translations - a Hugging Face Space by cstr: no description found


Skunkworks AI ▷ #papers (2 messages):

Links mentioned:


Skunkworks AI ▷ #off-topic (2 messages):

Links mentioned: