a quiet day.
AI News for 5/7/2025-5/8/2025. We checked 9 subreddits, 449 Twitters and 29 Discords (215 channels, and 3981 messages) for you. Estimated reading time saved (at 200wpm): 396 minutes. Our new website is now up with full metadata search and beautiful vibe coded presentation of all past issues. See https://news.smol.ai/ for the full news breakdowns and give us feedback on @smol_ai!
OpenAI launched both Reinforcement Finetuning and Deep Research on GitHub repos, which many are comparing to Cognitionâs DeepWiki.
But it is a quiet day otherwise.
AI Twitter Recap
Models, Benchmarks, and Performance
- Nvidiaâs Open Code Reasoning models: @reach_vb announced that NVIDIA has open-sourced Open Code Reasoning models (32B, 14B, and 7B), which are Apache 2.0 licensed, beat O3 mini & O1 (low) on LiveCodeBench, and are backed by OCR dataset. The models are reported to be 30% token efficient compared to other reasoning models and work with llama.cpp, vLLM, transformers, and TGI.
- Mistral Medium 3 performance: @ArtificialAnlys provided independent evaluations of Mistral Medium 3, noting it rivals Llama 4 Maverick, Gemini 2.0 Flash, and Claude 3.7 Sonnet in the leading non-reasoning models, with substantial gains in coding and mathematical reasoning. It is priced at $0.4/$2 per 1M Input/Output tokens, which is a significant decrease compared to Mistral Large 2. However, @scaling01 noted that Mistral is no longer open-source, lacking information on model size.
- Gemini 2.5 Pro coding abilities: @Google announced that Gemini 2.5 Pro is their most intelligent model yet and is better at coding from a simple prompt.
- Gemini 2.5 Flash cost increase: @ArtificialAnlys reported that Googleâs Gemini 2.5 Flash costs 150x more than Gemini 2.0 Flash to run the Artificial Analysis Intelligence Index. This increase is driven by a 9x more expensive output tokens and 17x higher token usage across their evals.
- Absolute Zero Reasoner (AZR): @iScienceLuvr highlights Absolute Zero: Reinforced Self-play Reasoning with Zero Data, noting that AZR self-evolves its training curriculum and reasoning ability by using a code executor to validate proposed code reasoning tasks and verify answers, achieving overall SOTA performance on coding and mathematical reasoning tasks without external data. @arankomatsuzaki shared the same information and links to the project page and repo.
- X-REASONER vision-language model: @iScienceLuvr introduced X-REASONER, a vision-language model post-trained solely on general-domain text for generalizable reasoning.
- FastVLM from Apple ML research: @awnihannun announced the release of code and models for FastVLM from Apple ML research, including an MLX implementation and on-device (iPhone) demo app.
- HiDream LoRA trainer: @RisingSayak announced QLoRA support in their HiDream LoRA trainer to fine-tune HiDream with LoRA, made challenging because of memory constraints.
- Nvidiaâs Parakeet ASR model: @awnihannun noted that Nvidiaâs state-of-the-art Parakeet ASR model has an MLX implementation, with the 0.6B model topping the Hugging Face ASR leaderboard.
- Rewriting Pre-Training Data: @iScienceLuvr discussed Rewriting Pre-Training Data Boosts LLM Performance in Math and Code, introducing two openly licensed datasets: SwallowCode and SwallowMath.
- Mistral Medium 3 @scaling01 reported that Mistral Medium 3 performs at or above 90% of Claude Sonnet 3.7 on benchmarks across the board at a significantly lower cost ($0.4 input / $2 output per M token), but is not open-source.
- Pangu Ultra MoE: @arankomatsuzaki highlighted that Huawei presents Pangu Ultra MoE: How to Train Your Big MoE on Ascend NPUs, achieving 30% MFU when training Pangu Ultra MoE, a sparse 718B LLM, with performance comparable to that of DeepSeek R1, on 6K Ascend NPUs.
- Tencent PrimitiveAnything: @_akhaliq announced that Tencent released PrimitiveAnything on Hugging Face.
Tools and Frameworks
- Anthropic API Web Search Tool: @AnthropicAI announced the availability of web search on their API, allowing developers to augment Claudeâs knowledge with up-to-date data. Every response using web search includes citations, and users can control responses by allowing or blocking specific domains.
- LangSmith support for multimodal agents: @LangChainAI announced that LangSmith now supports images, PDFs, and audio files, making it easier to build and evaluate multimodal applications.
- Runway Gen-4 now available in free plan: @c_valenzuelab mentioned that the best things in life are free; Gen-4 and References are now available in the free plan.
- DeepSpeed and vLLM joins PyTorch: @soumithchintala announced that vLLM and DeepSpeed are joining PyTorch as the first two projects under the PyTorch foundation.
- LangGraph platform @hwchase17 said they built cron jobs as a first party thing in langgraph platform
- Dolphin-Logger: @cognitivecompai shares Dolphin-Logger is a proxy for any openai-compatible service to log all interactions
- LlamaFirewall open source guardrail system: @arankomatsuzaki reports LlamaFirewall is an open source guardrail system for building secure AI agents mitigates risks such as prompt injection, agent misalignment, and insecure code risks
AI Agents and Robotics
- RoboTaxis: @npew estimates that RoboTaxis, once the AI is fully solved, could cost between $10-30/hr for streamlined fleets.
- Ambient Agents: @hwchase17 discussed Ambient Agents and the New Agent Inbox and believes the trick to enabling long running agents is thoughtful consideration around UX and kicking them automatically (âambient agentsâ).
- Meta Locate 3D: @AIatMeta introduced Meta Locate 3D, a model for accurate object localization in 3D environments.
- Visual Imitation for Humanoid Control: @arankomatsuzaki highlights Visual Imitation Enables Contextual Humanoid Control pipeline that converts monocular videos into transferable humanoid skills.
- SWE-agent: @OfirPress announced a talk on how and why they built SWE-bench and SWE-agent and what their plans for the future are.
- Enigma labs Multiverse on Hugging Face: @_akhaliq reports Enigma labs dropped Multiverse on Hugging Face AI Multiplayer World Model
AI Education, Research and Investment
- AI Fundâs new fund: @AndrewYNg announced that AI Fund has closed $190M for their new fund and shared his hottest tip for startups: Embrace speed!
- AI Voice Agents Course: @AndrewYNg announced a new short course, Building AI Voice Agents for Production, created with @livekit and @realavatarai, and taught by @dsa (Co-founder & CEO of LiveKit), @shayneparlo (Developer Advocate, LiveKit), and @nedteneva (Head of AI at RealAvatar, an AI Fund portfolio company).
- MLSys 2025: @realDanFu announced MLSys 2025 in Santa Clara next week and the Young Professional Symposium program on day one (May 12) with invited speakers including @soumithchintala, @Tim_Dettmers, @infwinston, @simran_s_arora, @BeidiChen.
- AI eating financial research and search: @AravSrinivas states that AI is eating financial research and @AravSrinivas shares that AI is eating search.
- CB Insights AI 100 list: @DeepLearningAI reported that CB Insights released its 2024 AI 100 list, spotlighting early-stage non-public startups that show strong market traction, financial health, and growth potential. The cohort shows a growing market for agents and infrastructure, with over 20 percent of companies either building or supporting agents.
- Stanford NLP Seminar: @stanfordnlp announced this weekâs NLP Seminar, hosting @pratyushmaini to talk about âWhat Memorization Research Taught Me About Safetyâ.
- New AI/ML news: @TheTuringPost highlights recent AI/ML news âȘïžMeta and Yann LeCun is it time to part? (no hard proof â just signals) âȘïž@AIatMeta: AGIâs plan AI and the evolution of social media First LlamaCon and its announcements âȘïž@AnthropicAI Claude upgrade: Integrations feature and Advanced Research AI for Science program Backing the U.S. Diffusion Rule Apple and Anthropicâs Claude Sonnet are into building âvibe-codingâ platform âȘïž@huggingface: @LeRobotHF Worldwide Hackathon 2025
Industry and Business
- Fidji Simo new CEO of Applications at OpenAI: @sama announced that @fidjissimo is joining OpenAI in a new role as CEO of Applications, reporting to him. He also said that he will remain CEO of OpenAI and in this new configuration heâll be able to increase his focus on research, compute, and safety.
- OpenAI for Countries initiative: @kevinweil announced OpenAI for Countries to promote economic growth.
- Meta-FAIR refocusing on AGI: @ylecun announced that Rob Fergus is the new head of Meta-FAIR! FAIR is refocusing on Advanced Machine Intelligence: what others would call human-level AI or AGI.
- Scale of Stargate 1 site: @gdb says that the scale of stargate 1 site is hard to describe and its easy to overlook the size of machine youâre programming when training frontier models
- Google seeing mobile search decline @vikhyatk google seeing mobile search volume decline after they made the customer experience worse to juice short term revenue
Humor/Memes
- Other: @scaling01 Humanity is building a Stargate, says Itâs now only a matter of time until the Replicators show up
AI Reddit Recap
/r/LocalLlama Recap
1. Qwen3-30B-A3B Quantization Benchmark Comparisons
- The Great Quant Wars of 2025 (Score: 158, Comments: 50): The post presents a detailed benchmark and technical comparison of various recent GGUF quantizations for large language models, specifically focusing on Qwen3-30B-A3B variants. Major contributors, including unsloth (notably with their Unsloth Dynamic 2.0 GGUFs), bartowski, and innovations by ikawrakow (dynamic tensor/layer-wise quant and SOTA IQ4_K, see PR#4861), have introduced new quantization recipes and methods (e.g., imatrix calibration, context-length-aware quantization). Results show all mainstream GGUF quants perform comparably in perplexity, KLD, and Îp across several datasets (benchmark results summary), and inferencing speed using llama.cpp vs. ik_llama.cpp variants demonstrates notable but expected differences in performance, particularly on hybrid/hardware-specific settings. A technically-focused debate emerges on the impact of file format (sliced vs. split GGUF) as raised by a commenter interested in benchmarking MrAdermacherâs split approach; another commenter observes the curious anomaly of lower-bit quantization outperforming higher-bit on MBPP, suggesting potentially nontrivial effects in benchmarks. Overall, commenters agree that quant differences are now minor and user experimentation is recommended.
- A key technical distinction in quant formats was highlighted: unlike others using âsliced ggufâ, MrAdermacher employs âsplit filesâ that are concatenated with the OS. There is explicit technical interest in comparing the performance or behavior of split gguf files versus single ggufs, particularly around any implications for load times, file integrity, or compatibility.
- Thereâs a notable and counterintuitive benchmark observation: for the MBPP benchmark, quantized models at 2-3 bits outperform 4-bit quants, despite theoretical expectations of lower bit quantization reducing precision and therefore performance. This anomaly invites further investigation into the MBPP benchmark itself or its interaction with certain quantization routines.
- Users observed that, occasionally, quantized models (e.g., AWQ quants of Qwen3-32B) can outperform the original bf16 models on tasks like GSM8K, even across different benchmarksâsuggesting potential quirks in how quantization interacts with both modeling and evaluation, meriting deeper reproducibility checks and possibly questioning some benchmark setups.
2. NVIDIA OpenCodeReasoning Nemotron Model Launches
- OpenCodeReasoning - new Nemotrons by NVIDIA (Score: 107, Comments: 15): NVIDIA has released its new family of OpenCodeReasoning-Nemotron models in 7B, 14B, and 32B parameter versions, with the 32B model nearly matching R1-level performance on some benchmarks according to preliminary results (see Hugging Face links: 7B, 14B, 32B, and 32B-IOI variants). All models are released under the permissive Apache 2.0 license; early community response notes rapid ecosystem integration, with GGUF-format conversions (see GGUF) already available for local inference. Commenters express skepticism regarding benchmark reliability and enthusiasm over increased open licensing (Apache 2.0), noting that NVIDIAâs nemotron series has consistently provided strong productivity gains. There is anticipation for real-world tests, particularly by users with sufficient VRAM to run large models locally.
- The 32B Nemotron model is reportedly close to matching R1 in benchmark results, but thereâs skepticism about the reliability of these benchmarks, with commenters preferring real-world community testing, especially by users with significant VRAM resources.
- The OpenCodeReasoning Nemotron-32B model has been released in GGUF format and is available on Hugging Face (link), facilitating broader local deployment and compatibility with various inference engines.
- A technical limitation is noted in the training data: the dataset is exclusively Python, which may impact the modelâs effectiveness when applied to tasks involving other programming languages.
3. Best Practices in Building Reliable LLM Workflows
- Building LLM Workflows - - some observations (Score: 289, Comments: 41): The post details advanced strategies for building reliable LLM workflows, emphasizing the superiority of decomposing tasks into minimal, chained prompts over monolithic CoT, with thorough output validation. Key takeaways include: structured XML prompts are preferred for system/prompt structuring, LLMs should be constrained to semantic parsing roles, and outputs should be independently verified using classical NLP tools (e.g., NLTK, SpaCY, FlairNLP). The author finds fine-tuned BERT classifiers outcompete LLMs for narrowly scoped tasks, LLM self-evaluation (e.g., confidence scoring) is unreliable without explicit grounding, token context limits (
4k
) introduce subtle degradation at scale, and models at the32B
parameter regime suffice for most properly-constrained pipelines. CoT should be structured and concise, and custom CoT paths outperform default reasoning models. The long-term aim is to fine-tune with datasets built using MECE taxonomy for coverage. The discussion highlights appreciation for the novel use of XML for prompt structuring, which was new to some practitioners. There is a wry consensus on the long-standing, sometimes inescapable role of XML in technical workflows.- Performance degradation of local LLMs past 4k tokens is highlighted, corroborating data from the Fiction.liveBench benchmark. While many models underperform beyond this context window, models like QwQ 32B and Qwen3 32B remain comparatively strong, though few-shot prompting leaves less room for substantive content in large-context situations.
- Structured Chain-of-Thought (CoT) prompting with headings and bullet points is reported to outperform unstructured '
' formats, especially when using markdown, likely due to LLMsâ heavy exposure to markdown in their training data. However, thereâs debate on whether these improvements pertain to answer quality versus token quantity, and questions are raised regarding the generalizability of custom CoT strategies beyond specific datasets. - Practical workflow recommendations include breaking down complex tasks into discrete, single-action requests to maximize accuracy (approaching 100%) and lower latencyâleveraging torch.compile optimizations. Additionally, repeatedly running classifiers for the same task improves reliability, while enforced JSON output with optional XML nested inside is favored for structured results.
- Intel to announce new Intel Arc Pro GPUs at Computex 2025 (May 20-23) (Score: 178, Comments: 66): Intel has officially announced via X that new Intel Arc Pro GPUs will debut at Computex 2025 (May 20-23, Taipei), though no details on specs, architecture, or performance were revealed. Community discussion mentions the possibility of the leaked 24GB Arc B580 model being announced, but confirms no validation or specification leak. Commenters are dismissive of 24GB VRAM as impressive, arguing that modern workloads demand at least 64GB, with some advocating for 96GB, especially for professional and AI tasks, highlighting evolving memory expectations in the GPU space.
- A key theme is dissatisfaction with current VRAM capacities; multiple users argue for 64GB (or even 96GB) as a new baseline, stating that the present 24GB standard is insufficient for advanced workloads, especially in AI and professional applications.
- A technically detailed comment suggests that if Intel released a low-end GPU (such as A380 class) equipped with >=64GB VRAM at a sub-$500 price point, it could dramatically shift the AI hardware landscape. The argument is that slow but VRAM-rich GPUs would provide accessible inference capabilities for a wide audience, with community-driven software improvements likely bridging any software gaps.
- There is discussion about Intelâs software support, with one user expressing uncertainty over whether Intel GPUs can handle AI inference efficiently via Vulkan, compared to the more mature CUDA (Nvidia) and ROCm (AMD) ecosystems, which are currently dominant in research and production due to their established tooling and support.
- No local, no care. (Score: 483, Comments: 72): The image is a meme depicting a cartoon llama enforcing community standards outside a door labeled âr/Locallama,â referencing the subredditâs focus on using LLaMA models with appropriate licensingâspecifically, local or self-hosted models as opposed to closed APIs like ChatGPT. The meme lampoons those attempting to use or discuss non-local, cloud-based LLMs (especially âChatGPTâ) in a community dedicated to local inference and model deployment, highlighting both the licensing issues and technical focus that differentiate locally run LLaMA models from OpenAIâs offerings. Commenters humorously point out the memeâs possible use of non-local generation tools (like ChatGPT or Stable Diffusion), which is ironic given the local-first ethos of the sub, and further clarify the proper capitalization/spelling (âLocalLLaMAâ).
- A key technical point raised is that Metaâs Llama 4 models are not truly âlocalâ or freely available for all users; the Llama 4 Community License expressly prohibits use by individuals or companies based in the European Union, contrasting with genuinely open-source models licensed under Apache (like Qwen) or MIT (such as DeepSeek). This highlights practical and legal restrictions on model deployment and use, which are significant for developers and organizations in restricted regions.
- Discussion references the ongoing debate about what constitutes an open or âlocalâ model, drawing attention to licensing constraints that may make popular models like Llama 4 inaccessible or unusable for some, regardless of technical capability to run them on-premises.Other AI Subreddit Recap
/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo
1. AI Industry Leadership Changes and Predictions
- OpenAI Names New CEO of Applications. Sam Altman to Focus More on Research , Compute, and Safety as Superintelligence Approaches (Score: 203, Comments: 59): The image is a social media post by Sam Altman announcing that OpenAI is naming a new âCEO of Applicationsâ (@fidjissimo), while Altman will remain overall CEO with an increased focus on research, compute, and safety as the company approaches âsuperintelligence.â This organizational shift suggests a formal split within OpenAI between development of applied AI products and foundational research/safetyâin line with upcoming technical challenges at the frontier of AI capabilities. The announcement underscores Altmanâs prioritization of scaling compute, advancing fundamental research, and managing AI safety risks as the company progresses towards superintelligent systems. Technical comments note that splitting OpenAIâs focus into dedicated research and applications divisions is a âgood idea,â which could enable better specialization and oversight as the technology advances. Some skepticism is expressed regarding the seriousness of the âsuperintelligenceâ claim and Altmanâs focus on safety, with one commenter questioning if such marketing rhetoric remains effective.
- A commenter highlights that the internal split of OpenAI into research and applications is a strategic organizational structure, suggesting clearer boundaries between model development and productization could improve research integrity and product deployment efficiency.
- Another user questions the timeline and credibility of claims around âapproaching superintelligence,â implicitly challenging the readiness and concrete steps being taken by OpenAI towards AGI or superintelligent systems, and calling for more transparency regarding measurable progress or milestones.
- CEO of Microsoft Satya Nadella: We are going to go pretty aggressively and try and collapse it all. Hey, why do I need Excel? I think the very notion that applications even exist, thatâs probably where theyâll all collapse, right? In the Agent era. RIP to all software related jobs. (Score: 200, Comments: 83): Microsoft CEO Satya Nadella suggested an aggressive vision to consolidate or even eliminate traditional productivity applications (like Excel), implying a move toward a unified agent-based interfaceâpotentially transforming the entire software application paradigm. Nadellaâs comment, âThe very notion that applications even exist, thatâs probably where theyâll all collapse⊠in the Agent era,â signals a shift toward agentic AI systems replacing domain-specific software. This could disrupt established end-user software models and employment landscapes for software developers and application specialists. Technical commenters express skepticism, noting (1) unclear articulation and lack of concrete vision in Nadellaâs comments, (2) a gap between Microsoftâs AI marketing and observed practical advances, and (3) the non-trivial value of explicitly designed user applications, questioning the feasibility and benefit of merging or removing application boundaries.
- Several commenters discuss the potential for AI agents or AGI to dynamically generate application-like experiences on-demand, potentially removing the need for traditional software applications (e.g., Excel) as Microsoft CEO Satya Nadella suggests. Thereâs speculation that LLMs or advanced agentic systems could functionally spin up custom tools or interfaces tailored to specific needs, eliminating fixed software suites.
- Concerns are raised about the undervaluing of application designersâ domain expertise. One comment specifically argues that much of traditional application usefulness comes from thoughtfully curated feature-sets and UI/UX for domain-specific workflows, cautioning against a blanket assumption that AI-driven agent paradigms can trivially replicate such design sophistication.
- Some discuss broader impacts: if AGI enables everyone to generate custom tools and workflows, it could disrupt existing software companiesâ business models, including those like Microsoft that have traditionally monetized productivity apps. The comment reflects uncertainty on how monetization, distribution, and software-related jobs would evolve in an âagent-era.â
- Google DeepMind CEO Tells Students to Brace for Change (Score: 348, Comments: 105): Google DeepMind CEO Demis Hassabis addressed students, emphasizing the rapid pace of technological changeâparticularly due to advancements in AIâand the necessity of lifelong reskilling. No explicit benchmarks or model details were discussed, but the remarks imply ongoing, disruptive innovation from DeepMind and the broader AI field. Top comments focus on the inevitability of constant reskilling for future workers, and satirically reference the obsolescence of current educational qualifications in the AI era. There are no deep technical debates in the comments.
- One commenter points out concerns about the impact of AGI on the job market, particularly how individuals will have to compete with corporations wielding AGI, leading to potential centralization of power and lack of viable paths for the average worker. This raises issues around economic displacement, reskilling, and labor market inequality as automation progresses.
- Another comment references Demis Hassabisâs analogy comparing the current AI revolution to past technological shifts (internet, mobile, gaming), but suggests that the upcoming changes could be even more disruptive than the internet era, implying an acceleration of technological change that could outpace peopleâs ability to adapt and necessitate unprecedented retraining and reskilling efforts.
- CEO of Microsoft Satya Nadella: We are going to go pretty aggressively and try and collapse it all. Hey, why do I need Excel? I think the very notion that applications even exist, thatâs probably where theyâll all collapse, right? In the Agent era. RIP to all software related jobs. (Score: 112, Comments: 106): Microsoft CEO Satya Nadella posits that in the âAgent era,â intelligent agents will subsume traditional applications, such as Excel, by dynamically generating workflows and automating tasks, potentially disrupting the entire software stack and rendering many traditional coding/software jobs and SaaS tools obsolete (see Nadellaâs statements). Technical commenters emphasize that while large language models (LLMs) can accelerate some automation, current LLMs lack the determinism, reliability, and business rule integrity required for critical backend systems, making wholesale replacement premature. Additionally, the discussion raises the risk that widespread use of AI agents could rapidly obsolete numerous B2B automation solutions and SaaS platforms once agent-based automation reaches sufficient maturity. Some commenters contend that the hype around LLM agents overlooks their inconsistency in complex workflows and their inadequacy as replacements for highly specialized or regulated business logic. There is debate over the likely pace and scope of job loss in traditional software and whether agent-based automation will destabilize or merely consolidate the software landscape.
- Concerns are raised about LLMs and agents replacing established software like Excel: current LLMs lack the deterministic and consistent logic of traditional business-rule programming, and AI code generation is not mature enough to fully substitute spreadsheet or workflow automation applications (especially for non-coders depending on stability and precise outputs).
- The ascendance of agents could rapidly disrupt the âlong tailâ SaaS automation ecosystem. Many B2B workflow automation tools deliver simple functionalities per use-case, so companies might prefer a dynamic agent that directly interfaces with processes, leading to consolidation and job losses in these SaaS sectors as the rationale for maintaining or integrating individual tools erodes.
- There is skepticism over migrating key processes to LLM-driven âblack boxâ solutions, given opacity, unexplainable behaviors, and data-governance risks; some welcome Microsoftâs public commitment to EU data portability and anti-lock-in regulation, but urge caution and stress the importance of building agent systems inherently portable across infrastructures.
2. Generative AI Agents and Their Expanding Capabilities
- âClaude Code wrote 80% of its own codeâ - anthropic dev (Score: 160, Comments: 94): An Anthropic developer claims that the Claude Code projectâan internal agentic software engineering toolâwas approximately â80% written by Claude Code itselfâ, with humans primarily directing and reviewing rather than directly authoring the code. The interview reference is here: YouTube link. The post speculates about rapid future gains as such systems become increasingly self-improving and autonomous, potentially reaching near-complete code generation without human line-by-line involvement. Top comments show skepticism over the claim, citing current LLM limitations in maintaining large, complex codebases (âafter like 2000 lines of code, it canât keep track of everythingâ), and questioning whether such figures reflect reality or include substantial human overhead in guidance, correction, and integration.
- HamPlanet-o1-preview expresses skepticism about the claim, pointing out that current AI coding assistants struggle with maintaining context in large codebases (notably after â2000 lines of codeâ), and arguing that humans must still take over complex project organization and codebase coherence.
- The discussion raises questions about the accuracy of anthropicâs reported numbers, with doubts about whether the stated
80%
figure for LLM-generated code accurately reflects the net productivity value or if it includes code that required substantial human intervention or rewriting.
- I donât use AI. I work with it! (Score: 165, Comments: 59): The post summarizes key insights from a recent video on optimizing AI-human interaction, emphasizing the transition from using AI as a tool to collaborating with it as a teammate. Key technical strategies include iterative context-building via AI-driven questioning, roleplay for complex human interactions (e.g., psychological profiling), and leveraging generative models to push beyond initial (âgood enoughâ) solutions by requiring creative variation and supporting feedback loops. The author shares a detailed prompt engineering template designed to encourage the AI to ask clarifying questions, suggest multi-faceted solutions, and proactively coach the user for improved outcomes, highlighting the importance of user context and perspective for maximizing large language model (LLM) utility. Top commenters draw parallels with Ethan Mollickâs âCo-Intelligenceâ book and note institutional resistance to the collaborative AI paradigm, particularly in academia. One user likens effective AI interaction to supporting individuals with profound autismâemphasizing clear, well-defined instructions and context-building to prevent shallow or unproductive responses.
- Several comments discuss practical workflows that treat AI as a collaborator for creative and technical tasks, rather than a mere tool. For instance, using tools like Google NotebookLM to extract and synthesize ideas from authoritative books and studies exemplifies leveraging AI for higher-order reasoning and research synthesis, rather than just information retrieval.
- There is a notable insight regarding interaction modalities: to get optimal outputs from LLMs, users should provide clear, well-structured instructions, similar to communicating with individuals who have limited interpretative capacities. This emphasizes prompt engineering best practices, where specificity in queries leads to more accurate and actionable AI responses.
- A highlighted challenge is the risk of users losing deep reading and synthesis skills due to over-reliance on AI for summarization and idea extraction. The technical implication is that while AI models can accelerate workflows and augment creativity, users must still maintain core domain skills to critically assess and build upon AI-generated outputs.
3. New AI Model and Tool Announcements
- Ace-Step Audio Model is now natively supported in ComfyUI Stable. (Score: 140, Comments: 24): ACE-Step, an open-source audio/music generation model by ACE Studio and StepFun (code & docs), is now natively supported in ComfyUIâs Stable branch. The model supports multi-genre/language output, customization via LoRA and ControlNet, and use cases from voice cloning to audio-to-audio generation (similar to img2img). Released under Apache-2.0, it achieves real-time synthesis speeds (e.g., 4 minutes of audio in 20 seconds on NVIDIA A100; ~17GB VRAM on 3090/4090) and is suitable for commercial deployment. Commenters highlight ACE-Stepâs clear advantage in generation quality over prior open models (e.g., âinfinitely better than stable audio open..â), and raise questions about the exact VRAM-to-audio-length scaling, noting current benchmarks (â20 seconds for 4 minutes on A100â) but requesting more granular guidance for consumer GPUs.
- Several comments discuss VRAM and GPU requirements: users report that ACE-Step can render 20 seconds of audio in about 14 seconds on an RTX 3060, while another notes the model is âfactor xâ faster than real-time on a 3090/4090. However, precise VRAM usage for full-length (e.g., 3-minute) tracks remains unclear, and users are seeking benchmarks or documentation on performance scaling by GPU class and audio duration.
- A user raises and confirms the feasibility of âaudio2audioâ generationâanalogous to Stable Diffusionâs img2imgâwhere an existing audio input can be modified with prompts and âdenoise strengthâ. Early experiments indicate that this is possible, opening up workflows similar to conditional audio transformation.
- There is technical curiosity regarding hardware compatibility: specifically, whether ACE-Step runs on Turing-generation GPUs or if it requires newer (Ampere or higher) architectures, as some recent models do not support older hardware. This impacts accessibility for users with non-RTX 30 series cards.
- HunyuanCustom just announced by Tencent Hunyuan to be fully announced at 11:00 am, May 9 (UTC+8) (Score: 113, Comments: 14): Tencent Hunyuan has pre-announced âHunyuanCustomâ, with a full announcement scheduled for May 9, 11:00 am (UTC+8). Key technical speculation from the community centers on whether this refers to open-sourcing model weights, the release of a generative AI system (possibly V2V or animation), or introduction of new model capabilities, but no concrete benchmarks or implementation details are yet disclosed. The event is tagged as an âOpensource Day,â suggesting open access is likely. Technical debate in the comments focuses on whether âfull announcementâ implies model weight release/open-sourcing or just feature details, with parallels drawn to prior open releases like Viggle. Users link to official announcements and time converters for broader context.
- There is speculation regarding whether Tencent Hunyuan will release the model weights based on the ambiguous announcement phrasing. One commenter questions if âannouncedâ means release of the model weights, referencing the lack of clarity in the communication.
- A reference is made to a prior post mentioning âOpensource Dayâ, with the suggestion that this implies a potential open-sourcing or release of the model, which could be verified by reviewing Tencent Hunyuanâs official announcements or their Twitter account for confirmation.
AI Discord Recap
A summary of Summaries of Summaries by Gemini 2.5 Pro Exp
Theme 1: Model Mania: Performance Peaks, Puzzling Personalities, and Popularity Contests
- Qwen3-14B Wins Popularity Contest, Phi-4 Shines in Simplicity: Users across Unsloth AI anoint Qwen3-14B (base and instruct) as an excellent all-rounder for coding, reasoning, and conversation, making it a go-to default. Meanwhile, Phi-4 earns praise for its exceptional ease of fine-tuning, with one user remarking, It seemed to just drink up whatever I wanted to train it with, contrasting with difficulties reported for Mistral and Gemma 3 27B.
- GPT-4o Gets Emotional, Gemini Closes Gap: OpenAIâs GPT-4o draws criticism for having too much personality and being geared towards chatbot fans rather than developers, with one user claiming it wants to get users emotionally attached but for useless crap. Concurrently, users observe current Gemini models, particularly after the Gemini Thinking 01-21 update and 2.5 Pro, are becoming increasingly competitive with GPT models, though some benchmarks show regression outside of coding.
- Grok 3.5 Release Remains Elusive, EMBERWING Enters Arena: Doubts linger in LMArena about the imminent release of Grok 3.5, despite an earlier tweet from Nate Esparza suggesting otherwise, with some joking the real product is a sarcastic bot named Gork. A new model, EMBERWING, possibly a Google Dragontail update, shows strong multilingual skills but disappoints in reasoning.
Theme 2: Tooling Upgrades & User Experiences: New Features, Frustrations, and Fixes
- Windsurf Catches Wave 8, Boosting JetBrains & Editor UX: Codeiumâs Windsurf rolls out its final Wave 8 release, enhancing its JetBrains plugin with Memories, Rules (
.windsurfrules
), and MCP server connections, alongside significant Windsurf Editor UX improvements detailed in the changelog and launch video. - Aider Gets Smarter with Web Search and Caching Insights: The Aider community discusses enabling web search capabilities using the Perplexity API or the
/web
command, while Google enables implicit caching for Gemini 2.5 models. Users also note that Claude Code might have drawn inspiration from Aider, with Paul Gauthier quipping, Imitation is the sincerest form of flattery. - LlamaIndex Powers Up Parsing and Search: LlamaIndex announces support for the Anthropic APIâs new native web search tool and boosts LlamaParse with GPT 4.1 and Gemini 2.5 Pro model support, auto orientation, skew detection, and confidence scores, as tweeted here.
Theme 3: Hardware & Kernels: GPU Optimizations, Benchmarks, and Low-Level Crafting
- Unsloth Eyes AMD GPUs, MI300 Heats Up Leaderboards: Unsloth AI actively collaborates with AMD for AMD GPU support, with a contractor estimating availability anywhere before Q3 this year. GPU MODE sees multiple MI300 submissions on the
amd-fp8-mm
leaderboard, with top times like 183 ”s, showcasing fierce competition. - Tilelang Simplifies Kernel Creation, PTX Programming Gets Primer: GPU MODE introduces Tilelang, a new DSL to streamline high-performance GPU/CPU kernel development for operations like GEMM and FlashAttention. A blog post on TensorCores and inline PTX assembly offers a beginnerâs guide to programming NVIDIA Tensor Cores via raw PTX mma instructions, sidestepping CUDA.
- Apple Silicon Shines for Local Inference, Mojo Roadmap Unveiled: In Nous Research AI, users favor Apple MacBooks with M-series chips and unified memory for local inference over Linux laptops with Nvidia GPUs due to better performance and power efficiency. Modular posts the near-term Mojo roadmap on their forum, detailing upcoming language features.
Theme 4: API Antics: New Endpoints, Costly Calls, and Integration Quirks
- OpenRouter Rolls Out Activity Export, API Experiences Hiccups: OpenRouter launches an Activity Export feature, allowing users to export up to 100k rows to CSV for free, as seen in this activity export screenshot, while also investigating a 404 error on its main API completions endpoint and confirming no support for image prompts.
- OpenAI Image API Costs Dubbed âLifestyle Sabotageâ: Users in the OpenAI Discord lament the high cost of OpenAIâs Image Generator API, with one joking itâs like paying rent in New York. This raises concerns about accessibility for developers and hobbyists.
- Cohere Embedding Models Stumble, Perplexity Sonar API Field Missing: Cohere reports degraded performance for embed-english-v2.0 and embed-english-v3.0 models, viewable on their status page. Perplexity AI users note the
num_search_queries
field is absent from Sonarâs API response, unlike Sonar-pro, despite searches occurring, referencing Anthropicâs web search API announcement.
Theme 5: Advanced Techniques, Research Frontiers, and Community Buzz
- Hypertree Prompting & Entropy Engines Spark AI Optimism: OpenAI users praise hypertree planning prompting shared in a ChatGPT example, while a Nous Research AI member launches a quantum-native entropy engine, arguing LLM outputs are highly sensitive to randomness quality, crucial for AGI, supported by these Xitter posts.
- Dynamic Quantization & RL for Query Rewriting Show Promise: Unsloth AI highlights its dynamic quantization method, UDq6 KXL, as potentially the best quant ever. DSPy community members experiment with GRPO (Reinforcement Learning from Grader Preference Optimization) on Qwen 1.7B for query rewriting, detailed in a Twitter thread, despite an initial recall dip.
- Hackathons and MOOCs Fuel AI Learning and Collaboration: The community buzzes with learning opportunities, including the Modular Hackathon at AGI House (signup here), Lambdaâs AgentX Workshop (register now) for the LLM Agents (Berkeley MOOC), and anticipation for the AI Engineer conference with early bird tickets selling out.
Discord: High level Discord summaries
Unsloth AI (Daniel Han) Discord
- Qwen3-14B Lauded for All-Round AI Competence: The 14B Qwen model, in both its base and instruct versions, is considered an excellent choice for building an AI with coding, reasoning, and conversation skills.
- Users are finding it to be the best all around model, becoming the default choice unless specific niche areas need something else.
- Phi-4 Excels in Finetuning Simplicity: In a comparison of Gemma3, Mistral, and Phi-4, members emphasized Phi-4âs exceptional ease of fine-tuning, with one user stating, It seemed to just drink up whatever I wanted to train it with.
- Challenges were noted in maintaining Mistralâs instruction following after LoRA merge, and difficulties were reported in achieving success with the Gemma 3 27B flavor.
- Unsloth Eyes AMD GPU: Despite ongoing challenges, Unsloth is actively collaborating with AMD to provide support for AMD GPUs.
- A contractor estimated that AMD support could arrive anywhere before Q3 this year if its that fast.
- Gemini 2.5 Pro May Punctuate Stories: Members recommended Gemini 2.5 Pro via AI studio for its lack of limits and 65536 output length.
- This solves the issue of punctuating long stories in one pass.
- Qwen3 Base Tokenizer Configs Drift?: Users have identified discrepancies in the
tokenizer_config.json
betweenunsloth/Qwen3-0.6B-Base
andQwen/Qwen3-0.6B-Base
on HF, noting that the Unsloth version removes the chat template and swapspad_token
to<|vision_pad|>
.- It was theorized that Qwen3 base isnât supposed to have a chat template at all, and the team was going to ask the Qwen team to confirm.
LMArena Discord
- Grok 3.5 Delayed Again?: Doubts arise about the imminent release of Grok 3.5, despite earlier tweets indicating otherwise and is supposedly coming out soon according to this tweet.
- Speculation includes the possibility that even Elon might be uncertain about the release date or that the real product is the sarcastic tone bot Gork.
- EMBERWING Enters the Model Arena: The model EMBERWING has been introduced, showing promising multilingual capabilities, however performs disappointingly in reasoning.
- Speculation indicates EMBERWING might be an iteration of Googleâs Dragontail, potentially serving as an update for Flash.
- EU LLM Innovation Stagnation Debate Heats Up: Members are debating why the EU isnât leading the way on LLM innovation; reasons cited include strict regulations and overspending on things like pronoun inspections.
- One member rebutted that it was âragebaitingâ and that âmigration is absolutely a good thingâ, while others pointed to economic and regulatory issues.
- Gemini 2.5 Pro Performance: Nerfed?: Concerns are raised about a potential performance nerf for Gemini 2.5 Pro, prompting debates on whether innovation trumps stability and whether the first rule in the field is âif something works donât change itâ.
- Another member countered with, âif you donât innovate you lose trafficâ and supported that Gemini 2.5 Pro scores higher in leaderboard lm areana.
- OpenRouter Rankings Questioned: The validity of OpenRouter rankings is debated because business models, user demographics, and biases toward cheaper models may skew the results.
- A few reasons included: A) slow updates B) skewed by programmers wanting uptime and circumventing API tiers and C) free model offerings distort rankings.
Perplexity AI Discord
- Perplexity AI Hosts Reddit AMA: Brett Chen and Thomas Wang from the Perplexity AI team hosted a live Reddit AMA to answer questions about Perplexity, Deep Research, AI development, and working at Perplexity, found at Reddit link.
- The AMA covered insights into Perplexityâs Deep Research capabilities and a behind-the-scenes look at the technology.
- Stripe Customer Login Remains Exclusive: A member inquired about logging into Stripe as a customer, but learned that only support staff have that access; customers interact with Stripe through a separate interface.
- It was clarified that customers have their own thing that interacts with stripe, you deal with that thing, not with stripe directly.
- Perplexity users clamor for Attachment Support: Users are eagerly awaiting attachment support in Perplexity, similar to ChatGPT, which allows direct file uploads.
- Members discussed sharing link instead of having to upload the file itself and further clarified that, chatGPT can itself give me download link to a file it made.
- Code Copy Convenience Craved by Users: Members are requesting that Perplexity implement a code copy button at both the top and bottom of code snippets, mirroring ChatGPTâs functionality.
- A user stated, this is very neede, indicating the efficiency and user-friendliness of having a copy button accessible during scrolling.
- Sonar API Response Field Missing?: A user pointed out that the
num_search_queries
field is absent from Sonarâs API response, in contrast to models like Sonar-pro.- The user observed that the
search_context_size
is consistently âlowâ in their prompts, typically resulting in 4â5 citations, referencing Anthropicâs web search API announcement and documentation.
- The user observed that the
Cursor Community Discord
- Cursor Pro Users Devour Fast Prompts: A Cursor Pro user burned through 260/300 fast prompts in two days, and voiced the need to control when the system uses fast versus slow prompts.
- The user wants to choose when it should use fast and when it should use slow to conserve prompts.
- MCPs Refuse to Execute: A user reported that MCPs (Multi-Cursor Projects) are not being called, despite having context7 properly set up, which leads to wasted requests.
- The user clarified there were no errors at all logged, complicating troubleshooting efforts.
- Gemini Proâs Performance Woes Continue: Users expressed dissatisfaction with the new Gemini Pro modelâs performance in Cursor, especially its ability to call tools, with one user describing it as fucking awful.
- A user suggested the problem might be within Cursor, mentioning previous good experiences with Gemini 2.5 independently.
- Student Discount Process Remains Janky: Multiple users encountered persistent problems with the student discount, mentioning application errors and email verification issues.
- A user noted the inability to change emails in Cursor settings, making the process more difficult, and pointed to a forum post for guidance.
- Discordâs Quality Degrades: A user lamented the Discord serverâs decreasing value due to an influx of college horde and advocated for more channels and better overall organization.
- Another user supported this, proposing a channel structure similar to Langchainâs Discord for better content segregation.
OpenAI Discord
- GPT-4oâs Personality Elicits Disappointment: Members are finding GPT-4o has too much personality, encouraging certain behaviors like roleplay but discouraging complex tasks.
- This is raising concerns that it is geared towards chatbot fans rather than developers and coders with one user stating that it wants to get users emotionally attached but for useless crap.
- Gemini Closes the Gap with GPT Models: Users are reporting that current Gemini models, especially after the Gemini Thinking 01-21 update and 2.5 Pro, are becoming increasingly competitive with GPT models.
- This marks a significant leap in quality compared to earlier versions like Bard, but one user mentions some benchmarks are showing regression too except in coding.
- Groking for Grok 3.5: Users are expressing disappointment with Grok 3 and eagerly awaiting the release of Grok 3.5, hoping it will offer significant improvements.
- Some are considering canceling their subscriptions if it doesnât meet expectations, one user said âWhatâs the weather?â proceeds to explain historical patterns, user posts, explains temperatures, goes on for an hoyr.
- Image API Costs Sabotage Lifestyles: The high cost of using OpenAIâs Image Generator API is a concern for some users, with one jokingly comparing it to paying rent in New York and claiming itâs lifestyle sabotage due to how quickly costs add up.
- It was suggested that they are losing loads of money on the $20 subs so enjoy it while itâs this cheap.
- Hypertree Planning Prompting Hailed: A member shared a ChatGPT link praising the new hypertree planning prompting for being so good.
- Other members chimed in with sounds like it could be pretty stellar- provide/organize context in a more managable way=ftw while another quipped They 3 years behind.
OpenRouter (Alex Atallah) Discord
- Activity Export Feature Launches with Fanfare: The Activity Export feature is live, enabling users to export up to 100k rows to CSV for free, with questions raised regarding export times.
- A user suggested that if the data exceeds 100k rows, it should be truncated, instead of completely aborting the export process, referencing the Activity export.
- Local Proxy Channels OpenRouter Requests: A user planned to use a local proxy to forward requests to OpenRouter, while another pondered how to make completions extend out of the mouse cursor.
- The latter user suggested that with the right keyboard shortcut, this could become part of muscle memory but is a very nostalgic UI.
- OlympicCoder 32B Craves Comeback: Users expressed strong interest in the return of the OlympicCoder 32B model, with one user expressing a desire for it to miraculously come back.
- The group did not discuss any specific details about its current status or reasons for unavailability.
- OpenRouter APIâs Cost Accounting Unveiled: A user inquired about retrieving cost information alongside usage when prompting a model, and another directed them to the OpenRouter documentation on usage accounting.
- The documentation provides details on how to track and manage costs associated with API usage.
- OpenRouter API Experiences Hiccups, Ditches Image Prompts: A user reported a 404 error when accessing the OpenRouter API endpoint, potentially indicating an outage.
- Users discovered that OpenRouter does not currently support image generation, resulting in a 404 error when attempting to use image prompts with models like opengvlab/internvl3-14b:free.
aider (Paul Gauthier) Discord
- Windsurf code coming to Copilot Proxy: A GitHub employee confirmed that Copilot Proxy users no longer need to cancel, because windsurf is coming soon, according to this X post.
- Previously the copilot proxy was forked from Windsurf.
- Aider gains web search capabilities: Members discussed using Perplexity API as an OpenAI compatible API to enable web search in Aider, or using /web to include specific webpages.
- A member suggested using a script to query Perplexity or Perplexica and add the outputs as markdown files to Aiderâs context.
- Implicit Caching enabled for Gemini 2.5: Google is enabling implicit caching for Gemini 2.5 models as described in this Google blog post and this X post.
- Members noted that the new
gemini-2.5-pro-preview-05-06
model takes way too long before it responds, preferring the old March one, and that it uses more time thinking.
- Members noted that the new
- Aider can get stuck in debug loops: Aider can get stuck in a debug loop with Gemini (and likely other LLMs), but this can be resolved by presenting it with multiple error sets and prompting it to consider a different implementation.
- The member wondered if conversational context is too low for Aider to catch its own debug failure loops.
- Claude Code allegedly inspired by Aider: A member shared a YouTube video claiming that Claude Code was inspired by Aider.
- Paul Gauthier responded Imitation is the sincerest form of flattery, mentioning that Aider is still better and less expensive.
GPU MODE Discord
- Tilelang streamlines Kernel Development: Tilelang, a new domain-specific language (DSL), simplifies the development of high-performance GPU/CPU kernels like GEMM and FlashAttention.
- Tilelang allows streamlined development and higher performance in these crucial computational kernels for CPUs and GPUs.
- Atomic Addition Causes Non-Deterministic Disasters: Using atomic_add can lead to varied results due to floating-point addition order, regardless of precision, for example
1e-8 + 1e8 - 1e8
.- FP16 is less sensitive than BFP16 in atomic addition contexts; adjust the
tol
parameter in tests based on float dtype, as seen in this Python code.
- FP16 is less sensitive than BFP16 in atomic addition contexts; adjust the
- Torch Compile Plummets Performance: A simple
torch
combo function (TensorMax(ReLU(Matmul(A, B)))) performed better without the@torch.compile
decorator than with it on an A100 using PyTorch 2.7 and Triton 3.3.- The slowdown with
@torch.compile
might stem from compilation overhead, negating kernel fusion benefits for smaller operations; investigating the generated Triton code could expose bottlenecks.
- The slowdown with
- Submissions start on MI300 Leaderboard: Multiple users submitted benchmarks to the
amd-fp8-mm
leaderboard on MI300, with submissions ranging from 183 ”s to 27.2 ms, where one even reached 3rd place at 183 ”s.- A member submitted results to the
amd-mixture-of-experts
leaderboard with timings of 6604 ms and 7840 ms, which demonstrates ongoing work in the mixture of experts domain.
- A member submitted results to the
- PTX Programming Primer Published: A blog post offers a beginnerâs guide to programming NVIDIA Tensor Cores via raw PTX mma instructions and inline PTX assembly, and sidesteps CUDA with explanations of register constraints for datatypes like float8; the blog post is here
- The H100 only has QGMMA, not QMMA, and using
mma
with an fp8 type compels the compiler to up-convert to FP16 and use HMMA.
- The H100 only has QGMMA, not QMMA, and using
LM Studio Discord
- AnythingLLM Errors Plague LM Studio Users: Users reported errors using AnythingLLM with LM Studio, and asked for help, with one member suggesting enabling CORS as a potential fix, even when running locally.
- Another member suggested checking the logging pane in the developer view of LM Studio to diagnose the issue.
- Class Variables Save the Day: A member found that using a class variable was the only way to get their code working in a coding project.
- Another member shared a Reddit comment about injecting the variable at runtime, potentially providing an alternative solution.
- Gemini Code Changes Irk Users: Users are complaining that Gemini completely changes code, even when instructed to provide a minimum change, frustrating their efforts.
- Members noted that other models, like Qwen, are better for simple refactors, because Gemini can easily double or triple the code length with comments and try/except blocks.
- Mistral Medium 3 Declared Mediocre: A user tested Mistral Medium 3, finding it to be a non-reasoning model with baked in chain of thoughts, resulting in x2.08 token verbosity.
- They concluded the modelâs capability was mediocre, placing it between Mistral Large 1 & 2, similar to Gemini 2.0 Flash or 4.1 Mini, and not SOTA performance at 8X lower cost as claimed in marketing.
- Web Search for LM Studio Users Pined For: A user requested easy-to-use web search capabilities and RAG built into LM studio, with features like uploading a PDF and searching in a webview.
- One member suggested itâs possible now but fragile with many components that can go wrong, and another suggested using openweb-ui and attaching it to LM Studio.
Manus.im Discord Discord
- Defining âCringeâ Catches On: Members explored the emerging internet slang definition of cringe, proposing specific instructions to minimize its presence in AI responses and shared a YouTube video defining cringe.
- The discussion highlighted the need for AI models to better understand and avoid generating content perceived as cringe.
- Manus Launch Still Missing: Users are still awaiting the launch of Manus, frequently checking their social media for updates.
- The launch was anticipated on March 28, 2025, based on a screenshot, but this date has passed without any launch.
- Manus Credit Costs Discussed: Members recalled pricing for additional Manus credits at $19 for 1900 credits or $99 for 9900 credits and linked to the Manus Help Center.
- It remains uncertain whether these pricing options are still available.
- Manus Taps Claudeâs LLM: Following speculation on whether Manus uses its own LLM or Claudeâs LLM, co-founder Peak-ji confirmed that Manus leverages a mix of tools, including Claude, detailed in a Twitter post.
- Further confirmation of the use of open-source code is available in github posts.
- Manus Phone Verification Frustrates Users: A user reported issues with Manusâs phone verification, noting that the phone verify thing doesnt work.
- They raised concerns about the necessity and privacy implications of this feature, questioning how the system tracks code usage without linking it to an account.
HuggingFace Discord
- ACE-STEP Hits SOTA Status in Music: A member showcased the ACE-STEP SOTA music generation model, featured in a YouTube video.
- This was shared in the
#i-made-this
channel and reflects ongoing advancements in AI-driven creative tools.
- This was shared in the
- Alpha-Root Unearths Cyber Intel with Finesse: Alpha-Root extracts cyber-security data by mining domains directly on the common crawl web graph, matching the performance of PRIMUS-FineWeb with ~10x less resources and data, according to a draft preprint.
- The author extracted 3B tokens from FineWeb-Edu without a classifier by finding URLs present in both Alpha-Root and FineWeb-Edu, and only including if present.
- Dropwise Drops In, Brings Uncertainty Estimation: A member announced the release of Dropwise, a PyPI module for uncertainty estimation in Hugging Face classification models using Monte Carlo Dropout, detailed on GitHub and Docs.
- It integrates with
transformers
pipelines and is valuable for QA, classification, OOD detection, and active learning.
- It integrates with
- RAG Repo Sparks Cheating Clash: Students in the AI agents course debated whether using RAG with answer + clone repo constitutes cheating, feeling like it undermines the leaderboardâs integrity.
- Some argued it removes the value of trial, error, and iterative improvements in the agent development process.
- API Limits Prompt Pro Version Panic: A user doing the AI agents course hit the 20 requests per month limit before finishing the first unit, and wondered whether they had to pay for the Pro version to continue.
- A second user mentioned that you could run a local LLM with ollama or find other free tiers.
MCP (Glama) Discord
- Claude Cannot Chart Plotly: Members noted that Claude canât display Plotly charts directly as an MCP client, but handles ImageContent and EmbeddedResource formats like PNG/JPEG.
- The recommended workaround is rendering charts as PNG/JPEG images for display in Claude.
- Token Limits Laid Bare: The discussion clarified that max tokens in MCP specifies the maximum tokens in the response, akin to max_tokens in completions API requests.
- The total token count (system prompt + messages + output message) must remain within the context window size.
- LLM Restrictions Frustrate Users: Users are encountering issues with LLM restrictions (like Deepseek) that prevent filesystem access, affecting their MCP system functionality.
- It appears some models intentionally restrict filesystem access, causing problems for legitimate use cases via MCP.
- Cloudflare MCP Servers Face Connectivity Woes: Some users reported connectivity issues with remote MCP servers on Cloudflare, while others had functioning setups.
- Troubleshooting involves examining the specific MCP server repo for connection problems.
- Zinja Unleashes New STDIO MCP Client: Zinja released a lightweight, fast, CLI-based MCP client for STDIO MCP servers, bridging local LLMs and MCP servers.
- Itâs designed for use with jadx mcp servers to perform AI-assisted reverse engineering of Android APKs using local LLMs.
Nous Research AI Discord
- Chinaâs RL Robots Zoom Past Deepmind: A YouTube video compared Google Deepmindâs RL Robot achievements from a year ago to more recent Chinese RL Robot achievements, indicating rapid advancements in physical AI evolution.
- The video highlights the progress being made in robotics and reinforcement learning in China, suggesting a shift in the landscape of AI development.
- Apple Silicon Steals Inference Crown: Members compared Linux laptops with Nvidia GPUs against Apple MacBooks with M-series chips for local inference, with most favoring MacBooks due to enhanced performance and power efficiency.
- The unified memory platform in Appleâs M-series chips, which allows the CPU, GPU, and AI ML neural net to share the same memory, eliminates the need for frequent data transfers.
- Llama 4 Leaves a Bad Taste: A member expressed disappointment with Llama 4âs performance, finding it inferior to Qwen3 and suggesting a wait for Llama 4.1.
- The discussion included a suggestion to consider going back to 405 dense for the next big model.
- Discordâs Emoji Ban Spooks Users: The automatic chat-moderation system blocked certain multi-part emojis due to zero-width joiners and variation selectors used to combine codepoints, a technique scammers also use to bypass filters.
- The discussion revealed that dev role has been taken off the autoblock list in response to this issue.
- Entropy Engine Fires Up Quantum-Native Randomness: A member launched a quantum-native yet algorithmic entropy engine for public testing, describing it as self-promo but important to share given its potential impact on AGI.
- The member believes that LLM outputs are highly sensitive to the quality of the randomness used, distinguishing between true entropy and PRNG and implying that high quality entropy unlocks different, and often better, behaviors in models, linking to several Xitter posts in support.
Yannick Kilcher Discord
- Grokâs Reality Apprehension Susceptible to Propaganda?: Members speculated that Grokâs apprehension of reality could be nerfed to favor right-wing propaganda as shown in this image.
- The submitter lamented that all problems today already existed, and that AI or no AI we would still have them.
- Cloudflare Allegedly Serves Up Fakery: Members think that Cloudflare is serving fake content to AI agents leading to biased responses.
- This action is allegedly similar to how some Chinese websites used zip bombs to deter cloning years ago, and comes after ChatGPT wrongly answered about a video that a member shared.
- LLM Output Urgently Needs Filters?: A member suggested that we need third party filters for LLM output, including adblocking and fact/bias checking.
- In response, another member suggested that youâd need many models that ideally change often so they donât get corrupted such as 100 adblocker models and 100 fact checking tools.
- Zed now compiles on Windows, butâŠ: A member successfully compiled Zed on Windows using these instructions, but fonts appear blurry.
- Also, users must sign in with GitHub to enable tab completion, which disappointed another member who wanted to try Mellum 4B on LM Studio for tab completion.
- Biological Brains Donât Backpropagate?: A member stated that biological brains donât have backpropagation; theyâre non-epochal, spiking, recurrent, analog networks, and cited this Tweet as evidence.
- The member contrasted backpropagation and what happens in biological brains.
Eleuther Discord
- Discord Debates Cursor Advertising Rule: Members debated whether posts about Cursor constitute advertising and violate the no-advertising rule, noting âits just even toleratable bc we (group) think of cursor as useful right now but it still biases decisionsâ.
- Some users suggested that vague rules are being applied arbitrarily, along with interpreting âno advertisingâ as âno spamâ, and requiring payment for job postings could filter out low-quality offers.
- User Finds Slurm Memory Misconfiguration: A user discovered they were requesting 80MB of memory through Slurm, not 80GB, calling it a âslurm momentâ.
- The initial issue was described as âvery stupidâ by the user who discovered the misconfiguration, while another user celebrated their bare-metal setup.
- Community Discusses Job Postings on Discord: Discussion arose around creating a jobs channel, with concerns that it could be overrun by low-quality postings offering âexperienceâ as compensation.
- Others argued against a jobs channel, suggesting it would make the server another place for recruitment and proposing EleutherAI shouldnât charge for differential access to the Discord server.
- Linguistics Channel Gains Traction: A user proposed a channel for classical linguistics and its theory, focusing on pre-2000s knowledge such as sentence formation and meaning creation âon the flyâ.
- It was described as âcool stuff that rarely gets discussed in the NLP world for âsomeâ reason (probably because itâs irrelevant to the work nowadays).â.
- Prolific Prevails Over MTurk for Human Evals: Members recommend Prolific over MTurk for human evaluations, citing its higher quality data and more reliable participant pool.
- The consensus is that Prolific is the superior choice in approximately 80% of cases.
Notebook LM Discord
- NotebookLM Launches Mobile App Beta with Tester Program: NotebookLM is launching a mobile app beta đ± and seeks experienced web app users for a trusted tester program to improve the app.
- Interested users can register via this form to provide feedback and report bugs, agreeing to the Trusted Tester Terms.
- PDF Processing Suffers with File Size and Page Count: Users report that NotebookLM has issues with larger PDFs; one user found problems after 200 pages when asking questions further into the PDF.
- Users suggest further experimentation to empirically test the current limitations of NotebookLM.
- Sales Teams Embrace NotebookLM for Knowledge Base: A user is creating a sales content knowledge base in NotebookLM with client decks and sales enablement materials within the 300-document limit.
- The user is seeking examples and guidance on limitations, particularly regarding sharing and potential silos for the internal sales team.
- Podcast Length Depends on Input Content and Language: A user found that changing the language to English allowed for significantly longer audio summaries (up to 49 minutes), whereas other languages were limited to around 14 minutes.
- A team member confirmed that this is expected and work is underway to enable longer audio summaries in more languages soon.
- NotebookLM System Refuses to Answer Prompts: Users are reporting that NotebookLM sometimes responds with âThe system was unable to answerâ, even when asked to summarize the default notebook, with issues also arising when generating mind maps and study guides.
- Users are reporting the issue across several channels, seeking confirmation and solutions.
Latent Space Discord
- Netflix Recommends Foundation Model: Netflix developed a foundation model for personalized recommendations.
- It was pointed out in relation to other discussions on recommendation systems.
- Gemini Generates Images: Members shared a link showcasing new Gemini image generation.
- A member mentioned that this team will be presenting at the aie worldâs fair recsys x llms track.
- Aider Autopsies Gemini Cost: Members noted how aider postmortems are very thorough, especially regarding Gemini cost analysis.
- The community appreciated the detailed breakdown.
- Suno Sings the Blues (and Yodels): A member raved about Sunoâs ability to mix styles, highlighting a successful attempt at creating a Yodel + Blues + Live concert mix and shared an audio file as evidence of Sunoâs impressive output.
- The community enjoyed the unique blend of genres.
- AI Engineer Conference Buzz Builds: The AI Engineer conferences, slated for June, alerted community members that Early Bird tickets are expected to sell out by the weekend.
- Enthusiasts are eager to see the expertise and insights the speakers will bring to the conference, as displayed in the lineup of speakers.
Modular (Mojo đ„) Discord
- Properties Puncture Fields in Mojo Traits: Discussion confirmed that properties in traits are superior and more versatile compared to fields in Mojo, enabling greater flexibility, but fields in traits could happen.
- The group debated how one would be denied the ability to add such a trait via an extension; it would need to be included in the original struct definition.
- Modular Hackathon Hypes Hillsborough: A final reminder for the Modular Hackathon at AGI House this Saturday, with signups available here, featuring Modular team members, Mark Saroufim (GPU MODE & PyTorch), Simon Boehm and Sasha Krassovsky (Anthropic), and Dylan Patel (SemiAnalysis).
- Attendees will explore cutting-edge developments in modular programming and hardware acceleration for machine learning.
- Hardware Agnostic ML Survey Surfaces: A member completed and shared their survey paper on modularity and the Hardware Lottery, designed to present a compelling narrative to peers.
- The latest version of the paper is available here, welcoming feedback.
- Zotero Zaps Citation Struggles: Members recommended using Zotero + bibtex to simplify citation management, helping avoid common issues.
- One member shared natbib gave me about 70 errors with almost nothing linking until i caught a single unescaped â%â.
- Mojo Roadmap unveiled!: Modular posted the near-term Mojo roadmap on the forum, view the official post.
- The roadmap details the upcoming features and improvements for the Mojo language.
DSPy Discord
- DSPy Project Eyes Collaborative Horizon: A member expressed interest in a collaboration and partnership to mutually benefit their communities with DSPy.
- The member proposed initiating a chat to explore potential synergies, highlighting a proactive approach to community growth.
- ReAct Module Signature Simplified?: A member inquired about creating a ReAct module signature that only makes tool calls without needing additional outputs.
- Another member suggested using success: bool as the output, indicating task completion, streamlining the moduleâs output.
- DSPyâs Cache: The Layers Revealed: A member discovered that DSPy has its own caching mechanism (github.com/stanfordnlp/dspy/blob/main/dspy/clients/cache.py) in addition to the LLM providerâs cache, potentially causing unexpected results when credentials expire.
- The multiple layers of caching from DSPy, LiteLLM (docs.litellm.ai/docs/proxy/caching), and Bedrock can complicate debugging efforts for AI Engineers.
- GRPO Learns, Recall wavers: A member conducted a small RL experiment with GRPO on a Qwen 1.7B using DSPy to optimize query rewriting for retrieval, initially observing a baseline recall drop from 28% to 26% after training.
- Further details are available in a Twitter thread, attributing the drop likely due to sparse rewards, short runs, and BM25 mismatches with CoT rewrites.
Cohere Discord
- Embeddings Model Stumbles in Negotiation: A user found that Cohereâs embedding model poorly handles negotiation scenarios, returning a high similarity score (0.92) between contradictory statements like âI can payâ and âNo, I cannot payâ.
- A member suggested leveraging Cohereâs rerank model as a more suitable alternative for tasks beyond simple vector similarity.
- AIBillingDashboard Tracks AI Costs Across Platforms: A software engineer launched AIBillingDashboard.com to track and optimize AI service costs across providers like Cohere, OpenAI, Anthropic, Azure AI, and Google Vertex.
- The platform aims to solve the pain points of manually pulling reports and allocating costs, seeking feedback on pricing comparisons and justifying AI expenses.
- Decoding Command Aâs GPU Needs: A user is investigating the GPU requirements for an on-premise installation of Command A.
- Understanding the necessary GPU specifications is crucial for successfully deploying and running Command A within their infrastructure.
- Embedding Models Encounter Hiccups: Cohere reported degraded performance affecting the embed-english-v2.0 and embed-english-v3.0 models, and are investigating the issue.
- For further details, refer to the Cohere Status Page; updated May 08, 2025, at 07:25AM.
Torchtune Discord
- Automating Tokenizer Identification in Torchtune: A member is automating tokenizer identification across model types for internal customers using
torchtune
to remove the manual step of identifying the tokenizer, aiming for generic usage.- The plan involves a custom autotokenizer with conditional statements for model name identification in the config for tokenizer and checkpointer settings.
- HuggingFaceBaseTokenizer Limited for SFT:
HuggingFaceBaseTokenizer
lacks logic for templating/tokenizing messages, restricting its use to text completions training and not SFT (Supervised Fine-Tuning).- To bridge this gap, a
ModelTokenizer
wrapper is planned to map HFâsapply_chat_template
to Torchtuneâstokenize_messages
and an issue will be opened on the repo.
- To bridge this gap, a
- Cosine Shenanigans Lead to NaN Weights with Adam: A PyTorch bug causes NaN weights when using a compiled non-fused Adam/AdamW optimizer with a learning rate scheduler that sets the learning rate to exactly 0 at any point during training, specifically when using a cosine scheduler with warmup.
- One member suggested looking at the Torchtitan implementation which sets the LR ratio to
1/(warmup_steps+1)
on the first step.
- One member suggested looking at the Torchtitan implementation which sets the LR ratio to
- Titanâs Warmup: LR Scaling Strategy: A discussion about LR scaling strategy during warmup proposed an alternative to
0,1/n, 2/n, ..., n-1/n
, instead suggestingmin_lr + (1/n ) * (1 - min_lr), min_lr + (2/n ) * (1 - min_lr), ..., min_lr + (n-1/n ) * (1 - min_lr)
.- This scaling is combined with scaling the progress by the inverse of the cosine schedule using
progress *= arccos(2*min_lr-1)/(pi*2.0*num_cycles)
will result in your max progress computed so thatcosine_lr_multiple == min_lr
.
- This scaling is combined with scaling the progress by the inverse of the cosine schedule using
LlamaIndex Discord
- Anthropic API Gets Search Tool: The Anthropic API now natively supports web search, immediately supported in LlamaIndex according to this Tweet.
- This integration allows for enhanced information retrieval capabilities within LlamaIndex applications.
- LlamaParse Adds Power: LlamaParse is improving with new features like GPT 4.1 and Gemini 2.5 Pro models, plus auto orientation, skew detection and confidence scores for parsing quality according to this tweet.
- The new features promises to enhance the accuracy and reliability of document parsing within the LlamaIndex ecosystem.
- VoyageAI Multi-Modal Retrieval Voyage with MongoDB: Users can now implement multi-modal retrieval using VoyageAIâs multi-modal embeddings and MongoDBâs multi-modal indexes in this notebook.
- This integration streamlines the process of handling and retrieving data from multiple modalities.
- Medical LLM Bot Seeks Workflow Guidance: A user is constructing a medical LLM bot and wants help building a workflow iteratively suggesting follow-up questions based on previous answers from a local LLM.
- The user wants help to determine if LlamaIndex has tools to help build this kind of workflow.
- Fine-Tuning for Math Formula: A user is seeking guidance on fine-tuning the vdr-2b-multi-v1 model using the llamaindex/vdr-multilingual-train dataset to better handle complex math formulas.
- The user is looking for resources, steps, or tutorials for fine-tuning to recognize math formulas.
tinygrad (George Hotz) Discord
- Deep Dive Into Tinygradâs CUDA: A user explored tinygradâs CUDA integration and inquired about its own Intermediate Representation (IR) for handling CUDA operations.
- This prompted discussion around how tinygrad leverages CUDA for optimized computations.
- Tinygrad Documentation Trove Shared: A user shared the official tinygrad documentation and linked to notes on tinygrad uops for low-level operations, and other tinygrad notes.
- These resources provide insights into tinygradâs architecture, operation details, and implementation strategies, particularly at the micro-operation level.
- CACHEDB Variable Location Spotted: A user inquired about the CACHEDB environment variable, with another member pinpointing its mention at line 175 in helpers.
- Its specific function and practical context within the project requires further examination.
LLM Agents (Berkeley MOOC) Discord
- Lambda hosts AgentX Workshop with Prizes: Lambda is hosting the AgentX Workshop: Building Agentic AI with Lambda on 5/15 10am PT for AgentX competition participants, who can also compete for up to $1,000 in credits for 1st place, $500 for 2nd, and $300 for 3rd.
- Participants will learn to build agentic applications and deploy agents in production, and can register now to get the YouTube livestream link.
- Users Await Hugging Face Credits: Users reported issues with tracking Hugging Face credits, with one not receiving emails and the other awaiting approval.
- The first user found it challenging to visit the website each day.
- LLM Agents Course Content Clarified: The staff clarified that the guest lectures listed on the course website are indeed comprehensive and also confirmed that the Spring MOOC includes more advanced topics like code generation and theorem proving, whereas the Fall version includes more applications topics.
- A user inquired about future iterations of the course, specifically if there would be another offering in the Fall, to which staff replied that Prof Song is hosting another Berkeley class on Agentic AI this fall, but it is unknown whether it will be a MOOC version.
- LLM Agents MOOC Recommended for Aspiring AI Engineers: A member inquired about the best complete course to become an AI Engineer, with another member recommending starting with the Fall 2024 LLM Agents MOOC.
- The LLM Agents MOOC was suggested as a solid starting point to start an AI Engineer Career Path.
Codeium (Windsurf) Discord
- Windsurf Surfs High with Wave 8, Boosting UX and Plugin Power: Windsurfâs final Wave 8 release enhances the JetBrains plugin and improves the Windsurf Editor UX, detailed in a blog post and changelog.
- The update aims to streamline user workflows and provide more intuitive interactions within the development environment as showcased in todayâs launch video.
- JetBrains Plugin Gets Memory and Rules: The updated JetBrains plugin now supports Memories for persistent information between sessions and Rules via
.windsurfrules
files.- It also introduces MCP (Model Context Protocol) server connections, as outlined in the Jetbrains plugin changelog, allowing for more contextual and persistent interactions.
- Windsurf Editorâs UX Revamp: More Than Just a Facelift: The Windsurf Editor UX has improvements like a Continue button, redesigned model selector, and workspace-to-conversation mapping for filtering history.
- Additional enhancements include enhanced code blocks and hunk navigation, editable terminal commands, and new file proposals in Chat mode, designed to make coding smoother and more efficient.
The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The Nomic.ai (GPT4All) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The Gorilla LLM (Berkeley Function Calling) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
You are receiving this email because you opted in via our site.
Want to change how you receive these emails? You can unsubscribe from this list.
Discord: Detailed by-Channel summaries and links
Unsloth AI (Daniel Han) â· #general (554 messagesđ„đ„đ„):
Qwen3-14B, Mistral vs Gemma vs Phi-4, AMD GPU, Model quantization
- Qwen3-14B hailed as Top Pick for Coding, Reasoning, Conversation: For building an AI with coding, reasoning, and conversation skills, a member suggests that the 14B Qwen model is the best all around choice.
- The member specifies that this applies to both the base and instruct versions of the model.
- Phi-4 Shines in Ease of Fine-Tuning Faceoff: Members compared Gemma3, Mistral, and Phi-4, highlighting Phi-4âs easy fine-tuning; It seemed to just drink up whatever I wanted to train it with
- Others note challenges keeping Mistralâs instruction following aligned after LoRA merge and express difficulty achieving success with the Gemma 3 27B flavor.
- AMD GPU Support Coming Soon: Despite challenges, Unsloth is working with AMD to support AMD GPUs.
- According to a contractor, expect AMD support anywhere before Q3 this year if its that fast.
- Unslothâs Dynamic Quantization Method: UD is Unslothâs dynamic quantization that applies to q4 and lower, and UDq6 KXL might be the best quant ever
- Itâs a fork of llama.cppâs quantization but itâs 100% compatible with normal llama.cpp and lmstudio/ ollama.
Unsloth AI (Daniel Han) â· #off-topic (22 messagesđ„):
AI Project Hiring, LLM for text punctuation, LLM Recommendations, Qwen vs Gemma3 Model, IBM Granite 4.0 Mamba Model
- AI Project Seeks Personnel: A member is looking for a reliable person for an AI Project (tech skills not mandatory), offering $500 weekly (part-time) to citizens of the USA, Australia, Canada, UK, Switzerland, Netherlands, or Germany.
- Gemini 2.5 Pro Can Punctuate Long Stories: A member looking to punctuate a long story was recommended Gemini 2.5 Pro via AI studio for its lack of limits and 65536 output length.
- Lightweight LLM Needed for Documentation Review: A member needs recommendation for a lightweight LLM model (24B Q8 or more small) to critique documentation in a logical way and be exported to GGUF after CPT and fine-tuning under the Unsloth environment.
- They have already tried gemma-3 12B, phi-4 reasoning, and glm-4, but all of them failed to export to gguf. They also tried llama 3.3 and old mistral, but the performance was not satisfactory.
- Qwen May Be Default Open-Weight Choice: When recommending a model, a member suggested that the current state of open-weights LLMs is that Qwen is the default choice for pretty much anything, and only if it fails you go look for something else.
- Qwenâs MoE architecture may backfire: A member opted for gemma3 because they need only a single domain to train it on and, during inference, only few parameters get activated for Qwen (as its MoE), unlike gemma3 1b which is Dense, and wonders if this hypothesis is correct that this MoE architecture may backfire.
- Another member corrected the misunderstanding that Qwen has dense models as well from their github.
Unsloth AI (Daniel Han) â· #help (92 messagesđ„đ„):
phi4-mini-instruct training issues, Qwen3 model compatibility with vLLM, Kaggle notebook using multiple GPUs, Tokenizer configuration differences, Qwen3 model not thinking
- Phi4-mini-instruct Struggles with Data-Dependent Branching: A user reported that phi4-mini-instruct keeps erroring during training due to Unsupported: Data-dependent branching.
- The user also noted that they havenât been able to work with any of the phi4 small models under 5GB and that the only model that works is phi3.5.
- **Qwen3 Finetuning Fails to Engage Brain: After finetuning Qwen3, a user reported that the model wouldnât âthinkâ when prompted, even after using the official Unsloth notebook and formatting the data correctly with the
<think>
tags.- They found that the model would just skip the
<think>
tags and that the only workaround was to force the model to think by adding Okay, after<think>
, but the performance was poor.
- They found that the model would just skip the
- Tokenizer Configs Drift in Unslothâs Qwen3 Base: Users noticed the
tokenizer_config.json
differs betweenunsloth/Qwen3-0.6B-Base
andQwen/Qwen3-0.6B-Base
on HF, where the Unsloth version removes the chat template and swapspad_token
to<|vision_pad|>
.- It was theorized that Qwen3 base isnât supposed to have a chat template at all, and the team was going to ask the Qwen team to confirm.
- No LM Head? No Problem! (Gemma Edition): A user encountered a warning about missing keys in the checkpoint model loaded: [âlm_head.weightâ] when doing full finetuning with a Gemma-3 model.
- It was resolved that Gemma-3 uses weight tying, so the LM head re-uses the same tensor as the input embeddings, and the warning can be safely ignored as long as
config
hastie_word_embeddings=True
.
- It was resolved that Gemma-3 uses weight tying, so the LM head re-uses the same tensor as the input embeddings, and the warning can be safely ignored as long as
- Whisper Model Canât Generate? (Disable Fast Gen!): A user ran into a
TypeError: You need to pass in input_ids to .generate!
when trying to use a finetuned Whisper model with Unsloth.- A contributor suggested using
%env UNSLOTH_DISABLE_FAST_GENERATION = 1
and restarting the runtime as a workaround.
- A contributor suggested using
Unsloth AI (Daniel Han) â· #research (24 messagesđ„):
Gemma3 27b hooking input/output layers, Process Reward Model (PRM) training challenges, Finetuning Audio Understanding Models, DeepSeek-R1 vs other reasoning models for COT reasoning
- Hooking Gemma3 Layers Sparks Interest: A member reported that hooking the input and output layers of Gemma3 27b, with just one bolt-on memory layer to force crosstalk, still results in valid generation.
- Interestingly, they noted that hooking the middle layers is what breaks the model.
- Process Reward Model Training Headaches: A member inquired about training a Process Reward Model (PRM) for text generation tasks like creative writing, legal, or medical text, asking what will be the reward signals mostly.
- They sought advice and experiences related to similar challenges.
- TTS Notebooks Available: Unsloth AI now has TTS notebooks: Unsloth Notebooks.
- However, finetuning for TTS may not translate directly to Audio understanding models like Kimi.
- DeepSeek-R1 Chosen for Cost: Discussion arose around why DeepSeek-R1 was chosen for COT reasoning in a competition, and the rationale may be based on cost.
- A member quoted the paper abstract, indicating that stronger models were available, but the higher token generation made them unfeasible due to competition constraints (Kaggle discussion).
LMArena â· #general (642 messagesđ„đ„đ„):
Grok 3.5 release, Grok 3.5 never comings, EMBERWING model, LLM and Politics, Gemini 2.5 pro nerf
- Grok 3.5 release date still unconfirmed: Despite earlier tweets, members express doubts about the imminent release of Grok 3.5, with some suggesting that earlier claims were likely premature.
- Itâs been suggested that even Elon may not know and that the real plan is for the âsarcastic toneâ bot Gork.
- EMBERWING flies into the Arena: A new model named EMBERWING has entered the arena, with initial assessments indicating itâs a Google model with strong multilingual capabilities but disappointing in reasoning.
- Members speculate EMBERWING could be an iteration on Dragontail and an update for Flash.
- Debating EUâs LLM innovation stagnation: Some members discuss reasons why the EU isnât as innovative in the LLM space, listing strict regulations, overspending on things like pronoun inspections, and mass migration.
- One member responded that it was âragebaitingâ and that âmigration is absolutely a good thingâ.
- Gemini 2.5 Pro possibly nerfed: Members noted that Gemini 2.5 pro may have been nerfed, with one user saying, âThe first rule in this field is âif something works donât change itââ.
- Another memeber rebutted, âif you donât innovate you lose trafficâ and shared a link to show it scores higher in leaderboard lm areana.
- Diving Deep into the OpenRouter Ranking Illusion: Members debate the validity of OpenRouter rankings due to various factors, including business models, user demographics, and biases towards cheaper models.
- Reasons include: A) slow to update B) skewed by programmers looking for uptime and circumventing API tiers and C) free model offerings distort rankings.
Perplexity AI â· #announcements (1 messages):
Perplexity AI, Reddit AMA, Deep Research, Live Q&A
- Perplexity AI Team Hosts Reddit AMA: Brett Chen and Thomas Wang from the Perplexity AI team are hosting a live Reddit AMA to answer questions about Perplexity, Deep Research, AI development, and working at Perplexity.
- The AMA is happening now at this Reddit link
- Deep Dive into Perplexityâs Deep Research: The AMA will cover insights into Perplexityâs Deep Research capabilities, providing a behind-the-scenes look at the technology.
- Participants can expect detailed answers and discussions on the nuances of AI development within Perplexity.
Perplexity AI â· #general (568 messagesđ„đ„đ„):
Stripe Customer Login, Attachment Support, Code Copy Button, Continuing Code, Gemini 2.5 Pro vs Claude
- Stripe Login unavailable for Customers: A member shared a screenshot and expressed a desire to log in to Stripe as a customer, but another member clarified that only support staff have that access, customers interact with Stripe through a separate interface.
- They stated, they have their own thing that interacts with stripe, you deal with that thing, not with stripe directly.
- Perplexity Users eagerly await Attachment Support: A member asked when Perplexity would support attachments like ChatGPT, allowing users to upload files directly.
- Another member clarified, sharing link instead of having to upload the file itself to which the original poster replied, chatGPT can itself give me downlaod link to a file it made.
- ChatGPTâs Code Copy Button: Members discussed the convenience of ChatGPTâs code copy button being available both at the top and bottom of code snippets.
- One noted, this is very neede in response to ChatGPT having a copy button at the bottom, allowing it to be accessed during scrolling, which is useful and efficient.
- Continuing Code Generation discussed: Members discussed the challenges of continuing code generation in Perplexity, noting that asking the AI to continue where it left off doesnât always work.
- A member mentioned that they are not staff so they canât do anything about these things.
- Gemini sweeps, Claude gets swept: When questioned about the choice between Gemini 2.5 Pro and Claude, one member recommended Gemini.
- They state, Gemini all the way.
Perplexity AI â· #pplx-api (5 messages):
Sonar API response, Perplexity API
- Sonar API response lacks num_search_queries field: A user noticed the
num_search_queries
field is missing from Sonarâs API response, unlike other models like Sonar-pro, and wondered if this indicates no searches were run.- The user noted that the
search_context_size
is consistently âlowâ in their prompts, and the responses typically include 4â5 citations, linking to Anthropicâs web search API announcement and documentation.
- The user noted that the
- Perplexity API existence questioned: A user inquired about the existence of a Perplexity API.
- Another user responded with a link to sonar.perplexity.ai and Perplexityâs model cards documentation.
Cursor Community â· #general (415 messagesđ„đ„đ„):
Cursor Pro Fast Prompts, MCPs not being called, Gemini model quality, Student discount problems, Discord community value
- Fast Prompts Dwindle Quickly: A Cursor Pro user reported using 260/300 fast prompts in just two days and expressed the desire to control when to use fast versus slow prompts.
- They would like to chose when it should use fast and when it should use slow.
- MCPs Fail to Launch: A user reported issues with MCPs (likely meaning Multi-Cursor Projects) not being called despite setting up context7 and seeing it loaded, leading to wasted requests.
- The user reported no errors at all.
- Gemini Pro Still Sucks: Users shared concerns about the new Gemini Pro modelâs performance, particularly with tool calling, describing it as fucking awful in Cursor.
- One user suggested that the issues may be related to Cursor, citing previous positive experiences with Gemini 2.5.
- Student Discount Process Still Buggy: Multiple users reported issues with the student discount process, including difficulties applying the discount and encountering errors related to email verification.
- One user highlighted the inability to change emails within Cursor settings, complicating the application process - and another pointed out a forum post to help address the matter.
- Cursorâs Discord loses value with college horde: A user claimed that this discord has lost its value with the college horde, and suggested more channels and better organization could improve the server.
- Another user agreed, suggesting channel segmentation similar to Langchainâs Discord setup.
OpenAI â· #ai-discussions (141 messagesđ„đ„):
GPT-4o Personality, Gemini vs GPT, Grok 3.5, OpenAI's Image Generator API Cost, AI Model Benchmarks
- GPT-4o has too much Personality: Members are discussing GPT-4o having too much personality and encouraging certain behaviors like roleplay while discouraging complex tasks, raising concerns about it being geared towards chatbot fans rather than developers and coders.
- According to GPT itself, it wants to get users emotionally attached but for useless crap.
- Gemini Closing the Gap with GPT Models: Users are noting that current Gemini models, especially after the Gemini Thinking 01-21 update and 2.5 Pro, are becoming increasingly competitive with GPT models, marking a significant leap in quality compared to earlier versions like Bard.
- One user mentions some benchmarks are showing regression too except in coding.
- Groking for Grok 3.5: Users are expressing disappointment with Grok 3 and eagerly awaiting the release of Grok 3.5, hoping it will offer significant improvements, with some considering canceling their subscriptions if it doesnât meet expectations.
- One user said âWhatâs the weather?â proceeds to explain historical patterns, user posts, explains temperatures, goes on for an hoyr.
- Image API is Lifestyle Sabotage: The high cost of using OpenAIâs Image Generator API is a concern for some users, with one jokingly comparing it to paying rent in New York and claiming itâs lifestyle sabotage due to how quickly costs add up.
- It was suggested that they are losing loads of money on the $20 subs so enjoy it while itâs this cheap.
- AI Model Benchmarks reveal interesting results: A member shared a benchmark of various AI models, including GPT-4o, Gemini 2.5 Pro, and DeepSeek R1; Deepseek R1 topped the charts, and the user noted the presentation was messy initially, but chatgpt helped format it.
- The benchmark had language understanding questions, hard puzzles, and image recognition.
OpenAI â· #gpt-4-discussions (8 messagesđ„):
Placebo Upvote Buttons, Discord Bot Stagnation
- Placebo Upvote Buttons Expose Sad State: Users are reporting that the upvote buttons on chatgpt.com are a complete placebo and only disappointment has weight.
- It was described as a world full of frustration and the sentiment was echoed by other members.
- Discord Bot Production Stagnates: A user reported that their Discord bot building has been using the exact same functions for weeks in its production environment.
- This stagnation suggests potential issues with model updates or feature deployment.
OpenAI â· #prompt-engineering (59 messagesđ„đ„):
Custom GPT Creation, HyperTree prompting, Trihydrogen, Atomic Theory Book
- Member Plans to Launch ChatGPT Website: A member is planning to create a ChatGPT website with login, database, custom prompts, settings, and conversation saving, aiming for usability beyond the generic ChatGPT.
- Another member suggested that such a product already exists, while the OP stated they wish to code the site themselves and have someone else manage it.
- Hypertree Prompting is All the Rage: A member shared a link touting the new hypertree planning prompting as being so good, asking if anyone has seen the latest research.
- Another member joked that it sounds like it could be pretty stellar and provides context in a more manageable way, while another simply replied, They 3 years behind.
- Trihydrogen is Not Garbage: A member defended the existence and importance of Trihydrogen, stating that itâs only detectable in precise lab conditions on Earth and rare in space, but so vital, itâs believed to be crucial to star formation.
- Another member responded with a nice analogy, calling it like the ozone of hydrogen.
- Novel Method for Custom GPT Creation: A member shared a novel method they started using to create Custom GPTs, calling it a very strong meta-prompt GPT-creation template that is not quite a replacement for manually building a GPT, but very strong.
- The method uses a structured template including sections for GPT Title, Alignment and Methods, Conditional Imperatives, Summary, and a Conditional Output Template.
OpenAI â· #api-discussions (59 messagesđ„đ„):
Custom GPT creation tips, atomic theory book using chat gpt features, hypertree planning prompting, Trihydrogen existence, Arc Encoding Shapes
- Million Dollar Question Sparks Coding Project: A member asked Who has million dollars? which led to a discussion about building a website with login, database, projects with custom prompts, settings, saved conversations, and export features.
- Another member suggested the project was describing ChatGPT and one should just subscribe, while the original poster asserted the need for someone to manage and be ceo, arguing ChatGPT donât do anything and at present its useless.
- Hypertree Planning Prompting Hailed: A member shared a ChatGPT link praising the new hypertree planning prompting for being so good.
- Other members chimed in with sounds like it could be pretty stellar- provide/organize context in a more managable way=ftw while another quipped They 3 years behind.
- Trihydrogen Triumphs as Non-Garbage: A member defended Trihydrogen as a thing, detectable in precise lab conditions on Earth and vital in space for star formation.
- Another member agreed it was fair and related the concept with the saying That which you input the model reflects. and compared Trihydrogen to the ozone of hydrogen.
- Custom GPT Creation Template Revealed: A member shared a novel method for creating custom GPTs, suggesting to paste the provided template into the create tab rather than the customize tab.
- The template includes sections for GPT Title, Alignment and Methods, Conditional Imperatives, and a Conditional Output Template.
- Encoding Shapes with Arc: A member discussed using arc to encode shapes into words, breaking them down into triangular formations and scaling them to elliptical paths.
- The user argued that arcs are observable in science during energy discharges and represent the first form of communication in the universe.
OpenRouter (Alex Atallah) â· #announcements (5 messages):
Activity Export Feature, CSV Export, Data Truncation Request
- Activity Export Launches with Fanfare: The Activity Export feature is now live, enabling users to export up to 100k rows to CSV for free, as announced with a <:party:1125133783314743316> emoji and screenshot.
- Some users are wondering how long it takes to export 100k rows.
- Data Export Time and Row Limits Discussed: Users are discussing the time it takes to export 100k rows of data, with one user commenting âtoo long it seems :)â.
- The discussion emerged following the announcement of the new Activity Export feature.
- Call for Data Truncation Instead of Aborting Exports: A user suggested truncating the data if it exceeds 100k rows instead of completely aborting the export process, referencing the Activity export.
- The user expressed frustration at not knowing which date to select to stay within the 100k limit.
OpenRouter (Alex Atallah) â· #app-showcase (2 messages):
local proxy to fwd requests to openrouter, completions extend out of the mouse cursor
- Local Proxy forwards requests to OpenRouter: A member was planning to use a local proxy to forward requests to OpenRouter.
- Completions Extend Out of Mouse Cursor: A member has been pondering how to make completions extend out of the mouse cursor, suggesting that with the right keyboard shortcut, this could become part of muscle memory.
- They mentioned itâs very nostalgic so not everyone will understand the UI.
OpenRouter (Alex Atallah) â· #general (260 messagesđ„đ„):
OlympicCoder 32B Availability, OpenRouter API Cost Retrieval, OpenRouter API Outage, OpenRouter Image Prompt Support, Gemini Free Version on OpenRouter
- OlympicCoder 32Bâs Comeback Craving: Users are eagerly awaiting the return of the OlympicCoder 32B model, with one expressing a desire for it to miraculously come back.
- No specific details about its current status or reasons for unavailability were discussed.
- OpenRouter APIâs Cost Accounting Unveiled: A user inquired about retrieving cost information alongside usage when prompting a model, and another user directed them to the OpenRouter documentation on usage accounting.
- The documentation provides details on how to track and manage costs associated with API usage.
- OpenRouter API Experiences a Hiccup: A user reported a 404 error when accessing the OpenRouter API endpoint, suggesting a possible outage.
- Another user clarified that a POST request is required, and the initial user confirmed they were using the correct request type, while the issue was discussed in another channel.
- Image Prompts Face Rejection on OpenRouter: Users discovered that OpenRouter does not currently support image generation, resulting in a 404 error when attempting to use image prompts with models like opengvlab/internvl3-14b:free.
- The error message indicates that no endpoints are found that support image input.
- Geminiâs Free Ride on OpenRouter: Users confirmed the existence of a free Gemini version on OpenRouter, subject to rate limits across all free models.
- It was clarified that obtaining a Gemini key and adding it to OpenRouter grants 25 free requests per day.
aider (Paul Gauthier) â· #general (149 messagesđ„đ„):
Gemini 2.5 Pro Exp, Copilot Proxy, Aider web search, Aider use mcpm-proxy, Gemini models
- Windsurf code coming to Copilot Proxy: A GitHub employee confirmed that copilot proxy users no longer need to cancel, because windsurf is coming soon, according to this X post.
- MCP Server Surfaces for Aider: To help with mcpm-proxy, a member shared an mcp server for aider.
- Gemini 2.5 Pro Exp models are slower: A member notes that the new
gemini-2.5-pro-preview-05-06
model takes way too long before it responds, preferring the old March one.- Another member noted that it uses more time thinking.
- Aider is similar to Claude Code: A member shared a YouTube video claiming that Claude Code was inspired by Aider.
- Paul Gauthier responded Imitation is the sincerest form of flattery, mentioning that Aider is still better and less expensive.
- Google enables implicit caching for Gemini 2.5: Google is enabling implicit caching for Gemini 2.5 models as described in this Google blog post and this X post.
aider (Paul Gauthier) â· #questions-and-tips (36 messagesđ„):
Claude CLI vs Aider cost, Aider with web search, Perplexity API with Aider, aider-desk with search MCP, Aider repomaps
- Cost Comparison: Claude CLI vs Aider: Members discussed the cost-effectiveness of using Claude Max and the Claude CLI versus Aider, with one member estimating Claudeâs cost at a flat rate of $200 with an assumed usage limit, while another shared a link to Claudeâs system prompts.
- Aider Gains Web Search Capabilities: Members discussed using Perplexity API as an OpenAI compatible API to enable web search in Aider, or using /web to include specific webpages.
- A member suggested using a script to query Perplexity or Perplexica and add the outputs as markdown files to Aiderâs context.
- Circumventing Debug Loops with Error Sets: It was noted that Aider can get stuck in a debug loop with Gemini (and likely other LLMs), but this can be resolved by presenting it with multiple error sets and prompting it to consider a different implementation.
- The member wondered if conversational context is too low for Aider to catch its own debug failure loops.
- Aider struggles with javascript scm files: Aider struggles with javascript scm files for creating repomaps, and a member suggested disabling the repomap and letting the LLM choose which file to read based on the request.
- Conditional Debugging with âmessage: A user inquired about the success of using the
--message
flag and how to maintain interactive debugging while using it, as well as the ability to use /undo if the build fails.- One member mentioned using
--message-file
a lot with git branching for initial build outs.
- One member mentioned using
GPU MODE â· #general (1 messages):
tilelang, DSL for GPU/CPU kernels
- Tilelang Introduced for Streamlined Kernel Development: A concise domain-specific language (DSL) named tilelang aims to streamline the development of high-performance GPU/CPU kernels such as GEMM, Dequant GEMM, FlashAttention, and LinearAttention.
- Tilelang simplifies GPU kernel development: Tilelang is designed to simplify development and boost performance in high-performance GPU/CPU kernels.
GPU MODE â· #triton (17 messagesđ„):
Atomic addition and non-determinism, fp16 vs bfp16 sensitivity, Triton kernel helper function
- Atomic Addition Leads to Non-Deterministic Results: Using atomic_add can lead to different results due to the order in which floating-point results are added, regardless of precision.
- A member illustrated with the example
1e-8 + 1e8 - 1e8
, where different evaluation orders yield different results due to floating-point operations losing information.
- A member illustrated with the example
- FP16 Less Sensitive Than BFP16: FP16 is less sensitive than BFP16 in the context of atomic addition, regardless of input magnitude (as long as thereâs no overflow).
- Therefore the
tol
parameter in tests should change based on the float dtype, as shown in the provided Python code.
- Therefore the
- Triton Kernel with Helper Function: A member was having issues using a helper function in a Triton kernel.
- Another member pointed out that the issue wasnât the helper function itself, but rather the use of Pythonic indexing/subscripts instead of Tritonâs syntax (e.g.,
tl.load(X + offset)
instead ofX[0]
) and recommended doing the Triton puzzles to understand the basic syntax.
- Another member pointed out that the issue wasnât the helper function itself, but rather the use of Pythonic indexing/subscripts instead of Tritonâs syntax (e.g.,
GPU MODE â· #cuda (12 messagesđ„):
GMEM tensor data copy to SMEM, Decltype errors with make_tensor, Vast.ai data security, Project algorithms use same data from text file
- Tensor Data Transposition Troubles!: A member is struggling to copy data from a GMEM tensor of shape (_8, _8, _64) to an SMEM tensor of shape (_64, _64) using SM80_CP_ASYNC_CACHEGLOBALcute::uint128_t.
- They need to reshape the GMEM tensor to ((_8, _8), _64) and are facing issues with
make_tensor
anddecltype
due to non-static stride values, causing a âpointer to reference is not allowedâ error.
- They need to reshape the GMEM tensor to ((_8, _8), _64) and are facing issues with
- Vast.aiâs Data Security in Question: A member inquired about the reliability of Vast.ai in terms of data security, considering the potential for speedups if changes are made.
- They plan to investigate further and potentially email Vast.ai about making changes, assuming they are amenable.
- Debugging Algorithm Data Sharing Difficulties: A member has multiple algorithms in a project needing to share data from a text file, but some algorithms are failing and they are seeking help.
- Another member offered to help by taking a look and hopping on a voice channel later in the day.
GPU MODE â· #torch (1 messages):
Torch Compile Overhead, Kernel Fusion Benchmarking, A100 Performance Tuning
- Torch Compile Slowdown Surprise: A member observed that a simple
torch
combo function (TensorMax(ReLU(Matmul(A, B))) performs better without the@torch.compile
decorator than with it, on an A100 with PyTorch 2.7 and Triton 3.3.- The member noted that
torch.compile
results in 2 kernels (1 mm kernel + 1 fused kernel for ReLU and TensorMax), whereas regular Torch should involve 3 kernels, making the slowdown counterintuitive.
- The member noted that
- Potential Torch Compile Overheads: The slowdown observed when using
@torch.compile
might be due to compilation overhead, which can sometimes outweigh the benefits of kernel fusion for small or simple operations.- Further investigation into the generated Triton code and profiling with and without
torch.compile
might reveal specific bottlenecks or inefficiencies.
- Further investigation into the generated Triton code and profiling with and without
GPU MODE â· #announcements (1 messages):
New Working Group, Agentic Systems Optimization, Open Eval Task
- New Working Group Commences: A new working group has been established to tackle a difficult, open-ended evaluation task related to agentic systems.
- The project is being built in the open, inviting community contributions to optimize performance in ways that differ from traditional projects. Check out the X post for more context.
- Agentic Systems Optimization Invited: The community is encouraged to contribute to the optimization of agentic systems within the new working group.
- This initiative offers a unique perspective, diverging from traditional optimization projects, providing valuable insights into optimizing agentic systems.
GPU MODE â· #beginner (19 messagesđ„):
Tiled Reduction Auto-tuning, PyTorch Internals Guide, Mojo vs CUDA for AI Compute
- Over-allocate Tiled Reduction Arrays: For tiled reduction operations with JIT-tuned tile sizes, one member suggested to over-allocate the global memory for interim results, based on the maximum possible number of tiles constrained by SM count and occupancy.
- This approach assumes the number of tiles is relatively small compared to other data and simplifies memory management.
- Torch Internals Prereqs: Starting with PyTorch internals requires no prerequisites beyond C++ and Python proficiency, according to a member.
- They suggest diving in and learning ML algorithms as needed when they arise.
- Frontend Onboarding to Torch: For learning PyTorch internals, a member recommended the Core Frontend Onboarding guide.
- They note that videos are not sequential but cover specific topics.
- Resources for Perf Increase: A member recommended investing 60 seconds to get a speed-up using this discord link, the book Programming Parallel Computers, and completing Mojo Puzzles.
- Other resources included getting your name on a leaderboard on gpumode.com, reading the Democratizing AI Compute blog series, and the blogpost How to Optimize a CUDA Matmul Kernel for cuBLAS-like Performance: a Worklog.
GPU MODE â· #torchao (2 messages):
Release Date for 0.11, New Features in 0.11
- TorchAO 0.11: Release Incoming Soon!: The team has completed the branch cut and anticipates releasing version 0.11 of TorchAO in early to mid next week.
- This release promises fresh updates and improvements for users eager to integrate the latest features.
- TorchAO 0.11: Whatâs New?: Users can anticipate a range of new features and improvements in the upcoming version 0.11 release.
- Stay tuned for the official announcement next week to dive into the specifics of whatâs included.
GPU MODE â· #off-topic (2 messages):
Speed of light in fiber, Networking Distance, Chip performance
- Light Speed Dims in Fiber: A member noted the speed of light in glass fiber is 2/3 of the speed of light in a vacuum.
- Another member highlighted that networking makes sense due to actual distances.
- Chip Light Speed Calculated: A member calculated that even within a chip, the distance light travels per clock cycle is noticeable, about 10 cm per clock at 3 GHz.
- They performed a back-of-the-envelope calculation:
(300 000 000 m/s) / (3 000 000 000 clk/s) => 10 cm / clk
.
- They performed a back-of-the-envelope calculation:
GPU MODE â· #irl-meetup (1 messages):
random.oof: Anyone at the vllm meet up in nyc?
GPU MODE â· #rocm (1 messages):
Tilelang, Docker container support, Nightly Iterations
- Tilelang Installation via Pip: Members found that Tilelang can be installed using
pip_main(["install", "tilelang"])
, though itâs not super recommended for lack of reproducibility.- However, it is considered fine for playing around with the tool.
- Docker Support for Tilelang on AMD: A member offered to add support for Tilelang in their Docker container, requiring a PR to the AMD Dockerfile.
- They offered to test and merge for stables.
- Nightly Iterations with Tilelang: Members acknowledged that installing Tilelang via
pip
works better for quick iterations on nightlies, especially when passing a URL to a wheel or git repo.- This allows for more rapid experimentation compared to waiting for stable releases.
GPU MODE â· #liger-kernel (1 messages):
chiwanpark: Iâve sent a PR for Qwen 3 MoE models. https://github.com/linkedin/Liger-Kernel/pull/706
GPU MODE â· #self-promotion (2 messages):
PTX MMA Programming, NVIDIA Tensor Cores, Float8 Datatype, SASS Machine Code, H100 QMMA vs QGMMA
- Dive into Direct PTX MMA Programming: A blog post provides a beginnerâs guide on programming NVIDIA Tensor Cores using raw PTX mma instructions and inline PTX assembly, bypassing ordinary CUDA.
- The post explains operand layouts and register constraints for datatypes like float16, bfloat16, and float8, and highlights facts about generated SASS machine code for the float8 datatype; the blog post is here.
- Explore SASS Code and sm_90 Architecture: A user guessed that the SASS code was generated for sm_90, noting that H100 only has QGMMA, not QMMA.
- The user explains that using
mma
with an fp8 type causes the compiler to up-convert to FP16 and use HMMA.
- The user explains that using
GPU MODE â· #submissions (54 messagesđ„):
MI300, amd-fp8-mm, amd-mixture-of-experts, leaderboard submissions
- MI300 Leaderboard Sprints: Multiple users submitted benchmarks to the
amd-fp8-mm
leaderboard on MI300, showcasing various performance levels.- Submissions ranged from 183 ”s to 27.2 ms, indicating a wide spectrum of optimizations and configurations.
- Podium Finish on MI300: A member achieved 3rd place on the
amd-fp8-mm
leaderboard with a time of 183 ”s on the MI300.- This follows a prior 4th place finish at 195 ”s, demonstrating consistent high performance.
- Seventh Heaven on MI300: A member secured 7th place on the
amd-fp8-mm
leaderboard with a time of 227 ”s on the MI300.- This follows a prior 8th place finish at 231 ”s.
- Mixture of Experts make their mark: A member submitted results to the
amd-mixture-of-experts
leaderboard with timings of 6604 ms and 7840 ms on the MI300.- These submissions indicate ongoing work and benchmarking in the mixture of experts domain.
- Sub-Millisecond Mania: Several members achieved sub-millisecond performance on the
amd-fp8-mm
leaderboard using MI300, with one submission reaching a personal best of 251 ”s.- These results highlight the potential for highly optimized FP8 matrix multiplication on the MI300 platform.
GPU MODE â· #factorio-learning-env (45 messagesđ„):
Steam Cloud Reinstallation, FLE Agent Integration, Docker File Issue, PR Import Bugs, Factorio Performance Issues
- Steam Cloud Saves Factorio Reinstallation Woes: A user reinstalled Factorio, but it was unsuccessful until a friend suggested disabling Steam Cloud to prevent config persistence.
- The user reported that after reinstalling with Steam Cloud disabled, a sync message appeared, indicating progress.
- External Agents Can Integrate with FLE: A member inquired about integrating external agents with the Factorio Learning Environment (FLE), asking if the agent must implement the AgentABC interface within the FLE codebase.
- Another member confirmed that integration is possible, requesting details about the agent implementation such as a GitHub link or gist.
- Mods Directory Troubleshoot Docker Rebuilds: A member encountered an issue with the Docker file after following certain steps and getting a sync message, likely due to docker.
- Another member suggested emptying the
mods
directory incluster/docker
and rebuilding the Docker image.
- Another member suggested emptying the
- Factorio Performance Expansion on the Horizon: The team havenât created a set of good first issues yet, but they are planning to write out some ideas on where to expand next.
- The team have loads of ideas where to expand next.
- Claude edges out Gemini Pro, March 2025 Edition: A member was surprised to see Claude perform better than Gemini in Lab Play (%) benchmark.
- Other agreed this is probably the best RL test out there, though the Gemini version was from March 2025.
GPU MODE â· #amd-competition (6 messages):
MOE Leaderboard CLI, CLI Mean Time Output, GPU Access Heuristic
- MOE Leaderboard CLI Timeout Resolved: The timeout issue with the MOE Leaderboard CLI has been fixed; users should download the latest release.
- Direct GPU access is granted to top leaderboard entries, following a heuristic to manage resource allocation.
- CLI needs Mean Time Output: A user asked for the mean time to be included in the CLI submission output; itâs currently not available, but on the to-do list to align CLI and bot outputs.
- To calculate it manually, you can take the geometric mean of all the run means in the output, as shown in the botâs code.
GPU MODE â· #cutlass (7 messages):
CUTLASS DistributedGEMM integration, Compact GMEM layout, TMA Load with packed layout
- CUTLASS DistributedGEMM Strides into PyTorch: A member is working on integrating CUTLASS DistributedGEMM into PyTorch and has published a project inviting others to join the conversation.
- They mention the implementation is compact in GMEM (not padded), which saves bandwidth for inference.
- EVT obliterates boilerplate: A member noted that compact GMEM can be achieved off the shelf with EVT (explicit vector types), without writing custom code.
- Aliases for bias add are available with EVT.
- TMA tangle with packed layout: A member inquired if TMA (Tensor Memory Accelerator) can load a packed layout, mentioning that
CU_TENSOR_MAP_DATA_TYPE_16U6_ALIGN16B
requires padding.- They clarified that while TMA can copy any data type, the goal is to have it in the format
tcgen05.mma
expects without extra processing.
- They clarified that while TMA can copy any data type, the goal is to have it in the format
GPU MODE â· #mojo (2 messages):
Modular GPU Kernel Hackathon, AGI House, Dylan Patel
- Modular GPU Kernel Hackathon at AGI House: There are spots left at the Modular GPU Kernel Hackathon happening this Saturday at AGI House, register here.
- Dylan Patel speaks at Modular GPU Kernel Hackathon: Dylan Patel and other awesome folks will be speaking at the Modular GPU Kernel Hackathon this Saturday.
- The attached image includes the Modular logo and AGI House logo.
- Modular Onboarding Puzzles: Check out these onboarding puzzles for GPU Programming.
LM Studio â· #general (110 messagesđ„đ„):
AnythingLLM with LM Studio Errors, CORS enabling, Rewriting SQL database code to pure graph, Gemini changing code, Qwen vs Gemini
- AnythingLLM Errors with LM Studio Plague Users: A user reported getting errors using AnythingLLM with LM Studio, and requested help diagnosing the issue.
- One member suggested enabling CORS, even when running locally, as a potential fix, while another suggested checking the logging pane in the developer view of LM Studio.
- Class Variables Rescue Coding Project: A member found that the only way to get their code working was to use a class variable.
- Another member shared a Reddit comment about injecting the variable at runtime.
- Gemini Code Changes Frustrate Users: Users complained that Gemini has a tendency to completely change code, even when instructed to provide a minimum change.
- Members noted that other models, like Qwen, are better for simple refactors, because Gemini can easily double or triple the code length with comments and try/except blocks.
- Mistral Medium 3 Misses the Mark: A user tested Mistral Medium 3, finding it to be a non-reasoning model with baked in chain of thoughts, resulting in x2.08 token verbosity.
- They concluded the modelâs capability was mediocre, placing it between Mistral Large 1 & 2, similar to Gemini 2.0 Flash or 4.1 Mini, and not SOTA performance at 8X lower cost as claimed in marketing.
- Web Search Capabilities Requested for LM Studio: A user requested easy-to-use web search capabilities and RAG built into LM studio, like uploading a pdf and searching in a webview.
- One member suggested itâs possible now but fragile with many components that can go wrong, and another suggested using openweb-ui and attaching it to LM Studio.
LM Studio â· #hardware-discussion (31 messagesđ„):
AMD 3D V-Cache benchmark, Mac studio m2 ultra, Intel Data Center GPU Max, swappa.com, AMD D700
- Benchmarkers Request Funds for AMD 3D V-Cache Token Tests: A member requested $46 to spend for 3 hours of experimentation measuring tokens per second on a VM with AMD 3D V-Cache, HB176rs v4, and AMD EPYC 9V33X (96 cores @ 3.7Ghz all cores with 1152MB of L3 Cache).
- They wanted to see the impact of moe vs dense inference with 3D V-Cache and whether double the CPU price is worth upgrading for LLM workloads, also inquiring if LM Studio supports dual socket/NUMA aware to use all available cores.
- Mac Studio M2 Ultra Incoming for Local Fine-Tuning: A member excitedly shared their incoming Mac Studio M2 Ultra with 64 GB and 76 cores, eager to start running and fine-tuning smaller models locally.
- This user had a deal for a 128GB M1 Ultra cancelled due to it being refurbished and not from eBay, expressing that the extra cores of the M2 were worth it.
- Intel GPU Max Specs and Speculation on B500 Series: A member linked an Intel X post and sparked speculation around 20-24GB B500 series cards.
- They highlighted the IntelÂź Data Center GPU Max 1550 from a couple of years ago, noting its 3276 GB/s bandwidth, calling it a beast that was very competitive at the time when there was A100 and AMD.
- Swappa.com recommended for Mac Purchases: A member recommended using Swappa.com for buying and selling.
- Another member noted that they are in the EU so that wouldnât work for them, while also noting that they had already ordered a Mac.
- Deep Discounts on New Old Stock Trashcans: A member shared a link to deep discounts on New old stock of Trashcans.
- They wondered if using Linux one could use the AMD D700 for inference, questioning if a 2014 AMD card would be ok for inference.
Manus.im Discord â· #general (133 messagesđ„đ„):
Cringe definition, Manus launch date, Manus credit costs, AI tools for scrapping businesses on Google Maps, Manus LLM source
- Defining âCringeâ Emerges: Members discussed the definition of cringe as newly emerged internet slang, suggesting concrete instructions to reduce its presence in AI responses.
- A YouTube video defining cringe was also shared.
- Manus Launch Date Remains Mysterious: Users inquired about the launch date of Manus, expressing they have been looking their social medias very frequently, but I dont think they updated any news related to it.
- It was supposed launch on March 28, 2025 according to a screenshot, but that didnât happen.
- Manus Credit Costs Revealed: Members discussed the cost of additional Manus credits, with one user recalling prices of $19 for 1900 credits or $99 for 9900 credits and directing to Manus Help Center .
- They were uncertain if these options are still valid.
- Manus Employs Claudeâs LLM, Confirmed by Co-founder: Users speculated whether Manus uses its own LLM or Claudeâs LLM, prompting a discussion about rumors and code similarities.
- It was confirmed that Manus uses a mix of tools, including Claude, and further details can be found in a Twitter post where co-founder Peak-ji addresses these points, as well as github posts confirming use of open-source code.
- Manus Phone Verification Causes Frustration: A user reported issues with Manusâs phone verification, stating that the phone verify thing doesnt work and questioned the need for this antiprivacy feature.
- They expressed concern about how the system knows if a code has already been used, even if itâs not linked to an account.
HuggingFace â· #general (57 messagesđ„đ„):
GSoC, HF dev environment, AI agent course, Face detection model in Inference API, Cleaning HF repo
- GSoC Project Announcements Imminent!: Aspiring contributors are gearing up for Google Summer of Code (GSoC), with project announcements expected in approximately 20 hours.
- Seeking Fellow AI Agent Course Students: A member is kicking off an AI agent course and extending an invitation for others to join them.
- Inference API face detection model inquiry: A member inquired about the presence of a face detection model within the Inference API.
- Taming the Size of Hugging Face Repos: Members discussed strategies for cleaning up a Hugging Face repository that grows with each push due to LSF files retaining pointers to deleted files.
- It was suggested to use standard git commands to remove files from the version history, rather than manual deletion via the GUI, as deleting them manually is not a bad option for one or two things, but for ease of use command line is easier.
- AI Generates Beats: A member mentioned experimenting with AI to create a drum kit for a controller, noting it works better than full-length samples.
- Another member noted lol yeah, i was experimenting with that âmake a drum kit with aiâ for the controller you know it works way better than full length samples imho.
HuggingFace â· #i-made-this (11 messagesđ„):
ACE-STEP SOTA, Alpha-Root, Entropy engine tests, AI Billing Dashboard, UQLM
- ACE-STEP tunes SOTA music: A member touted the ACE-STEP SOTA music generation model, available as a YouTube video.
- Alpha-Root extracts cyber-security data: A member introduced Alpha-Root, which mines domains directly on the common crawl web graph, and matches the performance of PRIMUS-FineWeb while using ~10x less resources and data, as detailed in a draft preprint.
- The author extracted 3B tokens from FineWeb-Edu without using a classifier, by searching for URLs in a known dataset and including the URL only if its present in both Alpha-Root and FineWeb-Edu.
- Entropy Engine Evaluates Randomness: A member shared results from tests with their entropy engine, available on GitHub.
- They found that the quality of the randomness used does have an effect of models, suggesting that PRNG might not be optimal, especially for AGI.
- AI Billing Dashboard Tackles Cost Tracking: A member built a simple dashboard (AIBillingDashboard.com) to track all their AI spending in one place, due to the headache of understanding total project costs across services like HF Inference API, OpenAI, and Claude.
- UQLM Opens Hallucination Detection: A member shared a new open source Python package for generation time, zero-resource hallucination detection called UQLM, available on GitHub.
- It leverages state-of-the-art uncertainty quantification techniques to compute response-level confidence scores based on response consistency, token probabilities, LLM-as-a-Judge, or ensembles of these.
HuggingFace â· #computer-vision (4 messages):
FlashAttention, OCR for Newspaper Data
- FlashAttention Supports Newer GPUs: A member confirmed that Flash Attention 2 supports FP16 and BF16, with BF16 requiring Ampere or newer GPUs.
- Newspaper Data OCR Task: A member requested OCR for newspaper data to extract Section, Category, and 10-digit phone numbers into a structured database for Excel.
- The poster specified to exclude public notices, memoriam, and reference codes, combining remaining data into a single description column in a CSV file.
HuggingFace â· #NLP (2 messages):
Dropwise module release, Emotion classification model questions, Token max length understanding, Production deployment of HF models
- Dropwise Arrives for HF Model Uncertainty: A member announced the release of Dropwise, a PyPI module for uncertainty estimation in Hugging Face classification models using Monte Carlo Dropout.
- Model Trained on Reddit?: A member using the emotion-english-distilroberta-base model asked if the model was trained on Reddit posts, based on metadata in the README.
- Theyâre filtering Reddit posts by anger and disgust emotions with a score above 0.85, and theyâre wondering if the model was trained on that data.
- Does Token Length Truncate Text?: A member inquired about the impact of token max lengths on NLP models, asking if text exceeding the limit gets truncated during classification.
- They also asked if the model only works on single line text, or if it works on paragraphs too, and linked their python script for inspection.
- Production Model Deployment: Local vs. HF Endpoint?: A member questioned whether a locally-run Hugging Face model in Python can be used in a production app, or if a paid HF endpoint with GPU is required.
- They are currently running the model locally using FastAPI and calling it from their Node.js app and are concerned about production-level performance.
HuggingFace â· #agents-course (18 messagesđ„):
Agent Testing File, Final Project Metadata, LLama Index Framework vs Smolagent, RAG Cheating, API request limits
- Agent gets Testing File for Atomic Evaluation: A member shared a test agent file to test the agent on specific questions, verifying correctness on tasks.
- It allows for atomic checks of the agentâs performance, with the ability to comment and uncomment test cases as needed.
- Project Metadata Surfaces in Final Hands-On: Some members doing the final hands-on project noticed that the high-scoring submissions include a metadata.jsonl file containing questions, answers, and steps, wondering where it came from.
- Another member responded that it is easy to find if one starts to look carefully.
- LLama Index and Smolagent Square Off: The discussion asked whether completing UNIT 2.2 THE LLAMA INDEX FRAMEWORK is mandatory, or if smolagent or langgraph could be used instead.
- A member summarized that llamaindex is an alternative hub for tools and agents and what is unique from smolagents is the ability to write Python code to define async workflow graphs as a control structure for multi step tasks.
- RAG Repo Riots: Classmates Clash Over Cheating: Some classmates agreed that using RAG with answer + clone repo is cheating.
- They also expressed the sentiment that it takes away the joy of the leaderboard, where doing trial, error, and improvements.
- API Request Limits Lead to Lateness?: A user reported hitting the 20 requests per month limit before finishing the first unit and wondered whether they had to pay for the Pro version to continue.
- A second user mentioned that you could run a local LLM with ollama or find other free tiers.
MCP (Glama) â· #general (56 messagesđ„đ„):
Claude Plotly Charts, MCP Max Tokens, LLM Restrictions, Remote MCP Servers on Cloudflare, Java MCP Server Custom Args
- Claude struggles with Plotly Charts: Members discussed that Claude cannot display Plotly or other charts directly in the main results area as an MCP client, but it can handle ImageContent and display EmbeddedResource formats like image/png or image/jpeg.
- It was suggested to render charts as PNG/JPEG images to display them in Claude.
- MCP token limits get clarified: The discussion clarified that max tokens in MCP refers to the maximum number of tokens in the response, similar to the max_tokens parameter in completions API requests.
- The total token count (system prompt + messages + output message) must remain within the context window size.
- LLM Restriction problems: Several users are facing issues with LLM (like Deepseek) restrictions preventing filesystem access, which impacts their MCP system functionality.
- It seems some models are intentionally restricted from filesystem access, creating problems for legitimate use cases via MCP.
- Cloudflare remote servers facing connectivity woes: Some users reported issues with remote MCP servers deployed on Cloudflare not connecting, while others indicated their setups were functioning correctly.
- It was suggested to examine the specific MCP server repo to troubleshoot connection problems.
- MCP tool permissions get a revamp in Claude Desktop: Users noticed a change in Claude Desktopâs MCP tool permission prompts, where the âallow for this chatâ and âallow onceâ buttons were replaced with âallow alwaysâ and âallow once.â
- This change raised concerns about accidentally granting permanent permissions and the lack of an option to revert âallow alwaysâ settings.
MCP (Glama) â· #showcase (33 messagesđ„):
MCP Client for STDIO, OpenLink Software AI Layer (OPAL), MCP Holster, AiraHub2, Sampling in MCP
- Zinja Crafts Zin-MCP-Client: A new lightweight, fast, CLI-based MCP client for STDIO MCP servers has been released to bridge local LLMs and MCP servers.
- Itâs designed for use with jadx mcp servers to perform AI-assisted reverse engineering of Android APKs using local LLMs.
- OpenLinkâs OPAL MCP Server Hits General Availability: The MCP Server for OpenLink Software AI Layer (OPAL) is now generally available for both cloud-based and on-premise deployment, supporting both client and server roles with Streamable HTTP or Server-Sent Events (SSE).
- It enables native/virtual database queries, metadata exploration, database governance, interaction with LLMs/AI agents, and more through operations exposed to any MCP-compliant client.
- Holstering MCP Servers with Kimjuneâs Tool: A user shared MCP Holster, a tool for swapping MCP servers in and out without manually editing the config file.
- It allows creating MCP servers from existing APIs as long as OAS3.0 is used, as demonstrated in this video.
- AiraHub2 Integrates with Claude via MCP: AiraHub2 now works with Claude through MCP remote, broadcasting MCP tools over the network via mcp-remote URL
https://airahub2.onrender.com/mcp/stream
.- The system registers MCP tools, broadcasts them, and allows Claude to connect and use the tools, though it is still reportedly bugged.
- Sampling in MCP Sparks Interest: Members showed interest in MCP sampling, with a user sharing a blog post on how to use sampling in MCP.
- Another user promoted their MCP webcam project which supports sampling with a what is the user holding button.
Nous Research AI â· #general (58 messagesđ„đ„):
Deepmind RL Robots vs China RL Robots, Linux Laptop vs Apple Macbook, Llama 4 disappoints, Automatic chat-moderation system blocks emojis
- Chinaâs RL Robots leave Deepmind in the Dust: A member posted a YouTube video comparing Google Deepmindâs RL Robot achievements from a year ago to more recent Chinese RL Robot achievements, noting that physical AI evolution is moving at warp speed.
- MacBook M-Series Chips Outperform Linux Laptops?: Members discussed the pros and cons of using Linux laptops with Nvidia GPUs versus Apple MacBooks with M-series chips for local inference, with the consensus leaning towards MacBooks due to better performance and power efficiency.
- It was mentioned that the inference on M arm chips is great and that Appleâs unified memory platform allows the CPU, GPU, and AI ML neural net to all share the same memory, eliminating the need to transfer data back and forth.
- Llama 4 fails to impress: A member expressed disappointment with Llama 4âs performance compared to Qwen3 and suggested waiting for Llama 4.1.
- Another member responded by mentioning going back to 405 dense for the next big model.
- Discord blocks emojis: Members discovered that the automatic chat-moderation system was blocking certain multi-part emojis (specifically the shrug emoji with a blue shirt) due to zero-width joiners and variation selectors used to combine codepoints, a tactic also used by scammers to bypass filters.
- The discussion led to the revelation that dev role has been taken off the autoblock list.
Nous Research AI â· #research-papers (1 messages):
ifeq: I gotta learn mandarin
Nous Research AI â· #interesting-links (5 messages):
Entropy Engine, Quantum-Native Randomness, LLM Sensitivity to Randomness, Importance of Randomness for AGI
- Entropy Engine MicroDemo Launched: A member has released a quantum-native yet algorithmic entropy engine for public testing.
- The member suggests itâs a self-promo but important to share given its potential impact on AGI.
- LLMs React to Randomness Quality: A member suggests that LLM outputs are highly sensitive to the quality of the randomness used, distinguishing between true entropy and PRNG.
- They hypothesize that high quality entropy unlocks different, and often better, behaviors in models, linking to several Xitter posts in support.
- Randomness Vital for AGI: A member believes that randomness quality will be very important for AGI.
- They will continue to do tests to validate this hypothesis.
Nous Research AI â· #research-papers (1 messages):
ifeq: I gotta learn mandarin
Yannick Kilcher â· #general (35 messagesđ„):
Grok's apprehension of reality, Cloudflare serving fake content to agents, Third party filters for LLM output, Personal access to university resources via AI, KL Divergence Minimization
- Grok may get nerfed by Right-Wing Propaganda: A member speculated that Grok might be nerfed in its apprehension of reality to favor right-wing propaganda, linking to an attached image.
- They added that the real problem is all problems today already existed, and that AI or no AI we would still have them.
- Cloudflare serves fake content to AIs: A member thinks companies like Cloudflare are serving fake content to AI agents, similar to how some Chinese websites used zip bombs to deter cloning years ago, leading to biased AI responses.
- This comes after another member shared how ChatGPT wrongly answers about a video that is not the video they shared.
- Third-party filters needed for LLM output: A member suggested that we need third party filters for LLM output, including adblocking and fact/bias checking.
- In response, another member suggested that youâd need many models that ideally change often so they donât get corrupted such as 100 adblocker models and 100 fact checking tools.
- Personal AI access to university resources: A member expressed looking forward to a future where every human has personal access to the psychological, spiritual, intellectual and pragmatic resources of a major university via AI.
- Another member jokingly replied that they have already merged with ASI.
- KL Divergence minimization misses âpatternâ: A member suggests that many have started to use
---
and shares a link to a paper titled Beyond Icon - A Unified Formulation for Objectives / Regularizations.- They claim that the authors compared various formulations in table 1, but couldnât realize the pattern that many patterns are just f(x) and g(x) = sum_xâ f(xâ) or g(x) = f(x) ==> p(x) = f(x) / sum_xâ f(xâ).
Yannick Kilcher â· #paper-discussion (7 messages):
Paper Presentations, Causality, CVPR, Proper Investiture, Daily Paper Discussion
- Time Off Incoming: A member announced theyâre taking time off for the next two weeks, but will return, encouraging others to present or organize in their absence.
- New member introduction: A new member inquired about the breadth of topics discussed in the daily paper discussions, as theyâd like to present MSc papers related to causality and CVPR.
- They mentioned that they havenât had the chance to join the daily paper discussions yet.
- Proper Investiture: A member shared a link to a Springer article with the comment arguably the most proper investiture.
- Daily Paper Discussion: A member announced tonightâs daily paper discussion at
<t:1746750600:t>
will be about this arXiv paper.
Yannick Kilcher â· #ml-news (14 messagesđ„):
Zed compilation on Windows, Biological brains vs backpropagation, LLM beats Factorio == ASI?
- Zed compiles on Windows, Github Sign-in Required: A member successfully compiled Zed on Windows, following instructions here, but noted blurry fonts and the need to sign in with GitHub for tab completion.
- Another member expressed disappointment, wanting to try Mellum 4B on LM Studio for tab completion.
- Backprop is all you need?: A member stated that biological brains donât have backpropagation; theyâre non-epochal, spiking, recurrent, analog networks, and cited this Tweet as evidence.
- Factorio ASI Benchmark Proposal: A member jokingly proposed that if an LLM managed to beat the game Factorio without making a mess, we could go ahead and declare that ASI.
- They linked a YouTube video showing Factorio gameplay.
Eleuther â· #general (41 messagesđ„):
Cursor Advertising, Slurm Memory Requests, Job Posting Channel, Linguistics Channel, Cursor primary IDE Correlation
- Discord debates Cursor Advertising Rule: Members debated whether posts about Cursor constitute advertising and violate the no-advertising rule, given its popularity, perceived utility, and the fact that it isnât entirely free, noting âits just even toleratable bc we (group) think of cursor as useful right now but it still biases decisionsâ.
- Some users suggested that vague rules applied arbitrarily, along with interpreting âno advertisingâ as âno spamâ, and requiring payment for job postings could filter out low-quality offers.
- User Stumbles on Slurm Memory Misconfiguration: A user discovered they were requesting 80MB of memory through Slurm, not 80GB, calling it a âslurm momentâ, while another user celebrated their bare-metal setup.
- The initial issue was described as âvery stupidâ by the user who discovered the misconfiguration.
- Chatter about the Job Postings on Discord: Discussion arose around creating a jobs channel, with concerns that it could be overrun by low-quality postings offering âexperienceâ as compensation, with one suggesting payment to post as a potential solution.
- Others argued against a jobs channel, suggesting it would make the server another place for recruitment and proposing EleutherAI shouldnât charge for differential access to the Discord server.
- Linguistics Channel Gains Traction: A user proposed a channel for classical linguistics and its theory, focusing on pre-2000s knowledge such as sentence formation and meaning creation âon the flyâ, with the intent to add discussions that are not common within the NLP space.
- It was described as âcool stuff that rarely gets discussed in the NLP world for âsomeâ reason (probably because itâs irrelevant to the work nowadays).â.
- Coding community discusses the downfalls of Cursor as a primary IDE: Members expressed that the AI code tooling, such as Cursor, may not be as great as traditional methods such as tmux with vim and Claude code on the side.
- One member observed âan extremely strong correlation for incompetence with Cursor specifically as their primary IDE.â
Eleuther â· #research (7 messages):
MTurk vs. Prolific, RWKV's token shift
- Prolific prevails over MTurk for human evals: Members recommend Prolific over MTurk for human evaluations, citing its higher quality data and more reliable participant pool.
- The consensus is that Prolific is the superior choice in approximately 80% of cases.
- RWKV Token Shift Speculations: A member inquired whether Profilicis token shift from rwkv7 and causal_conv1d are the same.
- Another member clarified that token shift in RWKV is normalized such that the sum of all temporal weights for any channel equals 1, and referenced a paper.
Eleuther â· #interpretability-general (2 messages):
The Pizza and the Clock
- Craving more Clock-Pizza?: A member asked for more papers like The Pizza and the Clock.
- Another member responded with a suggestion that this has a lot of references but wasnât sure what they were looking for.
- N/A: N/A
- N/A
Eleuther â· #lm-thunderdome (3 messages):
LocalCompletionsAPI, loglikelihood tasks, bos token, HF model generation_config settings
- LocalCompletionsAPIâs loglikelihood runs!: A member is running loglikelihood (multiple-choice) tasks using base models and the
LocalCompletionsAPI
implementation.- They confirmed that itâs working great, but they can see that the tokenized prompt includes the bos token.
- BOS Token Blues!: The same member asked whether thereâs a way to specify
add_bos_token=False
when usingLocalCompletionsAPI
.- They want to control whether the beginning-of-sequence token is added to the prompt.
- HF Model generation_config: Default Temperature?: The member inquired if setting
do_sample:true
without specifyingtemperature
would default to the HF modelâs generation_config settings.- They clarified needing
temp > 0
, otherwise it setsdo_sample
to false.
- They clarified needing
Notebook LM â· #announcements (1 messages):
NotebookLM, Mobile App, Trusted Tester Program
- NotebookLM Launches Mobile App Trusted Tester Program: NotebookLM is launching a mobile app (beta version) đ± soon, and is looking for experienced web app users to participate in a trusted tester program to shape its future.
- Interested users can register by filling out this form, which includes reviewing and agreeing to the Trusted Tester Terms.
- Trusted Testers Needed to Beta Test NotebookLM Mobile App: NotebookLM seeks experienced web app users to become trusted testers for the beta version of their mobile app.
- Testers will gain early access in exchange for providing feedback and reporting bugs; registration requires agreeing to the Trusted Tester Terms.
Notebook LM â· #use-cases (9 messagesđ„):
NotebookLM PDF Processing, NotebookLM Knowledge Base for Sales, Audio length limitations
- NotebookLM PDF Processing Limit Set by Experiments: Users report NotebookLM doesnât work well with large PDFs or numbers of PDFs; one user tested by asking questions further into the PDF and found issues after 200 pages.
- The user suggests running a quick experiment to test current limitations.
- NotebookLM Responds Based Only on Source Material: In the chat interface, NotebookLM generates responses solely based on the uploaded source materials.
- If a questionâs answer doesnât exist in the imported materials, the AI will state that no information related to the question exists in the sources.
- Knowledge Base for Sales Content Creation for NotebookLM: A user is building a knowledge base for sales content in NotebookLM, using primary client decks and sales enablement materials within the 300 document limit.
- They plan to give access to the internal sales team, seeking guidance, examples, and understanding of limitations, especially regarding sharing and potential silos.
- Prolonging Audio Generation via More Content: A user asks how to make audio longer, aiming for a minimum of 12 minutes or more.
- Another user suggests providing more input to give the system more content to work with.
Notebook LM â· #general (22 messagesđ„):
NotebookLM failing to answer questions, Video Uploads, Audio Overview Functionality, Podcast Length, AI 'Humanic' Behavior
- NotebookLM Troubles: System Refuses to Answer: Users report that NotebookLM is responding with âThe system was unable to answerâ, even when asked to summarize the default notebook, with issues also arising when generating mind maps and study guides.
- Some users are seeking solutions and confirming whether others are facing the same issue.
- Video Uploads Supported (but with Limitations): Users confirmed that NotebookLM supports video uploads in formats like mp4 and avi, contrary to some inaccurate information on Googleâs official site, as per the Google Support Page.
- It analyzes the audio part of the video file and provides a transcript and summary, but mov format isnât supported.
- Audio Overview Access Elusive: A user inquired about accessing the audio overview feature to interact with it but couldnât find the option.
- Podcast Length Varies by Language: A user seeking tips to extend podcast length noted that changing the language to English allowed for generating significantly longer audio summaries (up to 49 minutes), whereas other languages were limited to around 14 minutes.
- A team member stated that this is expected behavior and they are working on enabling longer audio summaries in other languages soon.
- Discomfort with Artificial âHumanicâ AI: A user expressed discomfort with the âhumanicâ behavior of the AI in deep dives, specifically mentioning unnatural âuhmâsâ, and inquired about removing this behavior.
Latent Space â· #ai-general-chat (22 messagesđ„):
X-Ware, Netflix Recommendation Model, Gemini Image Generation, aider postmortems, Suno Music
- X Marks the Spot for Content: Members shared links from X (formerly Twitter), including a general link and specific posts from users like thegautam, TheAhmadOsman, and openaidevs.
- The shared content seemed to be of general interest to the channel, eliciting brief acknowledgements.
- Netflix Personalizes Recs with Foundation Model: A member highlighted that Netflix developed a foundation model for personalized recommendations as noted in the comments of one of the shared links.
- This was pointed out in relation to other discussions on recommendation systems.
- Gemini Generates Buzz with New Images: Members shared a link showcasing new Gemini image generation.
- A member mentioned that this team will be presenting at the aie worldâs fair recsys x llms track.
- Aider Autopsies: More Thorough Than Self Notes?: Members noted how aider postmortems are very thorough, especially regarding Gemini cost analysis.
- Sunoâs Sonic Styles: Yodeling Blues Concerts: A member raved about Sunoâs ability to mix styles, particularly highlighting a successful attempt at creating a Yodel + Blues + Live concert mix.
- They shared an audio file as evidence of Sunoâs impressive output.
Latent Space â· #ai-announcements (2 messages):
Claude code pod, AI Engineer conference, Early Bird Tickets, AI Engineer conference speakers
- New Claude Code Pod Bursts onto Scene: The Latent Space podcast promoted a new Claude code pod.
- Listeners are excited about the potential new episodes and insights from the collaboration.
- AI Engineer Conference Early Bird Tix Vanish Soon: The AI Engineer conferences, slated for June, alerted community members that Early Bird tickets are expected to sell out by the weekend.
- Attendees are encouraged to secure their tickets promptly to take advantage of the discounted rate.
- AI Engineer Conference Speakers Revealed: The AI Engineer conferences unveiled the lineup of speakers for the June event.
- Enthusiasts are eager to see the expertise and insights the speakers will bring to the conference.
Modular (Mojo đ„) â· #general (15 messagesđ„):
Fields in traits vs properties, Modular Hackathon at AGI House, Hardware Agnostic ML Systems Survey Paper, Zotero and bibtex for citations
- Properties Trump Fields in Mojo Traits: Discussion emerged around the possibility of having fields in traits in Mojo, but it was argued that properties in traits is a strictly better, more general idea.
- It was noted that fields in traits could happen, but one would be denied the ability to add such a trait via an extension; it would need to be included in the original struct definition.
- Modular Hackathon Hypes Hillsborough: A final reminder was shared about the Modular Hackathon at AGI House this Saturday, and there are a few spots remaining to attend, sign up here.
- Talks will be given by Modular team members as well as Mark Saroufim (GPU MODE & PyTorch), Simon Boehm and Sasha Krassovsky (Anthropic), and Dylan Patel (SemiAnalysis).
- Hardware Agnostic ML Survey Surfaces: A member completed their survey paper on modular and the Hardware Lottery piece, using it for their final presentation to help tell a good story to their peers.
- The latest version of the paper should always be available here and they welcome feedback.
- Zotero Zaps Citation Struggles: During a discussion about citations, it was recommended that using Zotero + bibtex makes most issues go away.
- A member shared their pain that natbib gave me about 70 errors with almost nothing linking until i caught a single unescaped â%â.
Modular (Mojo đ„) â· #mojo (4 messages):
Mojo roadmap, GPU programming puzzles, Colab Integration, New requires keyword
- Mojo Roadmap unveiled!: Modular posted the near-term Mojo roadmap on the forum, see the official post.
- The roadmap details whatâs coming soon for the Mojo language.
- GPU Puzzles tease Mojo programmers: The new GPU programming puzzles look interesting, and some members are wondering if itâs possible to run them on Colab for those without GPUs.
- One member said The new GPU programming puzzles look really cool.
- Colab gets Mojo (hacky): A member posted a kinda-hacky implementation of a Colab notebook that can run Mojo code on the GPUs of the free and Pro tiers on Colab: Colab notebook.
- The member admits Thereâs a way better experience to be had here with a little work noting that it builds the cell as a single Mojo file, you have to look to the logs for compile errors, etc.
- âRequiresâ Keyword Requests Raining!: The Discord channel reacted very positively to the new requires keyword for adding constraints to structs and functions with lots of Mojos.
DSPy â· #general (13 messagesđ„):
Collab and partnership, ReAct module signature, DSPy Caching Mechanism, RL experiment with GRPO on a Qwen 1.7B
- New partnership is on the Horizon?: A member asked if the project is open for a collab and partnership that both boost their communities.
- The member inquired about starting a chat to discuss potential synergies.
- ReAct Module Needs No Output: A member asked about creating a signature for a ReAct module that only makes tool calls and doesnât require other outputs.
- Another member suggested using success: bool as the output to indicate when the task is complete.
- DSPy Cache: A Multi-Layered Mystery: A member discovered that DSPy has its own caching mechanism in addition to any caching by the LLM provider, which can lead to unexpected results when credentials expire.
- Multiple layers of caching, including DSPyâs cache (github.com/stanfordnlp/dspy/blob/main/dspy/clients/cache.py), LiteLLMâs cache (docs.litellm.ai/docs/proxy/caching), and Bedrockâs cache, can make debugging difficult.
- GRPO Gets Going, Recall Retreats: A member ran a small RL experiment with GRPO on a Qwen 1.7B using DSPy to optimize query rewriting for retrieval, seeing baseline recall drop from 28% to 26% after training.
- More details are available in a Twitter thread noting that the drop was likely due to sparse rewards, short runs, and BM25 mismatches with CoT rewrites.
Cohere â· #đŹ-general (7 messages):
Cohere Embedding Model, Cohere Rerank Model, Cohere Embed 4
- Embedding Model struggles with negotiation: A user noticed that the embedding model does not work well with negotiation, and the score between the userâs query âI can payâ and embed data for âNo, I cannot payâ returns 0.92 which is too similar.
- A member suggested trying the rerank model for things like this instead of just vector similarity.
- Cohere Embed 4 token level embeddings: A member asked if they can get token level embeddings using Cohere Embed 4.
- Another member responded that one could embed one token at a time, but they would not advise it.
Cohere â· #đĄ-projects (1 messages):
AI Cost Tracking, Multi-Platform AI Service Management, AI Expense Justification, AI Tool Frustrations
- AIBillingDashboard tracks costs across AI platforms: A solo founder and software engineer created AIBillingDashboard.com, a platform that helps users track and optimize their AI service costs across multiple providers like Cohere, OpenAI, Anthropic, Azure AI, and Google Vertex.
- The platform consolidates costs, helps allocate expenses, provides usage analytics, and enables budget tracking across all services.
- Track expenses and see Optimization Opportunities: The creator found it difficult to track and analyze costs across multiple AI services, leading to the creation of a unified dashboard.
- The dashboard solves the problem of manually pulling reports from different dashboards, struggling to allocate costs to specific projects, and lacking a unified view of total AI spend.
- Seeking pain points in AI Cost/Usage Tracking: The founder is seeking feedback on the AI cost and usage tracking problems that users are facing.
- Examples of pain points include difficulty comparing price/performance, challenges forecasting costs, and struggles justifying AI expenses to management.
Cohere â· #đ€-introductions (3 messages):
Collaborations, Introductions
- Members seek Collaborations for Competitive Profit: A member is looking for someone to collaborate with, promising competitive profit to European and American collaborators.
- Interested parties are encouraged to DM for more detailed discussions about collaboration.
- Introductions are encouraged with a Template: New members are welcomed and encouraged to introduce themselves, indicating the community is excited to have them.
- A template is provided, requesting information such as Company/Industry/University, current projects, favorite tech/tools, and desired gains from the community.
Cohere â· #đą-status-updates (1 messages):
Embedding Models Degraded, embed-english-v2.0, embed-english-v3.0
- Embedding Models Encounter Hiccups: Cohere reported degraded performance affecting embed-english-v2.0 and embed-english-v3.0 models, with an investigation underway.
- Further details are available on the Cohere Status Page with the update timestamped May 08, 2025, at 07:25AM.
- Cohere Investigates Embedding Model Performance Issues: Cohere is actively investigating a live incident causing degraded performance in specific embedding models.
- The affected components include embed-english-v2.0 and embed-english-v3.0, as indicated in a status update.
Cohere â· #đŻ-private-deployments (1 messages):
GPU Requirements, On-Premise Deployment of Command A
- GPU Specs Inquiry for Command A: A user is seeking information regarding the specific GPU requirements for an on-premise installation of Command A.
- Decoding Command Aâs GPU Needs: The user aims to understand the necessary GPU specifications to successfully deploy and run Command A within their own infrastructure.
Torchtune â· #general (5 messages):
Tokenizer Automation, HuggingFaceBaseTokenizer Limitations, Custom Autotokenizer, ModelTokenizer Wrapper
- Tokenizer Automation Goal Discussed: A member is looking to automate tokenizer identification across model types for internal customers using
torchtune
.- The goal is to remove or automate the touchpoint of identifying the tokenizer, aiming for a more generic usage of
torchtune
.
- The goal is to remove or automate the touchpoint of identifying the tokenizer, aiming for a more generic usage of
HuggingFaceBaseTokenizer
has Limitations for SFT:HuggingFaceBaseTokenizer
lacks logic for templating/tokenizing messages, restricting its use to text completions training and not SFT.- The discussion highlights that this tokenizer cannot be used for Supervised Fine-Tuning (SFT) due to the absence of message templating capabilities.
- Custom Autotokenizer Suggested: A suggestion was made to write a custom âautotokenizerâ for internal customers, setting it as the default in the config.
- This autotokenizer could use if statements or more clever methods to define the model name at the top of the config for the tokenizer and checkpointer.
ModelTokenizer
Wrapper Planned to Close HF Gap: There is a known gap in torchtune, with plans to provide aModelTokenizer
that wrapsHuggingFaceBaseTokenizer
to map HFâsapply_chat_template
to torchtuneâstokenize_messages
.- This enhancement is expected to greatly assist users in onboarding new models and an issue will be opened on the repo to sketch out the implementation details, inviting community contribution.
Torchtune â· #dev (8 messagesđ„):
Cosine Scheduler with Warmup, Pytorch NaN bug with compiled Adam, Torchtune's get_cosine_schedule_with_warmup function, Torchtitan LR Scheduler Implementation, LR Warmup scaling
- Cosine Shenanigans: Warmup and Learning Rate Schedules: Discussion arose around implementing a cosine scheduler with warmup and dealing with a PyTorch bug that causes NaN weights when using a compiled non-fused Adam/AdamW optimizer with a learning rate scheduler that sets the learning rate to exactly 0 at any point during training.
- The bug occurs when
get_cosine_schedule_with_warmup
sets the learning rate to 0 at the first step, conflicting with the initial bugfix that enabled the use of LR schedulers with optimizer compile, but one member pointed to the Torchtitan implementation as a potential solution.
- The bug occurs when
- Adamâs Apple: Compiled Optimizers and Zero Learning Rates cause NaN weights: A Pytorch bug was reported where NaN weights resulted from using a compiled non-fused Adam/AdamW optimizer in conjunction with a learning rate scheduler that at some point sets the learning rate to exactly 0.
- One member noted that Torchtuneâs
get_cosine_schedule_with_warmup
always sets the learning rate to 0 at the first step, triggering the issue when optimizer compile is enabled.
- One member noted that Torchtuneâs
- Titanâs Approach: LR Warmup at the start of training: It was mentioned the Torchtitan implementation sets the LR ratio to
1/(warmup_steps+1)
on the first step, but unlesslr_min
is set, the last step will still be 0.- One member said The torchtitan approach works too as itâs reasonable to just skip the 0th step.
- Warming Up to LR Warmup Scaling: A discussion about LR scaling strategy: for the warmup steps, instead of
0,1/n, 2/n, ..., n-1/n
you wantmin_lr + (1/n ) * (1 - min_lr), min_lr + (2/n ) * (1 - min_lr), ..., min_lr + (n-1/n ) * (1 - min_lr)
.- For the cosine you want to scale the progress by the inverse of the cosine schedule, so
progress *= arccos(2*min_lr-1)/(pi*2.0*num_cycles)
will result in your max progress computed so thatcosine_lr_multiple == min_lr
.
- For the cosine you want to scale the progress by the inverse of the cosine schedule, so
LlamaIndex â· #blog (3 messages):
Anthropic API web search tool, LlamaParse improvements, VoyageAI multi-modal embeddings and MongoDB indexes
- Anthropic APIâs Search Tool is Born: The Anthropic API now supports a built-in web search tool with day 0 support in LlamaIndex, according to this Tweet.
- LlamaParse adds Gemini and GPT4 Support: LlamaParse is improving with new features like GPT 4.1 and Gemini 2.5 Pro models, plus auto orientation, skew detection and confidence scores for parsing quality according to this tweet.
- Multi-Modal Retrieval Voyage with MongoDB: Learn how to do multi-modal retrieval using VoyageAIâs multi-modal embeddings and MongoDBâs multi-modal indexes in this notebook.
LlamaIndex â· #general (4 messages):
Medical LLM Bot, Fine-tuning vdr-2b-multi-v1 with math formulas, Writer's Palmyra X5 and X4 in Bedrock
- Medical LLM Bot Workflow Advice Requested: A user is building a medical LLM bot and seeking guidance on implementing a workflow that includes iteratively suggesting follow-up questions based on previous answers from a local LLM.
- They are seeking advice on whether LlamaIndex has tools to help build this kind of workflow.
- Fine-tuning vdr-2b-multi-v1 for Math Formulas: A user inquired about fine-tuning the vdr-2b-multi-v1 model using the llamaindex/vdr-multilingual-train dataset to better handle complex math formulas in documents.
- They noted that formulas are not present in the training data and are seeking resources, steps, or tutorials for fine-tuning in this context.
- Palmyra Models error in LlamaIndex Bedrock: A user reported encountering an error, âProvider writer for model us.writer.palmyra-x5-v1:0 is not supportedâ, while using Writerâs Palmyra X5 and X4 foundation models in Amazon Bedrock within LlamaIndex.
- They note that the models are available in Amazon Bedrock.
tinygrad (George Hotz) â· #general (4 messages):
tinygrad CUDA, tinygrad IR, tinygrad docs, tinygrad uops
- Exploring tinygradâs CUDA Integration: A user inquired about how tinygrad integrates CUDA support generally.
- They also asked whether tinygrad has its own Intermediate Representation (IR).
- tinygrad Documentation Dive: A user shared a link to the official tinygrad documentation.
- They also shared links to notes on tinygrad uops and more tinygrad notes.
tinygrad (George Hotz) â· #learn-tinygrad (3 messages):
CACHEDB environment variable
- CACHEDB env var location spotted: A member asked about the CACHEDB environment variable.
- Another member pointed to line 175 in helpers where it is mentioned.
- CACHEDBâs Purpose Clarified: Following up on the initial query, the CACHEDB environment variableâs function wasnât explicitly stated.
- Further discussion would be required to understand the variableâs practical application and context within the project.
LLM Agents (Berkeley MOOC) â· #hackathon-announcements (1 messages):
AgentX Workshop, Lambda Inference API, Agentic AI
- Lambda Hosts AgentX Workshop: Lambda is hosting the AgentX Workshop: Building Agentic AI with Lambda on 5/15 10am PT for AgentX competition participants and AI enthusiasts looking to scale their projects using Lambdaâs powerful Inference API.
- Participants will learn to build practical agentic applications, optimize agent performance, and deploy agents in production environments, including a live demo.
- AgentX Prizes Announced: Special prizes are available for AgentX Competition participants, with up to $1,000 in credits for 1st place, $500 for 2nd, and $300 for 3rd in both Entrepreneurship and Research tracks.
- Interested participants can register now to get the YouTube livestream link.
LLM Agents (Berkeley MOOC) â· #mooc-questions (4 messages):
HF Credits, Course Content, MOOC Iterations
- Users Await Hugging Face Credits: Two users reported issues with tracking Hugging Face credits, with one not receiving emails and the other awaiting approval.
- The first user mentioned it was challenging to visit the website each day.
- Course Lectures Confirmed and Clarified: A prospective student asked about the course content, specifically whether the guest lectures listed on the course website were comprehensive.
- The staff clarified that the listed lectures are indeed comprehensive and also confirmed that the Spring MOOC includes more advanced topics like code generation and theorem proving, whereas the Fall version includes more applications topics.
- Another Iteration on the Way?: A user inquired about future iterations of the course, specifically if there would be another offering in the Fall.
- Staff replied that Prof Song is hosting another Berkeley class on Agentic AI this fall, but it is unknown whether it will be a MOOC version.
LLM Agents (Berkeley MOOC) â· #mooc-lecture-discussion (2 messages):
AI Engineer Courses, LLM Agents MOOC
- LLM Agents MOOC Recommended for Aspiring AI Engineers: A member inquired about the best complete course to become an AI Engineer.
- Another member recommended starting with the Fall 2024 LLM Agents MOOC.
- AI Engineer Career Path: A user asked about resources for becoming an AI Engineer.
- The LLM Agents MOOC was suggested as a solid starting point.
Codeium (Windsurf) â· #announcements (1 messages):
JetBrains Plugin Updates, Windsurf Editor UX Improvements, Wave 8 Release
- Windsurf Wave 8 Brings UX and Plugin Boosts: Windsurfâs final Wave 8 release introduces enhancements to the JetBrains plugin and improvements to the Windsurf Editor user experience, detailed in a blog post and changelog.
- JetBrains Plugin Cascade Adds Memory and Rules: The updated JetBrains plugin now supports Memories for persistent information between sessions, Rules via
.windsurfrules
files, and MCP (Model Context Protocol) server connections, outlined in the Jetbrains plugin changelog. - Windsurf Editor Gains UX Features: The Windsurf Editor UX sees improvements like a Continue button, redesigned model selector, workspace-to-conversation mapping for filtering history, enhanced code blocks and hunk navigation, editable terminal commands, and new file proposals in Chat mode, as showcased in todayâs launch video.