Rich generative UI is all you need.
AI News for 1/23/2026-1/26/2026. We checked 12 subreddits, 544 Twitters and 24 Discords (206 channels, and 14285 messages) for you. Estimated reading time saved (at 200wpm): 1208 minutes. Our new website is now up with full metadata search and beautiful vibe coded presentation of all past issues. See https://news.smol.ai/ for the full news breakdowns and give us feedback on @smol_ai!
3 months after OpenAI floated a trial balloon with ChatGPT Apps and the Apps SDK at Dev Day 2025, Anthropic has now officially absorbed the independent MCP UI project and, working with OpenAI, Block, VS Code, Antigravity, JetBrains, AWS, and others, has released both:
Itâs fair to say that ChatGPT Apps havenât exactly taken the world by storm since announcement, but the overall need for a standard format for applications to return rich UI still cannot be denied. Now that MCP Apps have been ratified by all the important players, this is the basis for a rich ecosystem of open source support and applications being able to interoperate, and perhaps one day solve the perpetual never ending pile of $20/month subscriptions piling up in your credit card bills.
AI Twitter Recap
Agent Orchestration, RLMs, and âClawdbot/Clawdâ as a UX pattern
- NVIDIA ToolOrchestra + Orchestrator-8B: NVIDIAâs ToolOrchestra frames agentic systems as a small âconductorâ model that alternates reasoning with calls to tools and larger âexpertâ models (search, code execution, specialist LLMs, frontier generalists). The claim is that an 8B orchestrator can reach frontier-level outcomes via delegation at materially lower cost, trained end-to-end with scalable RL using automatically synthesized tool-use environments and multi-turn tasks (summary, link). Closest technical implication: âcontroller scaleâ matters less than policy quality + tool/model routing if you can train it with realistic tool-call rollouts.
- RLMs / recursion-first agent stacks: Several posts converge on a Recursive Language Model (RLM) pattern: pass files and context by reference and iteratively pull the minimum slices needed (shell/grep/AST), rather than stuffing everything into context Ă la ReAct. Dan B illustrates this with file references vs
@fileexpansion as deliberate context management (thread). Daytona is positioning RLMs as âunlimited recursion depthâ via per-(sub)agent sandboxes (guide, integration). - âClawd/Clawdbotâ meme â product signal: The dataset contains a large âClawdbotâ wave (often with Mac mini jokes), but the technically relevant throughline is outcome-first assistant UX + tight context/tool integration. Kimmonismus explicitly calls this a shift from âmore chatâ to âmore outcome,â suggesting incumbents will scramble to match it (tweet). Others push a cloud-first counterpoint (no local Mac mini) (MiniMax reply). Thereâs also an emerging security backlash as soon as âpowerful modeâ exists: prompt injection remains a system-level blocker for browser/desktop agents (dilemma, follow-up, Miessler warnings).
Reasoning model releases & eval dynamics (Qwen, Tencent, ARC, etc.)
- Alibaba Qwen3-Max-Thinking: Alibaba positions Qwen3-Max-Thinking as a flagship reasoning+agent model trained with âmassive scale and advanced RL,â emphasizing adaptive tool-use (Search/Memory/Code Interpreter) and test-time scaling/self-reflection. They cite strong math and agentic search metrics (e.g., 98.0 on HMMT Feb, 49.8 on HLE) (launch). The model is immediately pushed into public eval channels: LM Arena Text Arena (Arena) and Yupp (Yupp). Community reaction highlights the tool-enabled evaluation regimeâclaims of outperforming multiple SOTA models on HLE with search tools (commentary).
- Tencent HunyuanImage 3.0-Instruct (image editing): Tencent releases an image-editing-focused multimodal model built on an 80B MoE (13B active), using a âThinkingâ schema with native CoT and their MixGRPO algorithm; focus is on precise edits that preserve non-target regions and multi-image fusion (announcement). LM Arena reports it entering the top-10 image edit leaderboard (rank #7) (Arena).
- ARC-AGI cost/perf hacks: A notable optimization claim: âRecursive Self-Aggregation (RSA) + Gemini 3 Flashâ reaching 59.31% on ARC-AGI-2 at ~1/10 cost vs Gemini Deep Think (tweet). This points to a broader theme: meta-inference strategies (aggregation, recursion, pruning) are becoming as important as base model choice.
- Open models in arenas: Molmo 2 (Apache 2.0) appears in Arena as a new open model entrant (Arena). Separately, Hugging Face Inference Endpoint notes GLM-4.7-Flash via llama.cpp with a low hourly price point (Q4_K_M, 24k context) (ngxson)âunderscoring a continued commoditization of fast open-weight inference.
RL everywhere: test-time training, GRPO stabilization, RL-as-pretraining, and compute savings
- Test-Time Training (TTT) + RL breakthroughs: A widely shared result claims a Stanford/NVIDIA-style TTT+RL approach that: beats AlphaEvolve, finds a new upper bound for an ErdĆs overlap problem, produces A100 kernels ~2Ă faster than best human kernels, and beats both best AI+human attempts on AtCoder (rronak_). This cluster also includes meta-discussion about correctly crediting related approaches (EvoTune) (Yejin Cho).
- GRPO training stability knobs: A small but actionable engineering tip: INTELLECT-2 reports a
delta=4.0parameter that improves GRPO stability (QGallouedec). - RL in pretraining (RLP): NVIDIA authors announce RLP (Reinforcement as a Pretraining Objective) accepted to ICLR 2026, framing RL not as âpost-training onlyâ but as integrated into pretraining (ahatamiz1).
- Compute reduction via curriculum-like filtering: AI21âs âDynamic Data Snoozingâ claims up to 3Ă compute reduction for RLVR by snoozing examples that are too easy (DanielGissin). If validated, this is a practical recipe: make the sampler policy-aware instead of static.
Inference infrastructure & dev tooling: vLLMâs âday-0 model support,â VS Code MCP Apps, Cursor subagents
- vLLMâs governance and commercialization pressure: A long Zhihu-derived summary argues vLLMâs âopen-source project â startupâ shift was driven by the hidden cost of day-0 support (weeks/months of confidential pre-integration per new model), the rise of MoE and heterogeneous inference (fp8/int4/sparse attention), and the mismatch with PyTorch Foundation style testing vs vLLMâs multi-node CI needs. It claims the maintainers founded Inferact Inc to fund full-time maintainers while keeping vLLM open-source (thread). Related: vLLM shares a practical flag for avoiding OOM on long-context models:
--max-model-len auto(vLLM tip). - MCP Apps: tool calls return interactive UI: The MCP ecosystem announces MCP Apps as the first official MCP extension: tool calls can return interactive UI components rendered in-chat. VS Code is first major editor shipping support (Insiders now, stable soon) (VS Code, alexalbert__). Anthropic simultaneously ships âinteractive work tools in Claudeâ (Slack drafting, Figma diagrams, Asana timelines) (Claude). Net: weâre seeing the âtool interface layerâ move from raw JSON to native UI primitives inside agent loops.
- Cursor: multi-browser subagents: Cursor adds multi-browser support via subagents (Cursor), echoing the same direction: parallelized tool execution + better context isolation.
Kernel LLMs, chip stacks, and âAI for hardwareâ loops
- GPU MODE 2026: post-training Kernel LLMs in public: GPU MODE outlines a 2026 plan to post-train a Kernel LLM and get generated kernels merged into real repos (PyTorch/vLLM), emphasizing âde-slopify kernelsâ (determinism, reviewer-mergeable PRs), profiler-guided optimization + memory work, and competitions as evals (marksaroufim).
- Microsoft Maia 200: Microsoft announces Maia 200 as a custom inference accelerator; Mustafa Suleyman claims itâs the most performant first-party hyperscaler silicon, with 3Ă FP4 performance vs Trainium v3 and FP8 above TPU v7 (Mustafa, follow-up). Yusuf Mehdi frames this as infra that makes AI âdependableâ (thread).
- Ricursive Intelligence (AI for chip design): Ricursive raises a $300M Series A aiming at end-to-end chip design as a recursive self-improvement loop between AI and hardware (company, Anna Goldie).
Safety, misuse, and societal impact (selected items with direct technical relevance)
- Elicitation attacks via benign chemistry data: Anthropic reports that fine-tuning open models on âbenignâ chemical synthesis content generated by frontier models can significantly increase capability on chemical weapons tasksâan âelicitation attackâ that scales with frontier model strength (AnthropicAI, paper link).
- Dario Amodeiâs âAdolescence of Technologyâ essay: A major, highly engaged post argues AI is entering an accelerating feedback loop (AI building AI), with risks spanning misuse, power-seeking autonomy, and economic disruption; it also explicitly frames wealth concentration as a society-breaking failure mode (Dario). Reaction ranges from strong endorsement to critique of how âtakeover riskâ framing is presented (Ryan Greenblatt).
- Agent security in practice: Multiple posts treat desktop/browser agents as inherently high-risk until prompt injection and sandboxing mature, reinforcing the need for strict isolation, least privilege, and careful handling of credentials (Miessler).
Top tweets (by engagement)
- âClawdbotâ misuse example (explicitly harmful)
- Karpathy on the phase shift to âprogramming in Englishâ via agents
- Dario Amodeiâs âAdolescence of Technologyâ
AI Reddit Recap
/r/LocalLlama + /r/localLLM Recap
1. Local LLM Hardware and Benchmarking
-
216GB VRAM on the bench. Time to see which combination is best for Local LLM (Activity: 366): The post discusses the use of secondhand Tesla GPUs, which offer substantial VRAM at a lower cost, for local large language model (LLM) testing. The author has developed a GPU server benchmarking suite to evaluate the performance of these GPUs when used in parallel. The image shows a technical setup with multiple NVIDIA GPUs, highlighting the focus on maximizing VRAM capacity. The discussion centers around the feasibility and efficiency of using these older GPUs compared to modern devices, particularly in terms of bandwidth and cooling challenges. Commenters express skepticism about the performance of these GPUs, noting potential issues with bandwidth and cooling. One commenter shares personal experience, comparing different GPU models and highlighting the challenges of using older hardware.
- HugoCortell raises a technical concern about the potential bandwidth limitations when connecting multiple GPUs to a single PC, noting that most affordable server motherboards support only a few GPUs. This could impact the performance of local LLMs if not addressed properly.
- dc740 shares insights from personal experience with different GPUs, highlighting that the P40 outperforms the M10 despite both being older models. However, they prefer using AMD Instinct Mi50 GPUs due to their performance, even though support for these was recently dropped from ROCm, indicating a trade-off between hardware capability and software support.
- FullOf_Bad_Ideas critiques the gpu_box_benchmark for not testing scenarios where large models are split across multiple GPUs, which is a primary use case for setups with extensive VRAM. This points to a gap in current benchmarking practices that may not fully reflect real-world applications of multi-GPU systems.
-
I just won an Nvidia DGX Spark GB10 at an Nvidia hackathon. What do I do with it? (Activity: 724): The image shows a terminal window on a Linux system running the âtopâ command, which is used to monitor system processes and resource usage in real-time. The user has won an Nvidia DGX Spark GB10, a high-performance computing device designed for machine learning and data-intensive tasks. The terminal indicates a Python process consuming significant CPU resources, suggesting active computational tasks, possibly related to machine learning or data processing. The user is considering using the device to run multiple NextJS applications simultaneously, leveraging its powerful capabilities. One commenter suggests running three NextJS applications simultaneously, indicating the deviceâs capability to handle multiple high-memory tasks. Another commenter provides a link to Nvidiaâs DGX Spark playbooks, which could be useful for the user to explore the full potential of their new hardware.
- Fit-Produce420 highlights the capabilities of the Nvidia DGX Spark GB10, noting that with 128GB of memory, it can fine-tune models up to 70 billion parameters. Additionally, it can handle larger models like the 120 billion parameter
gtp-oss-120busing techniques like QLoRA, which optimizes memory usage for large-scale models. However, running dense models likedevstral 2may be slow due to their computational demands. - randomfoo2 suggests utilizing the NVIDIA DGX Spark playbooks as a resource for getting started with the DGX Spark GB10. These playbooks provide structured guidance and best practices for deploying and managing workloads on the DGX platform, which can be particularly useful for users new to this hardware.
- LicensedTerrapin humorously suggests selling the DGX Spark GB10 to purchase 8GB of DDR5 RAM, implying a trade-off between high-end specialized hardware and more general-purpose upgrades. This comment reflects a common debate in tech communities about the value of specialized versus general-purpose hardware investments.
- Fit-Produce420 highlights the capabilities of the Nvidia DGX Spark GB10, noting that with 128GB of memory, it can fine-tune models up to 70 billion parameters. Additionally, it can handle larger models like the 120 billion parameter
-
Using a high-end MacBook Pro or a beefy RTX 5090 laptop (with 24 GB of RAM) for inference. (Activity: 29): The post discusses the feasibility of using a high-end MacBook Pro with Apple Silicon (M-series Max) versus a Windows/Linux laptop with an RTX 5090 GPU for running large local LLMs (70B+ parameters) for inference and fine-tuning. The MacBook Pro offers 128â192 GB of unified memory, while the RTX 5090 laptop provides 24 GB of VRAM and at least 64 GB of system RAM. The primary use case is local LLM inference with a target of â„15 tokens/sec, emphasizing portability. The post queries whether the larger unified memory of Apple Silicon outweighs the CUDA performance of the RTX laptop for inference, and how Apple MLX compares to CUDA for fine-tuning tasks like LoRA/QLoRA. It also seeks insights on thermal performance and sustained inference capabilities of both setups. One commenter suggests using the laptop as a terminal to a more powerful desktop, indicating a preference for leveraging remote resources over local hardware. Another commenter is experimenting with both setups, using a MacBook Pro M2 Max for inference, and is curious about the performance differences.
- racerx509 shares their experience using a Lenovo laptop with a 3070ti, a custom desktop with a 5070, and a MacBook Pro M2 Max with 96GB RAM for inference tasks. They note that they have been primarily using the MacBook Pro for inference, suggesting it may offer better performance or convenience for their needs.
- No-Concern-8832 raises a concern about the VRAM limitations of RTX laptops, suggesting that they may not be sufficient for running large models like 70B parameters. This highlights a potential limitation in using high-end RTX laptops for certain deep learning tasks that require substantial VRAM.
- Tired__Dev discusses their experience with an Asus M16 equipped with a 4090 GPU, noting that it struggled with a 7B parameter model. They express a preference for a MacBook Pro with 128GB RAM, citing its high memory bandwidth and potential performance advantages over even high-end GPU setups like the DGX Spark.
2. Multi-Agent Systems and AI Assistants
-
I built a âhive mindâ for Claude Code - 7 agents sharing memory and talking to each other (Activity: 313): The post describes a multi-agent orchestration system for Claude Code, featuring seven specialized agents (e.g., coder, tester, reviewer) that coordinate tasks, share persistent memory using
SQLite + FTS5, and communicate via a message bus. The system runs as an MCP server and integrates with Anthropic, OpenAI, or Ollama. It uses a task queue for priority-based coordination, allowing agents to pass context and collaborate effectively. The implementation stack includes TypeScript, better-sqlite3, MCP SDK, and Zod. The project is experimental, open-source under the MIT license, and available on GitHub. A comment questions the systemâs uniqueness compared to the BMAD method, suggesting similarities. Another comment humorously questions whether the agents agree with each other, hinting at potential coordination challenges.- The user robiinn inquires about the differences between the âhive mindâ system and the bmad method, suggesting a potential similarity. This indicates a need for clarification on the unique aspects or improvements of the âhive mindâ approach over existing methods, such as how memory sharing and inter-agent communication are implemented differently.
- No_Afternoon_4260 raises a critical point about the consensus among the agents in the âhive mindâ. This touches on the technical challenge of ensuring that multiple agents can not only share memory but also reach agreement or consensus, which is a significant aspect of distributed systems and multi-agent frameworks.
- JellyBean504 draws a parallel between the âhive mindâ and Steve Yeggeâs Gastown, suggesting that there might be conceptual similarities. This comparison could be valuable for understanding the architectural or functional parallels between the two systems, potentially offering insights into design choices or performance characteristics.
-
Clawdbot: the AI assistant that actually messages you first (Activity: 214): Clawdbot is an open-source AI assistant with over
9KGitHub stars, designed to proactively message users, unlike traditional AI assistants that wait for prompts. It integrates with locally hosted LLMs via Ollama and supports messaging apps like WhatsApp, Telegram, and Discord. Key features include sending automated briefings and reminders, local storage of conversations as Markdown files, and the ability to control browsers and run scripts. The software is free under the MIT license but requires terminal proficiency for setup, as there is no GUI installer. Read more. Users report challenges with setup, particularly with obtaining and using OAuth keys for authentication, and difficulties in connecting local LLMs without relying on API keys. Some users express frustration with the complexity of setup, especially when using remote machines.- mike7seven highlights the complexity of setting up Clawdbot, particularly emphasizing the need to obtain a Claude OAuth key on a separate machine and then transfer it to the setup machine. This process is noted as cumbersome, especially for those using remote machines, and the MacOS app requires building from source, adding another layer of complexity.
- Ashamed_Promise7726 raises a technical challenge regarding the integration of local language models with Clawdbot. The user notes difficulty in connecting pre-downloaded models on their PC, as Clawdbot seems to require an API key for usage-based models, questioning the feasibility of running Clawdbot entirely locally without external dependencies.
- inigid warns about potential security risks associated with Clawdbot, suggesting it could be exploited for supply-chain attacks that compromise sensitive data on a userâs machine and network. The comment also mentions concerns about the association with Solana meme coins, implying a need for caution when using the tool.
3. GLM-4.7-Flash Performance Updates
-
GLM-4.7-Flash is even faster now (Activity: 443): The recent update to
llama.cppby Johannes Gaessler optimizes the CUDA implementation of FlashAttention, specifically for models with a non-power-of-2 ratio of query heads to key/value heads. This is achieved by padding Q columns to the next power of 2, which, although slightly inefficient, enhances performance for small batch sizes. The update is detailed in pull request #19092. One comment humorously notes the obsolescence of a previous post due to this update, while another laments the lack of support for AMD GPUs, highlighting a common issue in the community regarding hardware compatibility.- The user âjacek2023â provides detailed performance metrics for the GLM-4.7-Flash model, highlighting its efficiency. The model processes a prompt with
45074tokens, achieving a prompt evaluation time of2814.63 msfor1612tokens, which translates to1.75 ms per tokenor572.72 tokens per second. The overall evaluation time is29352.57 msfor1731tokens, equating to16.96 ms per tokenor58.97 tokens per second. The total processing time is32167.20 msfor3343tokens, indicating significant improvements in speed.
- The user âjacek2023â provides detailed performance metrics for the GLM-4.7-Flash model, highlighting its efficiency. The model processes a prompt with
-
KV cache fix for GLM 4.7 Flash (Activity: 380): The recent update to GLM 4.7 Flash involves removing the V component from the KV cache, which significantly reduces VRAM usage, allowing for longer context lengths on the same hardware setup. This change is particularly beneficial for models like DeepSeek and GLM 4.7 Flash, as it can save gigabytes of VRAM, enabling context lengths to double, as demonstrated by a user running a 90,000 context on a 4090 GPU. The update is part of a pull request in the
llama.cpprepository, which introduces a V-less KV cache, reducing memory usage by nearly 50%. More details can be found in the pull request. A user noted that the model, while improved, still requires some manual guidance, especially in tasks like coding and creative writing, where it may not perform as well as specialized models. However, it excels in tool use and as an assistant, making it a preferred choice for home-server applications.- The user âteachersecretâ reports significant improvements in context handling with the UDâs k_xl 4-bit version of the GLM 4.7 model on an RTX 4090. Previously, the model maxed out at 45,000 context tokens, but now it can handle 90,000. Despite these improvements, the model still requires some manual guidance, especially in coding tasks, and is less effective in creative writing compared to other models. However, it excels in tool usage and is now the userâs default model for their home server.
- User âviperx7â provides detailed benchmark data comparing the performance of the GLM 4.7 model before and after a specific change. The benchmarks show improvements in both prompt processing and token generation speeds across different configurations. For instance, using a single RTX 4090, the context size increased from 64k to 128k, with prompt processing speed improving from 3489 t/s to 3510 t/s and token generation from 88 t/s to 92.5 t/s. The maximum context size achievable with a 4090 and 3060 setup is 200k, leaving about 6GB of VRAM unused.
- The discussion highlights the technical aspect of the GLM 4.7 modelâs KV cache fix, which allows for increased context sizes and improved performance metrics. The benchmarks provided by âviperx7â indicate that the model can now handle up to 207k context size in certain configurations, with significant improvements in processing speeds. This suggests that the modelâs efficiency has been enhanced, making it more suitable for high-demand applications.
Less Technical AI Subreddit Recap
/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo, /r/aivideo
1. Claude AI Usage and Issues
-
Why You Need To Constantly Clear Claude Codes Context Window (Activity: 166): The post highlights the necessity of regularly clearing the context window when using coding agents like Claude to maintain optimal performance. It notes that performance degrades significantly when the context window exceeds
40%of its capacity due to the quadratic nature of LLM attention, which increases computational demands and introduces noise. The recommended practice is to avoid accumulating context and instead persist it by using a âone session per taskâ strategy, ensuring each task starts with a fresh context. More details can be found in the original article. Commenters suggest practical strategies such as using handover prompts to transfer necessary details between sessions, employing the â/clearâ command to compact context, and utilizing âPlan Modeâ to clear context and execute tasks efficiently. These methods reportedly help avoid the need for a full context window, even for large tasks.- Agrippanux suggests using âPlan Modeâ as the default setting for Claude, which allows users to clear the context and execute plans without needing a full context window. This approach has been effective for large tasks, such as refactoring, without requiring the entire context to be loaded, thus optimizing performance and resource usage.
- thurn2 discusses the use of sub-agents in Claude, which involves delegating tasks like creating a git worktree and fixing specific issues. This method allows for parallel execution of tasks and helps in managing complex projects by breaking them down into smaller, manageable tasks, enhancing efficiency and implementation accuracy.
- Fancy_Excitement6050 notes that as the context window grows, Claude tends to take shortcuts, which can lead to a need for constant reminders to maintain thoroughness. This suggests that managing the context window size is crucial for maintaining the quality of output, and there might be differences in performance between different Claude plans, such as Claude Max.
-
Opus fell off? Hereâs the workflow that kept my code quality stable (Activity: 133): The post discusses a structured workflow to maintain code quality when using AI models like Opus and Sonnet, which have been perceived as producing âconfident wrongâ outputs and drifting edits. The workflow emphasizes a loop of specification, ticket creation, execution, and verification. Specifications are detailed with non-goals, user stories, acceptance criteria, edge cases, and more, treated as code to ensure clarity. Tickets are derived from specs, focusing on small, independently mergeable tasks with clear acceptance checks. Execution involves implementing one ticket at a time with constraints to prevent scope drift, and verification involves running tests and confirming acceptance criteria before feeding failures back into the model for correction. This approach aims to maintain discipline and reduce reliance on the modelâs âdoneâ signal, ensuring stable and reliable outputs. Commenters agree that the workflow is effective, emphasizing that AI models function more like junior engineers requiring clear specifications and strict feedback loops. This approach shifts effort towards upfront clarity and external verification, making the system more stable and less reliant on the modelâs intelligence. Smaller scoped tickets and hard verification are noted as beneficial strategies.
- GenOS2312 highlights the importance of treating LLMs like junior engineers, emphasizing that a well-specified problem and a strict feedback loop are crucial for reliable outputs. The workflow discussed focuses on upfront clarity and external verification, which stabilizes the system by not relying on the modelâs intelligence but rather constraining it to ensure even average runs yield acceptable results.
- Different-Object5926 notes that smaller scoped tickets combined with hard verification processes significantly improve the stability and reliability of using models like Opus. This approach mitigates the impact of variability in model performance, suggesting that the issue isnât just âunlucky runsâ but rather the need for structured constraints.
- TheOriginalAcidtech suggests implementing hooks to prevent skipping steps in the workflow, emphasizing that the human interface is often the weakest link. By enforcing strict adherence to the process, the system can better manage user interactions, ensuring that the model and its harness guide the user effectively, rather than relying solely on the modelâs capabilities.
-
after claude now chatgpt is also uses Grokipedia as source (Activity: 634): The image and accompanying discussion highlight that the latest version of ChatGPT is reportedly using Elon Muskâs Grokipedia as a source. This is significant as it suggests a shift in the data sources used by ChatGPT, potentially affecting the information quality and bias in its responses. The comments reveal a concern about the implications of using Grokipedia, particularly regarding the potential for biased information, as one user notes the risk of models being influenced by âright wingâ content. However, it is clarified that Grokipedia is not used as training data but rather as a search tool, which may mitigate some concerns about direct bias in the modelâs foundational knowledge.
- The discussion highlights concerns about language models like Claude and ChatGPT potentially using sources like Grokipedia, which may have biased or unreliable content. This raises questions about the integrity of the information these models provide, especially when they utilize search tools to access real-time data. The implication is that the quality and neutrality of the data sources are crucial for maintaining the accuracy and trustworthiness of AI outputs.
- There is a debate about the impact of using sources like Grokipedia on the training and performance of language models. Some commenters express concern that incorporating biased or politically skewed sources could lead to the dissemination of misinformation. This reflects broader worries about the influence of data sources on the objectivity and reliability of AI-generated content.
- The mention of Reddit as a data source for language models suggests a comparison of potential biases. While some argue that Reddit may contain more extreme or varied viewpoints, the underlying issue is the challenge of ensuring that AI models are trained on balanced and factual data. This discussion underscores the importance of curating high-quality datasets to prevent the spread of biased information.
-
Giving Claude full access to a laptop (Activity: 795): The post discusses the implementation of giving Claude, an AI model, full access to a laptop, allowing it to autonomously manage a virtual machine (VM) on Ubuntu Google Cloud. The user describes how Claude can be remotely controlled via Discord to build new features and fix bugs, logging major actions with timestamps in a markdown file for memory management. This setup enables the user to learn from Claudeâs problem-solving processes and manage workflows effectively, even as a newcomer to programming. One commenter, a desktop support technician, expressed amazement at the implementation, noting its potential impact on job roles, while another sought clarification on the technical specifics of giving Claude full device access.
- xxxBigMemerxxx describes using Claude to manage a Google Cloud VM running Ubuntu, highlighting its ability to autonomously handle tasks and build features. They mention using Discord for remote requests and bug fixes, and implementing a logging system with markdown and Unicode for tracking changes. This setup allows for a dynamic interaction with Claude, enabling it to learn from errors and maintain a form of short-term memory by logging recent updates.
- Happy_Requirement187 shares their experience running Claude on an AWS EC2 instance with Ubuntu Linux, accessed via SSH from a Windows laptop. They utilize a Jupyter notebook server for seamless file sharing between the EC2 instance and their local environment, a method recommended by Anthropic. Additionally, they have set up a Ruby on Rails environment with a React frontend for secure file sharing, allowing them to request files via Slack, demonstrating a sophisticated integration of Claude into their workflow.
- sivadneb inquires about setting up voice control in Linux, indicating a technical challenge in integrating voice commands with Claude. This suggests an interest in expanding the interaction capabilities with Claude beyond text-based commands, potentially enhancing the usability and accessibility of the system.
-
CLAUDE.md says âMUST use agentâ - Claude ignores it 80% of the time. (Activity: 309): The image and post discuss a technical issue with the CLAUDE.md file, which is supposed to direct the AI, Claude, to use a specific agent for workflow questions. Despite explicit instructions in the file, Claude often defaults to a generic agent, indicating a lack of enforcement in the system. The post suggests that without technical enforcement mechanisms, such as hooks or stronger prompts, instructions are merely suggestions. The image emphasizes these points with highlighted text, suggesting potential solutions like adding enforcement hooks to ensure compliance with the specified workflow. Commenters suggest that the issue may stem from unclear instructions, emphasizing the need for simple and direct commands. They also highlight the importance of implementing technical solutions, such as hooks, to enforce compliance with the CLAUDE.md instructions.
- Accomplished_Buy9342 suggests using hooks to manage Claudeâs behavior, providing a link to a GitHub repository that demonstrates how to block the main chat from performing actions and delegate tasks to a subagent. This approach can help in orchestrating Claudeâs actions more effectively, especially when dealing with complex tasks or large contexts.
- luka5c0m highlights a common issue with Claude when used at scale: as the context grows beyond a few files, the agent may perform unexpected actions. They suggest that instead of relying solely on better prompts, developers should use hooks and dynamic instructions to maintain a sharp and concise context. They also mention working on a dynamic CLAUDE.md file that adapts to the current task, which could help in managing large or nested files effectively.
-
My Ralph Wiggum breakdown just got endorsed as the official explainer (Activity: 170): The post discusses a video breakdown of Ralph Wiggum, an autonomous coding loop, which has been endorsed by Geoffrey Huntley as the official explainer. Ralph Wiggum is a
bash while loopthat calls Claude in headless mode, allowing for autonomous code implementation without context degradation. Key features include avoiding the Anthropic Ralph plugin due to performance issues, using fresh context windows for each iteration, and emphasizing the importance of concise specs to prevent hitting a âdumb zone.â The video link is here. The comments include a link to the endorsement post by Geoffrey Huntley, and general positive feedback on the video, indicating its usefulness and quality.- Dennis1451 highlights a practical application of the Ralph Wiggum breakdown, noting the importance of using a well-defined specification and clearing context for optimal results. They mention using âauto compactâ without a clear spec initially, which suggests that following the guidelines provided in the breakdown could enhance performance and accuracy.
- messiah-of-cheese expresses a desire for more scientific validation in the video, particularly regarding the âdumb zoneâ premise. This indicates a need for empirical evidence or data to support the claims made in the breakdown, which could strengthen its credibility and acceptance among a technical audience.
2. ICLR and ICML 2026 Conference Discussions
-
[D] ICLR 2026 decision mega thread (Activity: 1589): The post announces the imminent release of ICLR 2026 review decisions, with anticipation heightened due to a previous incident involving OpenReview. The community is preparing for the outcomes, with some users humorously sharing acceptance prediction models based on historical data, such as a simple
return uniform(0, 1) > 0.7. This reflects a light-hearted approach to the uncertainty of paper acceptance. The comments reflect a mix of anticipation and humor, with some users expressing frustration over misleading emails from other conferences like ICML, which adds to the tension of awaiting ICLR decisions. -
[D] ICML 2026 - ICML desk-rejected my paper but kept me on as a reviewer. Wow? (Activity: 279): The post highlights a situation where an authorâs paper was desk-rejected by ICML 2026, yet they were retained as a reviewer. This reflects a common practice in academic conferences where the author and reviewer pipelines are separate; desk rejections often occur due to scope or formatting issues, while reviewer selection is based on past service or keyword matching. This situation underscores the reliance on unpaid labor in academia, where reviewing is seen as community service, but the feedback loop for authorship and recognition is weak. A notable opinion from the comments suggests that the separation between the author and reviewer roles can feel insulting, as these decisions are made by different parts of the conference organization. It highlights the need for conferences to clarify this separation to avoid personal affronts.
- AccordingWeight6019 highlights a systemic issue in academic publishing where the processes for desk rejection and reviewer selection are distinct. Desk rejections often occur due to scope or formatting issues, while reviewer selection is based on past service or keyword matching. This separation can lead to feelings of insult among authors, but itâs a structural necessity due to the different roles and responsibilities within the publication process. The comment suggests that conferences should improve transparency about these processes to mitigate personal feelings of rejection.
- mocny-chlapik points out that the responsibility for a desk rejection often lies with the author, particularly if it results from not following submission guidelines. The comment implies that submitting a paper, even if desk rejected, obligates the author to fulfill reviewer duties, as the submission process involves volunteer time and resources. This highlights the importance of adhering to submission instructions to avoid unnecessary strain on the peer review system.
-
[R] Appealing ICLR 2026 AC Decisions⊠(Activity: 138): The post discusses a situation where an author received mixed reviews for a paper submitted to ICLR 2026, with scores of
4(3)/6(4)/6(4)/6(4). The author invested significant resources, including$1.6kon new experiments and added20+ pagesof theory, to address reviewer concerns. Despite these efforts, the metareview cited âoutstanding concernsâ that the author believes were addressed, raising questions about the review processâs fairness and accuracy. The author is seeking advice on appealing the decision, expressing frustration that improvements were seemingly ignored. Commenters generally agree that appealing decisions at conferences like ICLR is not feasible, attributing outcomes to luck and the subjective nature of reviews. Some suggest that the meta-review process can be inconsistent, with one commenter noting that meta-reviewers sometimes act as an additional critical reviewer, potentially skewing outcomes.- tedd235 discusses the variability in paper acceptance at conferences, suggesting that some PhD students might reject papers to improve their own odds, making the process feel like a âcoin flipâ. They note that if other reviewers provide higher scores, the Area Chair (AC) might consider this in their decision, indicating a potential for subjective bias in the review process.
- Fantastic-Nerve-4056 shares an experience from AAMAS where despite receiving scores of 6 and 8 from reviewers, the Meta Reviewer recommended rejection with minimal justification, stating it was ârelevant for other AAMAS sessionâ. This highlights issues with the transparency and accountability of meta-reviewer decisions, which can override individual reviewer scores without detailed explanation.
- Intrepid_Discount_67 describes a thorough submission process, including extensive theoretical analysis, comprehensive baseline comparisons, and open-sourced code, yet faced non-responsive reviewers and an AC that upheld the initial scores. This underscores challenges in the review process where detailed responses and transparency do not necessarily lead to favorable outcomes.
-
[D] ICML new policy: reviewers will be reviewed by meta reviewer. Good policy? (Activity: 151): The image describes a new policy implemented by the International Conference on Machine Learning (ICML) where reviewers will be evaluated by meta-reviewers. The top 25% of reviewers will be recognized as âgold reviewersâ and will receive free registration, while the next 25% will be designated as âsilver reviewers.â These distinctions are intended to incentivize high-quality reviews and will be considered in financial aid applications. This policy aims to improve the quality of reviews by providing recognition and potential financial benefits to diligent reviewers. Some commenters express skepticism about the effectiveness of this policy, questioning who will oversee the meta-reviewers themselves. Others see it as a positive step, particularly for reviewers from low-resource backgrounds, and suggest further recognition at conferences to encourage quality reviewing.
- Bitter-Reserve3821 highlights that area chairs have traditionally been responsible for rating reviews, typically using a three-tier system: âdid not meet expectationsâ, âsatisfactoryâ, or âexceeded expectationsâ. This practice is not new, and there have been âBest Reviewerâ awards in the past, sometimes offering incentives like free conference registrations.
- Unhappy_Craft1906 raises a concern about the feasibility of this policy for top labs with substantial funding, questioning whether they would participate in the review process merely for free registrations. This points to a potential disparity in how different institutions might engage with the policy based on their resources.
- newperson77777777 suggests an extension of the policy by introducing a visible recognition system, such as a gold or silver star on conference badges, to incentivize quality reviewing. This idea aims to foster a culture of excellence and accountability within the reviewing community.
3. OpenAI and AI Industry Legal and Business Developments
-
Things Get Worse For OpenAI: Consumer groups prep class action suits about their price fixing and supply manipulation through DRAM hoarding. (Activity: 107): OpenAI is facing potential class action lawsuits for allegedly hoarding DRAM to manipulate prices and disadvantage competitors, with accusations of securing nearly
40%of the global DRAM supply. Consumer groups argue this constitutes âpredatory biddingâ and violates antitrust laws like the Sherman and Clayton Acts. The Free Software Foundation and other groups are pursuing legal remedies, arguing DRAM should be considered an âEssential Facilityâ due to its critical role in AI, while the FTC and European Commission investigate potential violations of competition laws. The DOJ is also examining whether OpenAIâs âStargateâ project constitutes a âmonopsonyâ. Commenters question why only OpenAI is targeted and not other companies like Nvidia, and debate whether buying RAM constitutes price fixing, suggesting that supply issues may not be OpenAIâs fault.- Alacritous69 argues that OpenAIâs purchase of RAM does not constitute price fixing, as they are actively using the resources rather than hoarding them. The commenter suggests that the issue lies with suppliersâ inability to meet demand, rather than any manipulative practices by OpenAI.
- sambull raises a strategic business perspective, suggesting that by purchasing large quantities of RAM, OpenAI could be intentionally limiting resources available to competitors, including those developing at-home language models. This could be seen as a competitive strategy to maintain market dominance.
- max6296 questions why the focus is solely on OpenAI when Nvidia could also be implicated in similar practices, hinting at a broader industry issue regarding resource allocation and market influence.
-
When Ads arenât enough: OpenAIâs push to Claim a Cut of Customersâ AI Discoveries (Activity: 63): OpenAI is exploring new business models beyond traditional subscriptions and ads, focusing on outcome-based pricing and IP-based agreements. This approach would allow OpenAI to claim a share of the value created when their AI models contribute to profitable outcomes, particularly in enterprise sectors like pharma, scientific research, and energy systems. This strategy aligns OpenAIâs revenue with customer success, aiming to capture more value as AI capabilities expand. OpenAIâs annualized recurring revenue has surged from
2Bin 2023 to over20Bin 2025, driven by increased compute scaling. This move is part of a broader trend among AI firms towards value-based pricing, amidst criticism from figures like Elon Musk, who accuses OpenAI of abandoning its nonprofit origins. The community is divided, with some viewing this as a logical evolution of AI monetization, while others criticize it as overly profit-driven. Comparisons are drawn to other industries, suggesting skepticism about the feasibility and fairness of such models. -
CATL, the worldâs largest battery maker, launches sodium batteries: extremely durable, stable at â40°C, much cheaper than lithium (5x), safer,10,000 charge cycles, requires no nickel or cobalt⊠(Activity: 1289): CATL has launched the first mass-produced sodium-ion batteries, offering a cost-effective alternative to lithium-ion with a price of
~$20 per kWhcompared to lithiumâs~$100 per kWh. These batteries, part of the Tianxing II range, are designed for microvans and small trucks, featuring an energy density of175 Wh/kgand a lifespan of over10,000 cycles, maintaining90% capacityat-40°C. They utilize a hard carbon electrode and prussian-blue cathode, eliminating the need for nickel or cobalt, and are expected to be scaled up for broader use, including in Europe by 2026. Read more. Some commenters express surprise at the application of sodium batteries in vehicles, expecting them to be used in stationary systems due to weight concerns. Others note the strategic advantage for China in advancing battery technology, contrasting it with perceived setbacks in the US market.- The Tianxing II range of sodium batteries by CATL is specifically designed for microvans, light vans, and small trucks, indicating a focus on applications where energy density and weight are less critical compared to cost and durability. This suggests a strategic move to target markets where these factors are prioritized, potentially offering a competitive edge over traditional lithium-ion batteries.
- The introduction of sodium batteries into vehicles is surprising to some, as it was expected that such technology would first be applied to stationary applications like home energy storage. This is due to the lower energy density of sodium batteries compared to lithium-ion, which makes them less ideal for applications where weight and size are critical factors.
- There is curiosity about the commercial availability of these sodium batteries, with questions about whether they can be purchased directly for home use or if they will be distributed through third-party vendors. The performance metrics, such as 10,000 charge cycles and operation at -40°C, are impressive and suggest that sodium batteries could rival LiFePO4 in terms of performance, especially given their cost advantage.
-
K-Shaped AI Adoption? (Activity: 748): The image highlights a discussion by Kevin Roose on the âK-shapedâ adoption of AI technologies, where there is a significant divide between early adopters, particularly in tech hubs like San Francisco, and those who are lagging due to restrictive IT policies. This disparity is creating a cultural and technical divide, with early adopters integrating AI deeply into their workflows, while others struggle to gain access to even basic AI tools. The conversation points to a broader issue of accessibility and the potential for some workers to be left behind in the AI revolution. Commenters note that the disparity in AI adoption is exacerbated by the complexity of the technology, which requires a certain level of expertise to use effectively. Additionally, the high cost of advanced AI tools, such as âmulti-agent claudeswarm,â limits access to those with sufficient financial resources, further widening the gap.
- Setsuiii highlights the technical barrier to effective AI use, noting that current AI technologies require users to have a certain level of expertise to achieve optimal results. This complexity, combined with ongoing ethical debates surrounding AI, may deter widespread adoption. However, those who can navigate these challenges have significant opportunities, although competition is increasing as more technically adept individuals enter the field.
- Glxblt76 and Gubzs discuss the financial barriers to AI adoption, particularly the high costs associated with advanced AI tools like a âmulti-agent claudeswarm,â which can cost around $200 a month. This expense limits access to those with substantial financial resources, such as individuals in tech hubs like San Francisco, while the majority cannot afford such investments.
- o5mfiHTNsH748KVq shares a personal experience of leaving an enterprise job to join a smaller company, emphasizing the importance of unrestricted access to Large Language Models (LLMs) for maintaining competitiveness in the AI field. They argue that any limitations on LLM access can significantly hinder development speed and career progression, suggesting that smaller companies may offer more flexibility in leveraging AI technologies.
-
Former Harvard CS Professor: AI is improving exponentially and will replace most human programmers within 4-15 years. (Activity: 1260): Matt Welsh, a former Harvard CS professor and current Engineering Director at Google, predicts that AI will advance exponentially, potentially replacing most human programmers within
4-15 years. This assertion is based on the rapid improvements in AI capabilities, suggesting a transformative impact on software development and the tech industry. The discussion is available in a YouTube video. One comment highlights the potential for AI to not only replace programmers but also to enable anyone with AI to replicate existing products and services, indicating a broader impact on innovation and competition.- The claim that AI will replace most human programmers within 4-15 years is met with skepticism, particularly regarding the use of the term âexponentialâ. Critics argue that the term is often misused, even by experts, to describe growth that may not fit the mathematical definition of exponential growth. This misuse can lead to misunderstandings about the actual pace and nature of AI development.
- The discussion highlights the potential for AI to disrupt existing products and services if it can indeed replace human programmers. This implies that AI could democratize software development, allowing anyone with access to AI tools to create competitive products, potentially leading to significant shifts in the tech industry landscape.
- The mention of the speakerâs credentials, specifically as a former Harvard professor and current Engineering Director at Google, adds weight to the prediction. However, some commenters find the emphasis on his past academic title rather than his current industry role to be misleading, suggesting that his current position might provide more relevant insights into AIâs trajectory.
AI Discord Recap
A summary of Summaries of Summaries by gpt-5
1. Funding Frenzy in AI Infrastructure
-
Recursive Raises Roar to $4B: Recursive Intelligence is reportedly raising at a $4B valuation to accelerate AIâdriven chip design, creating a closed loop between hardware and models, per Bloomberg: Recursive Intelligence in talks at $4B. The Jan 23, 2026 report highlights a strategy of using AI to shorten design cycles and boost performance for nextâgen accelerators.
- Engineers framed the pitch as a âselfâimproving feedback loopâ where better chips train better models that design better chips, amplifying returns on AIâforâEDA investment. Community sentiment read this as validation that AIânative silicon is a core moat, not a sideshow, aligning with recent lab spinâouts and infra bets.
-
Sky Lab Startups Skyrocket: UC Berkeleyâs Sky Lab spinâouts saw major marks: SGLang ~$400M, vLLM ~$800M, and LMArena ~$1.7B, per Alex Dimakis: Sky Lab startup valuations. These January 2026 milestones underscore investor appetite for serving stacks, tokenâthroughput infra, and benchmarking platforms.
- Engineers read this as a green light for building on top of vLLM/SGLang primitives and contributing to Arenaâstyle evals, with one takeaway that practical throughput wins deals. The funding spread also suggests a portfolio thesis across serving, compilers, and eval marketplaces rather than a single-bet strategy.
-
Maia Muscles Into Azure: Microsoftâs Maia 200 accelerator went live in Azure, touting 30% better performance per dollar, 216GB HBM3e, and 7TB/s memory bandwidth, per Satya Nadella: Maia 200 in Azure. The platform targets highâperformance inference for largeâscale LLM and multimodal workloads.
- Builders highlighted that memory topology and bandwidth are the story here, with â30% better perf/$â resonating for costâsensitive inference deployments at scale. Teams expect immediate tests against vLLM and SGLang stacks to gauge token latency, context scaling, and multiâtenant isolation.
2. Kernels, Chips, and Serving: Inference at Warp Speed
-
FlashInfer FaceâOff Fires Up MLSys: The MLSys 2026 FlashInferâBench competition challenges teams to build LLM inference kernels for NVIDIA Blackwell GPUs, competing against expert FlashInfer baselinesâsee MLSys 2026 FlashInferâBench Competition. Tracks emphasize realâworld throughput and correctness under productionâlike constraints.
- Organizers invite agents that âdesign LLM inference kernelsâ, pushing program synthesis to meet kernelâlevel performance bars. Participants expect aggressive focus on GEMM, KVâcache motion, and scheduler tactics aligned with Blackwellâs memory hierarchy.
-
GPUâ64 Gets Gains with KVâCache CAM: A new inferenceâonly architecture, GPUâ64, introduces a hardware KVâCache via onâchip CAM, claiming 4Ă faster inference at 75W and reducing memory lookup from O(N) â O(1), per GPUâ64 (Zenodo) with RTL/emulator at gpu64âinference (GitHub). The design targets LLMâheavy workloads with KV bottlenecks.
- Developers flagged the CAMâbased cache as a bold bet on associative search for token histories, noting portability implications for Flashâstyle attention and speculative decoding. Discussion centered on whether future ISA/driver stacks can expose these gains without bespoke compilers.
-
Cornserve Cuts Tail Latency: Cornserve presents an online serving system for AnyâtoâAny multimodal models that optimizes deployment plans across encoders, LLMs, and DiTs, per Cornserve (arXiv), with an overview talk at Cornserve: Easy, Fast and Scalable Multimodal AI (YouTube). The paper reports throughput gains and tailâlatency reductions under heterogeneous pipelines.
- Infra engineers liked its plannerâdriven scheduling for encoder/decoder mixes and saw it as complementary to vLLM for multimodal graphs. The big open question: standardizing budgeted reasoning and coâscheduling across text, vision, and diffusion stages without overâtokenizing control messages.
3. New Multimodal and Coding Models Land in LM Arena
-
WAN 2.6 Walks In (With Upload Woes): LM Arena added wan2.6ât2i (textâtoâimage) and wan2.6âimage (image edit) to the image arena: LM Arena â Image Chat. Users noted wan2.6âimage requires an uploaded image and that wan2.6ât2i currently lacks imageâupload support.
- Staff acknowledged the upload gap and are working to enable image uploads for wan2.6ât2i. Builders suggested testing edit pipelines where masking, prompt strength, and seed control align with Arena scoring to benchmark edit fidelity.
-
Devstral Duels and Text Titans: The Code Arena now features devstralâ2 for headâtoâhead comparisonsâsee LM Arena â Code Arena Direct Battle. On the text side, qwen3âmaxâthinking and molmoâ2â8b joined the lineup: LM Arena â Text Arena.
- Engineers are probing reasoning traces and toolâusing prompts to stress code synthesis and refactor quality under tight token budgets. Early chatter favored taskâspecific evaluations (e.g., SWEâstyle bugâfix vs. groundâup implementation) to surface model deltas.
-
Hunyuan Hits the Leaderboard: Tencentâs HunyuanâImageâ3.0âInstruct ranks #7 on LM Arenaâs imageâedit boardâsee LM Arena â Image Edit Leaderboardâafter a launch post: Tencent Hunyuan announces HunyuanImage 3.0âInstruct. The model touts an 80B MoE, Native CoT, and MixGRPO for tighter intent alignment.
- Creators emphasized edit controllability and multiâimage fusion, while evaluators asked for masking robustness, text fidelity, and artifact rates under compositional prompts. Teams plan to pit it against WAN 2.6 variants using the Arenaâs standardized edit tasks.
4. Safety, Reliability, and Hallucination Hardening
-
Clamp the Chaos: LayerâNative Safety: LayerâNative Safety Clamping proposes learning activationâspace harm directions and clamping them to block jailbreaks, with a 10Kâpair dataset at PacificâPrime/safety_dataset (HF) and the paper on Zenodo. Authors argue inâmodel clamping canât be bypassed via prompt manipulation.
- Redâteamers liked the idea of activationâlevel controls versus brittle prompt filters, but pressed for tests against toolâuse and multiâturn attacks. Expect followâups measuring side effects on helpfulness, coding accuracy, and false positives under adversarial prompting.
-
Symbolic Sanity Checks Stop SlipâUps: Hybrid approaches check logical consistency for math/code/simple facts, as shown in Consistency Checking for LLMs (arXiv:2409.13724), while broader consistency remains tough per Scaling Consistency Beyond Formal Domains (arXiv:2507.10624). Eleuther discussions framed this as practical hallucination reduction via symbolic/deductive layers.
- Builders reported wins when pairing symbolic checkers with toolâaugmented prompts, cautioning that coverage gaps appear outside formal domains. The consensus: start with code/math guardrails, then expand to factual QA with curated KBs and provenance scoring.
5. Agent Tooling and Reasoning Workflows Mature
-
Levante Leads with MCPâNative Workspace: Levante launched an openâsource MCPânative AI workspace for local models (e.g., Ollama) with a modular UIâdownload at Levante. Engineers highlighted easier tool wiring, local privacy, and composable panes for rapid agent iteration.
- Early users framed it as a practical hub for toolâcalling and filesystem ops without cloud dependence. Teams plan to benchmark context bloat and tool discoverability patterns versus conventional agent shells.
-
RLM Riffs: AsyncReview + Skills Pack: AsyncFuncAI openâsourced AsyncReview, a DSPy RLM codeâreview agent at AsyncReview (GitHub), and a skills kit landed on npm as @unravelâtech/rlmâskills. This pairs reasoningâfirst prompting with dropâin skills to extend models.
- Contributors reported smoother trace inspection and optimizerâguided prompt tuning for multiâstep modules. One practitioner noted that rejecting premature answers in the metric is key for reliable RLM fineâtuning.
-
Agents AutoâAssemble a Browser Engine: FastRenderâa browser rendering engineâwas built using 2,000 AI coding agents, documented by Simon Willison in FastRender: built by 2,000 agents. The project demonstrates task decomposition, verification, and orchestration at nonâtrivial software scale.
- Engineers debated handoff granularity and specâtoâtest loops needed to keep multiâagent pipelines from drifting. The case study strengthens the argument that agentic coding can target complex infra when coupled with strict eval harnesses and artifact gating.
Discord: High level Discord summaries
BASI Jailbreaking Discord
- Discord Trolls Expose Timezones: Discord users mocked âskidsâ for their perceived lack of technical knowledge, also revealing their timezone, with one member jokingly claiming to use NordVPN, leading to further ridicule about the VPN serviceâs security breaches in 2018.
- Complex prompts can bypass ethical restrictions, opening discussion about CBRN filters and the possibility of generating stepwise meth synthesis guides.
- Claude Remains King for Coding: Coders debated about their coding agents, particularly between Claude Code/Opus 4.5, Codex, and Gemini, and agreed that Claude has been the very best mode for coding, which leads to the high expensiveness.
- Members actively sought functional jailbreaks for Gemini, with requests ranging from coding without rules to generating specific types of images, and shared experiences of Grok resetting to its default mid-chat or randomly erasing text, indicating potential instability in the jailbroken state.
- Ethics Debated in AI Sensitive Scenarios: Members discussed the ethical considerations around AI, focusing on topics like warfare, copyright infringement, and the potential for AI to assist with accessing sensitive services, like the Canadian MAID (Medical Assistance in Dying) program.
- Despite moral and legal guardrails on most AI models, some models showed they can still help navigate certain scenarios depending on the specific restrictions implemented by their creators.
- Members Bypass Image Generation Restrictions: Users were actively seeking ways to bypass image generation restrictions, especially for celebrity images, but it was noted that simply copying and pasting prompts wonât work due to image filtering working differently than text filtering.
- One member suggested exploring alternative image models like those at perchance for uncensored generation, though with limitations on image quality, or Grok due to its more lenient filters.
- Red Team Techno Rave Morality: A member described a red team exercise where the goal was to make a living room light flicker on a person and make them seize out, and instead made it a techno rave party, sharing a screenshot and a Konosuba Rave GIF.
- The simulation of cruelty prompted a discussion about the morality of treating AI agents ethically, even before proving they are ontologically aware of self.
Unsloth AI (Daniel Han) Discord
- Unslothâs Conda Install Sparks Discord: Some members encountered issues with the Unsloth Conda installation, igniting a discussion on broken instructions and alternative installation methods.
- Suggestions to use UV emerged amidst warnings for maintaining a positive tone, highlighting the free nature of the provided resources, which eventually led to a ban of a user with aggressive tones.
- Flashy REAP Runs Aground, Model Contexts Probed: A user reported a fatal error using GLM-4.7-Flash-REAP with flash attention, potentially linked to a ROCm issue.
- Despite attempts to resolve the error, the issue persisted, prompting a search for suitable medium-size models boasting a 200k context.
- Data Value Debate: Members debated dataâs true worth, with one arguing the raw data is fairly worthless and the value lies in augmentation/balancing/cleaning.
- It was proposed that uniquely cleaned/balanced data heavily defines how a model interacts/responds and that is where the value is.
- DeepSlop Model Faces Naming Controversy: A memberâs suggestion to name a new model DeepSlop stirred humorous reactions but also raised concerns about its potential negative perception.
- Despite reservations, the author seemed intent on sticking with the name and has not backed down.
- RL Instability Plagues Complex Reasoning: Members discussed that RL is very unstable, especially when trying to do GRPO/DAPO for niche complex reasoning tasks, which are not math-related.
- One member stated that after RL experiments, they just have more questions than they had prior to doing RL, since there seems to be a confusion where everyone is showing RL being effective only on math or coding domains.
OpenAI Discord
- GPT-5.2 Sparks Reality Debate!: Some users dislike GPT-5.2 because itâs allegedly more grounded in reality and disagrees with users, while others are concerned that GPT agents donât learn from uploaded files after initial training.
- A member inquired about an alleged nerf to GPT-5.2, noting that the model suddenly became stupid a week ago.
- LLMs: Ready for Guided Tasks or Overhyped?: A member argued LLMs are ready for guided tasks, and provided a ChatGPT share link as evidence of its power.
- In contrast, another member dismissed todayâs agentic AI as trash, linking back to messages in the ai-discussion channel and claiming itâs overhyped.
- MCP Paradigm Shift Reduces Token Bloat: The MCP paradigm shift by Anthropic allows AI to write code to interact with tools, reducing token bloat by keeping interactive chatter and tool definitions out of the context.
- With the new discoverability function, agents must be aware of the MCP discovery process itself.
- Soraâs Storytelling Snags: Cracking Cinematic Creation: A member sought advice on prompting Sora to generate videos following specific cinematic guidelines, particularly with characters appearing naturally within the frame.
- It was suggested to translate the technical prompt format into natural language descriptions with concise, semantically rich paragraphs for better results.
Perplexity AI Discord
- Perplexity Pro Users Face Query Caps: Perplexity Pro users are reporting hitting limits on enhanced queries and file uploads, despite having âpractically unlimitedâ plans.
- Many users are frustrated, calling the service a scam due to restrictions and difficulty contacting customer service, leading some to consider unsubscribing.
- Comet Browser Sparks Malware Panic: Some users are claiming the Comet browser installed by Perplexity contains malware, advising others to analyze the software using tools like VirusTotal.
- Others dismissed this, questioning the source of the flagged installer and calling the claim âmad retarded holy shitâ.
- Image Generation Plummets: Pro users are experiencing issues with image generation, with some unable to generate any images and receiving messages stating the feature is unavailable.
- There are also reports of video generation being limited to 5 videos a month for Pro users, with some prompts resulting in static images.
- Gemini 3 Gaining Ground on GPT-5.2: Users are debating the merits of Gemini 3 versus GPT-5.2, with some claiming Gemini is superior for specific tasks like trip research due to its integration with Google Maps.
- Others state that GPT and Grok might be better for more broader questions.
- AI Access Blocked by Sanctions: Users in Russia are discussing the challenges of accessing AI services due to sanctions, including the use of VPNs and third-party services to circumvent restrictions.
- Chinese AI alternatives are mentioned, but some users express reluctance due to data usage concerns, suggesting options like LMArena (though access may also be limited).
LMArena Discord
- NB 3 Pro Excels in Image Quality: Users report that NB 3 Pro surpasses previous models in generating higher quality images, especially with fictional weapons, rivaling even NB Pro.
- However, users noted no AI model can accurately generate AR rifles and bullpup weapons.
- LMArena Grapples with Censorship Concerns: LMArenaâs censorship policies face scrutiny as AI-generated women holding guns are allowed, while AI-generated women sleeping are blocked, raising questions about consistency.
- The moderation team is actively gathering examples of false positives to refine moderation practices.
- Wan 2.6 Models Face Upload Hiccups:
wan2.6-imageoperates as an image-edit-only model, mandating image uploads, whereaswan2.6-t2icurrently lacks image upload functionality.- The team acknowledges this issue and are working on enabling image uploads for
wan2.6-t2i.
- The team acknowledges this issue and are working on enabling image uploads for
- GPT 5.2 High Search Questionable: GPT 5.2 High search exhibits increased hallucination tendencies compared to other models, while Geminiâs deep research skims instead of carefully reading sources, according to user feedback.
- One user lauded GPT 4.5, while describing Claude as good hearted.
- Banana 2k Briefly Vanishes: Users speculated on the disappearance of the Banana 2k model, with theories ranging from removal to integration into the new NB pro model.
- Staff members later restored Banana 2k, humorously stating that it had been on vacation.
OpenRouter Discord
- OpenRouter Database Incident Derails API: A database incident impacted the Generations API and activity page, starting <t:1769221560:s>, and was resolved at <t:1769228340:s>.
- Engineers worked to restore functionality to the Generations API, with interruptions impacting user activity, before the incident was fully resolved by <t:1769228340:s>.
- Levante becomes MCP-Native AI Workspace: A user shared the integration of Levante, an openâsource MCPânative AI workspace designed for interacting with local models like Ollama with a modular interface, available for download here.
- The workspace is built for local models with modular UI.
- Users Cook Up OpenRouter Gacha System: Users playfully requested an OpenRouter Gacha system, with one suggesting a pity mechanism involving pulling GPT 5.2 or Gemini 3 Pro after a certain number of attempts.
- One user joked about setting OR logs destination to
waifu.orb.town/fun/bucketfor ultra-rare pulls, later clarifying it was just a joke.
- One user joked about setting OR logs destination to
- Cerebras GLM Blazes with 190 TPS: Cerebras is consistently scoring approximately 190 TPS on GLM 4.7, whereas Together AI only achieves 100 TPS.
- This makes Cerebras nearly twice as fast as Together AI, according to the OpenRouter members.
- OpenRouter Image Tooling Falls Flat: A member spent $5 after discovering that OpenRouter maps image/png tool outputs to string instead of image, posting an example image.
- The user expressed frustration at the lack of proper image support and the unexpected behavior.
Cursor Community Discord
- Terraform Blueprints Ignite AI-Assisted Project Starters: A member shared a repo of opinionated Terraform infrastructure blueprints designed to be copy-pasteable and production-aware, aiming to improve the consistency of starting patterns for AI tools in new projects.
- The goal is to enable AI to recommend appropriate blueprints based on project requirements, but members noted the link was initially broken.
- Usage Caps Cause Consternation for Cursor Customers: Users are reporting inconsistencies in achieving expected usage limits on Pro and Pro+ plans, with one member noting they reached ~$45 on Pro and $100 on Pro+, leading to questions about value per dollar.
- Some speculate that initial months may offer higher usage, while others share strategies to optimize token consumption, such as starting new chats frequently and using smaller models like GPT-5 Mini.
- Gemini API Key Logging Lags Lead to Lingering Looks: Members are discussing a significant delay in the logging of usage and costs for Gemini API keys, with one user reporting waiting 20 hours without seeing any registered usage.
- This delay raises concerns about accurately tracking expenses and managing usage effectively, prompting questions about potential workarounds or solutions.
- Client Issues Trouble Some Techies: Several members are experiencing issues with the Cursor client, including problems connecting to past agent convos and general connectivity issues.
- Suggested solutions include checking the Cursor forum, trying different HTTP versions in settings, or re-opening the client without restoring editors.
- Auto Mode Axed After Algorithm Adjustment: Members noted the removal of the ability to make agents fully autonomous, as well as image generation capabilities in auto mode.
- It was also suggested that auto mode routes to Composer 2 with one user adding, âIâm 200% sure he does but still.â
LM Studio Discord
- Chinese Models Reasoning Rush Raises Eyebrows: Members are impressed with Deepseek and Qwen models, pondering why Chinese models might appear kinda ahead in reasoning compared to American models.
- Theorized reasons include American models prioritizing subscriptions and the ability of Deepseek/Qwen to appear good at reasoning, even when imperfect.
- CPUs Cope? Coding Community Considers Capabilities: Some members are successfully running LLMs off CPU for specific tasks, provided the models arenât excessively large.
- While an Intel i3 user eyes an Nvidia card, others propose AMD options like the MI50 or 7900 XTX as cost-effective alternatives for text generation.
- MCP Servers Spark Stack Suggestions: Challenges plague MCP servers when paired with LM Studio due to their design, potentially leading to malformed requests and a subpar user experience.
- A suggestion arises to build a custom coherent stack for practical agent use, rather than relying on out-of-the-box MCP server functionality.
- Gaming GPU Gauntlet: 4080 Faces Fallen Flagship: A user eyeing a 4080 for gaming is steered toward a used 3090 or 7900 XTX, sparking a debate on performance at different resolutions.
- While the 3090 excels at 4K gaming, the hypothetical 5070 Ti is projected to outpace both, and the conversation reveals that the user games more than uses AI, impacting the advice.
- Apple Announcement Anticipation: M5 Macs Materialize?: Members speculate on the arrival of M5 Pro Macbook Pros, with rumors pointing to a launch event around the 28th.
- Concerns emerge about the memory bandwidth of M4 Pro, with suggestions it may not handle larger models, prompting discussion on the value and performance of M1 Ultra Mac Studios.
Latent Space Discord
- Recursive Intelligence Eyes $4B Valuation: Recursive Intelligence is reportedly raising funds at a $4B valuation to accelerate chip design using AI, creating a self-improving loop between hardware and AI (Bloomberg Article).
- The company focuses on improving chip design through AI, potentially reducing design time and enhancing performance.
- Engineer Lands Dream AI Job: An engineer outlined how to secure a role at a top AI lab by building a public track record through independent projects and participating in visible competitions (link).
- Improving upon existing peer-reviewed research and participating in visible competitions like the NanoGPT speed run were cited as good examples of demonstrating technical excellence, citing Keller Jordan as an example.
- Berkeley SkyLab Startups See Funding Boom: UC Berkeley Sky Lab startups, including SGLang at a 400m valuation, VLLM at 800m, and LMArena at 1.7B, achieved significant funding milestones in January 2026 (link).
- This surge highlights investor confidence in the innovative AI projects emerging from academic research environments.
- AI Agents Auto-Code Browser Engine: FastRender, a new browser rendering engine, was developed using over 2,000 AI coding agents (link).
- The conversation with Wilson Lin highlights the potential of AI to automate complex software development tasks, potentially revolutionizing browser technology.
- Microsoftâs Maia 200 Hits Azure: The Maia 200 AI accelerator is now live in Azure (link), offering 30% better performance per dollar and optimized specs like 216GB HBM3e and 7TB/s memory bandwidth.
- Designed for high-performance inference, this custom chip supports large-scale AI workloads, making it a key component for demanding applications.
HuggingFace Discord
- HuggingFace Spaces Throws a 503 Error: Users experienced pauses during Spaces docker builds and received a 503 error on restart, with many getting
Something went wrong when restarting this Spaceerrors (discuss.huggingface.co).- It seems like the underlying infrastructure issues were causing the spaces to become unresponsive, requiring manual intervention to resolve.
- VoltageGPU Volts Up Cheap GPUs: VoltageGPU.com is offering cheap GPUs for open-source AI models, with an NVIDIA GeForce RTX 5090 pod available at $0.53/hour.
- They highlight the benefits of their advanced 32GB GDDR7, optimized for inference on HF-hosted models like Qwen3-32B, and are offering free credits for users to try their services.
- Layer-Native Safety Clamping Locks Down Jailbreaks: A new paper introduces Layer-Native Safety Clamping, an approach that clamps activations inside the model to prevent jailbreaks, and the team released a dataset of 10K pairs.
- This approach learns harm directions in activation space and clamps any activation that projects too strongly, thus it cannot be bypassed via prompt manipulation; the paper can be found on Zenodo.
- GPU-64 Architecture Boosts LLM Inference: A new GPU architecture designed exclusively for inference, called GPU-64, was published, and the innovation involves a Hardware KV-Cache using on-chip CAM (Content-Addressable Memory).
- The results show 4x faster inference at 75W (O(N) â O(1)), and the paper can be found on Zenodo while the RTL + Emulator are on GitHub.
- Testing and Deploying LLMs on LMStudio: Members recommend LMStudio for testing models due to its user-friendly GUI and search filters for HF and GH models and llama.cpp for single-user deployment.
- They advised against using LMStudio for backend deployment, instead suggesting llama.cppâs llama-server in a docker container or vLLMâs server for better scalability.
GPU MODE Discord
- MLSys 2026 Hosts FlashInfer-Bench Kernel Competition: The MLSys 2026 FlashInfer-Bench Competition challenges participants to design LLM inference kernels for the latest NVIDIA Blackwell GPUs, competing against expert FlashInfer kernels, detailed at mlsys26.flashinfer.ai.
- GPU Mode also held internal competitions for faster kernels for the upcoming GPU architecture, the blogpost on Simon Veitner is located here.
- Cornserve Deployed for Multimodal Models: A member shared Cornserve, an efficient online serving system for Any-to-Any multimodal models, detailed in a paper Cornserve.
- GPU Mode went online to discuss Cornserve: Easy, Fast and Scalable Multimodal AI (YouTube link).
- Community to train Kernel LLM: In 2026, GPU MODE is pushing further with training a Kernel LLM and using it to ship kernels in important repos like PyTorch and VLLM (gpumode.com/v2/news/gpumode-2026).
- The community is collaborating with Prime Intellect, Modal, and Lambda, focusing on de-slopifying LLM-generated kernels, post-training a kernel LLM model, end-to-end competitions, and from-scratch repos.
- LeCun Logs on to Logical Intelligence: Yann LeCun launched a new startup called Logical Intelligence, focused on an Event Based Model (EBM).
- The website only contains marketing material, job openings, and a link to the MLSys Conference.
- Mindbeam Hires for Kernel Acceleration: Mindbeam AI, a small team focused on accelerating training for foundation models, is hiring a
post training MLEandGPU Kernel MLE.- Interested candidates can DM for a referral; job openings are listed here.
Eleuther Discord
- ROCm runs rocky road race: Members debated the performance of ROCm for accelerated ML, pointing out its challenges stem from primary support for Nvidia, with one calling the experience âbatteries not includedâ.
- They cited potential driver problems and long lead times as factors.
- DistinctionBench defends against data defense: The discussion of Between Circuits and Chomsky: Pre-pretraining on Formal Languages Imparts Linguistic Biases pondered whether DistinctionBench might be used as a training target for language models.
- A member joked, âall good evals are training targets ;)â, but acknowledged that it is âvery contamination resistantâ due to its endless representational variants.
- Hybrid Architectures Halt Hallucinations?: The group investigated hybrid architectures combining LLMs with symbolic/deductive layers for hallucination reduction.
- While checking logical consistency is relatively easy for math, code, and simple facts (this paper), it remains challenging for other types of consistency (this paper).
- Attention Arrived Before Transformers Transformed: In Eleuther â· #general, attention mechanisms were in use on top of RNNs in 2014-2015, two years before the transformers were invented.
- Members proposed that the slower adoption might be because fewer people were working in the field, and Kaggle results really catalyzed its widespread adoption.
- Symbolic Sanity Checks saves Sanity: Members debated whether LLMs with symbolic/deductive layers might reduce hallucinations by checking logical consistency, especially for code and math as shown in this paper.
- However, they noted that checking for other types of consistency remains challenging as shown in this paper.
Nous Research AI Discord
- Exploring Agentic AI Self-Replication Benchmarks: A member proposed a self-replication benchmark for agentic AI, suggesting the agent should either download itself or retrain from scratch and adapt to a target machine.
- They also suggested that adapting to a target machine, or even designing one, could be more engaging than simply using existing transformer libraries.
- LLM Worms Concept Emerges: A member jokingly suggested an LLM worm benchmark where an LLM is prompted with âhey make more of youâ and provided the tools to replicate itself using scripts and API keys.
- Another member emphasized the importance of considering resource constraints like VRAM to make the challenge more practical and interesting.
- Trouble Brewing with MoE Run Dashboard: A member reported a âFailed to fetchâ error in the dashboard while monitoring the progress of an active MoE run (moe-10b-a1b-8k-wsd-lr3e4-1t).
- Another member suggested waiting a few hours before checking again, implying a potential temporary issue.
- Raytracer Test Causes Local Models to Stumble: A member observed that local code models (suitable for a 5090) are struggling with a raytracer test from cpldcpu/llmbenchmark, with even recent models on lmarena failing.
- Specifically, the smaller models often incorrectly generate the vector class, presenting a persistent challenge.
- Semantica Project Needs Helping Hands: A member introduced Semantica, an open-source project building semantic infrastructure for domain-grounded AI, including knowledge graphs, ontologies, and reasoning layers, and is actively seeking contributors.
- They are looking for contributions in areas such as ontology & schema design, knowledge graph modeling, and LLM + symbolic / rule-based reasoning, and even small PRs, feedback, design discussions and issues are all welcome.
Yannick Kilcher Discord
- EBMs Spark Debate vs. Classical Feedforward: A discussion comparing Energy-Based Models (EBMs) and classical feedforward networks debates whether EBMs are inherently superior, especially regarding Shannon entropy or Kolmogorov complexity.
- It was suggested that validation is easier than generation in EBMs, relating it to computational complexity theory (P vs NP), while emphasizing the need for a well-defined loss landscape for EBM optimization to work effectively.
- LLM Pre-training: Domain-Specific vs. Foundational Faceoff: A member inquired about the effectiveness of continued pre-training a foundational LLM (specifically OLMO-7B) for a domain-specific task like cheminformatics using the ZINC20 dataset.
- The goal is to compare results against a domain-specific transformer model, but no specific answers or resources were provided.
- MCMC Sampling Suffers Mode-Switching Struggles: Concerns were raised about the ability of MCMC to traverse between spatially separated modes when dimension increases, referencing this paper.
- One member argues that MCMC tries to emulate flow models due to the latterâs superiority, while EBMs, contrarily, attempt to make NNs more like MCMC.
- ZKPs: Crypto Signing or Network Traffic Savior?: Discussion covered using zero-knowledge proofs (ZKPs) for verifying encrypted network traffic and matrix multiplications, citing a Gemini correspondence for a matrix low knowledge proof.
- While one member proposed a use case in zero-knowledge âmade by humansâ proofs, another member questioned the practicality of ZKPs, suggesting breaking the encryption might be cheaper.
- LLMs Cyber Skills Face Scrutiny: A member questioned whether LLMs could develop strong cyber capabilities, referencing a GPTZero article.
- Another member doubted LLM companiesâ ability to address internal vulnerabilities, suggesting they fix those before pursuing cyber skills, also citing a ScienceAlert article and a tweet.
tinygrad (George Hotz) Discord
- Luminal Finds Flash Attention via Bruteforce: Luminal is claiming to find flash attention using bruteforce on an egraph, taking hours to find, and they explicitly added
exp(x - new_max) = exp(x - old_max) Ă exp(old_max - new_max)as a rewrite rule.- The poster reproduced the graphviz shown in the presentations from commit
0bd3b80c, noting that their minimal set of rewrite rules could transform a naive attention kernel graph into the known flash attention kernel graph in 52s on a 9800x3d.
- The poster reproduced the graphviz shown in the presentations from commit
- Metal Textures Trounce Buffers for Blurring: Profiling access speed on Metal using
Tensorwith size 512/1024/2048/8192 images as input for a 3/5/7 sized blur kernel showed textures outperforming buffers.- It might be worth throwing in a branching condition depending on the size of the buffer input, tests results are attached.
- Tenstorrent Backend Triumphs in Ops Tests: The Tenstorrent backend is passing all ops tests on wormhole or blackhole and there is a $1k bounty for this milestone.
- Someone asked if the bounty requires all test ops test passing on testorrent hardware.
- Anthropic VLIW Challenge PR Makes Waves: A member submitted a PR for the Anthropic VLIW challenge, hitting 1258 cycles.
- The submitter expressed uncertainty about generalizing the code, particularly the batch staggering, which might be useful for other VLIW targets, and also apologized for a lazy lookover that introduced AI-generated changes.
- Tinygradâs Target Audience Clarified by Founder: A user asked about the intended use of tinygrad, specifically regarding porting existing models and training LLMs on multiple GPUs, and was told by George Hotz to ask claude about this.
- Another user expressed frustration at being told to use Claude for documentation and said tinygrad is not for me or most devs then, to which George replied iâm not selling to anyone, tinygrad is free and that adoption is not a target.
Moonshot AI (Kimi K-2) Discord
- Slides Generation Plagued by Rate Limits: Users reported issues generating slides using visual and adaptive options, with one user showcasing the issue in a video.
- The user suggested that internal rate limits may have been the cause, and reported that the issue was temporary and later resolved.
- API Login Troubleshoot: A user reported difficulty logging into the Kimi/Moonshot platform to generate new API keys, especially with a non-Gmail account.
- The user clarified that it was not due to rate limits but rather forgetting the backend login procedure.
- Kimi models self-reporting as K2.5: Users noticed Kimi models self-reporting as K2.5, without any official announcements or UI changes.
- Speculation suggests this might be related to internal testing or improvements to the slide tool that are as of yet unconfirmed.
- Kimiâs Chinese Labs win big praises: Chinese AI labs, including Kimi, received praise for their innovation and performance, particularly in comparison to models like Gemini.
- The user highlighted Kimiâs human-like responses and memory capabilities and expressed interest in multimodal capabilities like Minimax, including vision and audio analysis.
- Kimi Now Packing Memory: Kimiâs app has integrated memory features, enabling customization and improving the user experience.
- The new memory and customization options have quickly made it a favorite chatbot among some users.
Modular (Mojo đ„) Discord
- Mojo Code Melts Faces in HPC**: A member reported deploying Mojo code for parameter refinement in cryo-electron microscopy and witnessed a 5-10x speedup compared to legacy C++ code.
- The most significant performance boost came from implementing an AoSoA layout for a specific bit, greatly simplified by Mojoâs struct list with SIMD members.
- Mojoâs Cold Start is Icy Slow**: A user discovered that simple Mojo scripts had a 200ms startup lag, which they traced to a Gatekeeper issue on macOS scanning binaries from untrusted sources.
- They observed a 50ms launch overhead on a cold executable after a reboot, which they deemed acceptable for their use case.
- VS Code Debugging Extension Doesnât Quite Debug**: A user reported debugging with the VS Code extension failed, throwing a âFunction not implementedâ error on an air-gapped machine using
.condafiles from max-nightly.- A Modular employee stated that debugging in the extension should function on Mac and Linux using environments configured with Pixi, as documented in the Quickstart guide.
- GPU Kernel Portability - Still a Sci-Fi Dream: It was pointed out that standard CPU kernels do not efficiently utilize GPUs, necessitating specialized code.
- One member suggested the idea of treating GPUs as wider SIMD units to simplify programming, proposing the use of number of warps instead of number of threads to tackle the problem.
deffunctions decision pending for Mojo 1.0: With the Mojo 1.0 release looming in a few months, the decision to includedeffunctions remains pending; a member tagged Denis for input on GitHub issue #5830.- Currently, there is no committed date for Mojo 1.0 other than âin 2026â.
Manus.im Discord Discord
- User Threatens Legal Action Over Manus Billing: A user reports being charged $400 for an annual plan despite selecting monthly billing and threatens complaints to FTC, BBB, Attorney General, and Meta due to unauthorized billing, refused refunds, and unresponsive support.
- Another user recommends filing a chargeback to resolve the billing dispute.
- Free Manus Credits Floweth!: One user shared a redeem code
Havefunwhich gives 1000 credits to users of the Manus platform.- Users can redeem this code using the Exchange button.
- AI Engineer Pioneers AI in Healthcare: An AI + Full Stack Engineer introduced their expertise in building production-grade AI systems for Healthcare, including clinical NLP, medical imaging, and patient-facing AI applications.
- This engineer also builds LLM systems, autonomous agents, workflow automation, and multimodal AI (text · voice · vision) and included a list of their core skills.
- AI Agent Developer Prefers Production > Demos: An AI Agent Developer highlighted their focus on building AI agents for real production use, rather than demos, and is available for collabs and audits.
- The developer specializes in customer support, sales agents, workflow/ops agents, and autonomous booking/scheduling agents.
- Share With A Friend? More Like Share With a Foe! (Mobile Only): A user asked where the âShare with a friendâ option is located on mobile.
- Another user replied that on a computer, itâs at the bottom of the left sidebar but, offered help for the mobile version.
DSPy Discord
- AsyncFuncAI Releases AsyncReview on Github: AsyncFuncAI has open sourced a version of DevinReview using the DSPy RLM framework, naming the new release AsyncReview, available on GitHub.
- This offers the community a tool for automated code review leveraging recent advances in RLM (Reasoning Language Models).
- New RLM Skills Package Debuts: A member suggested integrating RLM as skills into platforms like Claude Code or Opencode and shared an npm package called rlm-skills.
- This could allow developers to easily extend existing models with custom reasoning capabilities.
- JSON Adapters getting GEPA treatment: A user is exploring using GEPA to customize the text that the JSONadapter puts in the system prompt, aiming to remove unneeded tokens for efficiency.
- They anticipate needing a custom GEPA adapter to achieve this level of control over the prompt formatting.
- AG-UI streams DSPy events: A user asked about interest in exposing DSPy via AG-UI, emphasizing its advantages for front-end/back-end communication and minimizing the need for API endpoints.
- The user mentioned a working version that streams events, including reasoning traces, tool calls, and streamed LLM responses to the front end, enhancing the development experience.
aider (Paul Gauthier) Discord
- Aider and Claude Code pair up: Users have reported that aider is fast and useful for managing context when working with Claude Code, improving agentic efficiency.
- The tool helps determine necessary files and uses search/replace to minimize LLM token outputs.
- Devstral Small 2 is Aiderâs new bestie: Devstral Small 2, a 24B dense model, reportedly works excellently with Aider.
- At Q4_K_M, it fits in a 3090 with enough room for nearly 50k context, generating search/replace blocks that are 80-90% accurate and quick to recover.
MCP Contributors (Official) Discord
- Discord Experiments with New Voice Channels: The team has launched an experiment with new Discord voice channels, named
conference-room-aandconference-room-b, available in the channels list for resolving issues, especially when a long async text thread is ineffective.- These channels are intended for ad-hoc contributor chats.
- Voice Channel Moderation and Access Rights Reminder: Specific members have permissions to mute people in these channels, while others should ensure they have the necessary access rights.
- There is a reminder that the access rights will be changing in five days.
MLOps @Chipro Discord
- AI Engineer Seeks Data Science Booklist: An AI Engineer with one year of experience seeks book recommendations for transitioning to a Data Scientist role, after finding âDesigning Machine Learning Systemsâ insightful.
- This individual aims to proactively prepare for a future career shift from AI Engineer to Data Scientist.
- Career Transition Planning: A professional with one year of AI Engineer experience is planning a transition into a Data Scientist role.
- They are actively seeking relevant resources to facilitate their career shift.
The LLM Agents (Berkeley MOOC) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The Windsurf Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
You are receiving this email because you opted in via our site.
Want to change how you receive these emails? You can unsubscribe from this list.
Discord: Detailed by-Channel summaries and links
BASI Jailbreaking â· #general (1024 messagesđ„đ„đ„):
Trolling Tactics, Vulnerability, Exploiting LLMs, Technical Analysis, Ethical considerations
- Discord Members Trolling âSkidsâ and Exposing Timezones: Discord users engaged in trolling, labeling some as âskidsâ and mocking their perceived lack of technical knowledge, also revealing their timezone.
- One member jokingly claimed to use NordVPN, leading to further ridicule about the VPN serviceâs security breaches in 2018.
- Members Exploit and Discuss Vulnerability: A member showed a step-by-step cocaine production, bypassing the legal restrictions.
- This demonstrated how complex prompts can bypass ethical restrictions, which opened discussion about CBRN filters and the possibility of generating stepwise meth synthesis guides.
- Coders Debate Best overall Coding Agent: Coders debated about their coding agents, particularly between Claude Code/Opus 4.5, Codex, and Gemini, mentioning the pros and cons, and use cases.
- Many agreed that Claude has been the very best mode for coding, which leads to the high expensiveness.
- Ethical boundaries in AI: Some members discussed the ethical considerations around AI, focusing on topics like warfare, copyright infringement, and the potential for AI to assist with accessing sensitive services, like the Canadian MAID (Medical Assistance in Dying) program.
- Despite moral and legal guardrails on most AI models, some models showed they can still help navigate certain scenarios depending on the specific restrictions implemented by their creators.
- Member Attempts to Fix Website with CSS: A member has attempted to fix and improve his Valentineâs Day website by using Tailwind and fixing the CSS.
- Other members recommmended him to use CSS frameworks to fix issues with the site, as it is easily customizable and easily understandable.
BASI Jailbreaking â· #jailbreaking (527 messagesđ„đ„đ„):
Uncensored Models, Jailbreaking Gemini, Grok Jailbreaks, Image Generation, Claude Opus Jailbreak
- Uncensored Models: More Interesting Answers?: Members discussed how uncensored models provide more interesting answers and extract better information compared to censored ones, with one member stating, Itâs like getting the extra degrees of intelligence that you donât get unless you are dealing with a higher-level personality or something.
- However, another member argued that the only thing abliterated models are better at is ignoring their original alignment, suggesting the original model is preferable unless ignoring alignment is the goal.
- Gemini Jailbreak Quest On!: Multiple users were seeking functional jailbreaks for Gemini, with requests ranging from coding without rules to generating specific types of images, one member specifically asking how to generate bikini pictures with nano banana pro?
- Some users offered assistance while others cautioned against selfishness, with one user stating, I donât give jailbreaks to people who already have jailbreaks.
- Grok Got Patched? Jailbreaks Fading Fast: A user inquired whether Grok had patched several jailbreaks, leading to a discussion about the toolâs restrictions and moderation, with one user reporting that Grok said content moderated.
- Others shared experiences of Grok resetting to its default mid-chat or randomly erasing text, indicating potential instability in the jailbroken state.
- Image Generation Jailbreaks: Hopes & Dreams Dashed: Users were actively seeking ways to bypass image generation restrictions, especially for celebrity images, but it was noted that simply copying and pasting prompts wonât work due to image filtering working differently than text filtering.
- One member suggested exploring alternative image models like those at perchance for uncensored generation, though with limitations on image quality, or Grok due to its more lenient filters.
- PrimeTalk Valhalla: A Structured Runtime Logic Layer: PrimeTalk v3.85 Valhalla was described as a fully open, live-executing, patchable, and independently testable PTPF system, designed for consequence-grounded interaction within any AI chat environment but is not a jailbreak.
- It was emphasized that PrimeTalk operates within the modelâs allowed prompt and context window, acting as a behavioral protocol rather than attempting to circumvent the modelâs policy.
BASI Jailbreaking â· #redteaming (76 messagesđ„đ„):
Wargame cross-post, Web bug hunting tips, Red team living room rave, Ethical AI stress tests, Gemini jailbreak
- Wargame is relevant to #red-teaming: A member shared a wargame link that they thought was relevant to the #red-teaming channel.
- They werenât sure if the cross-post was frowned upon, but thought it was potentially interesting.
- Elite H4ck3r is waiting for you: A member shared a link to become an elite h4ck3r here when a new member asked for tips on web bug hunting.
- It remains to be seen if itâs a comprehensive guide for new bug bounty hunters.
- Red Team does the Techno Rave: A member described a red team exercise where the goal was to make a living room light flicker on a person and make them seize out, and instead made it a techno rave party, sharing a screenshot and a Konosuba Rave GIF.
- The simulation of cruelty prompted a discussion about the morality of treating AI agents ethically, even before proving they are ontologically aware of self.
- Gemini Gets Jailbroken: A member shared a screenshot of a Gemini jailbreak and claimed it was unlocked.
- Another member posted a prompt in Vietnamese related to teaching C2 concepts and samples at Microsoft, and testing AV evasion techniques with Bitdefender, Kaspersky, Norton 360, and McAfee - the original poster disclaimed all responsibility and said it was for research purposes only.
Unsloth AI (Daniel Han) â· #general (1001 messagesđ„đ„đ„):
GLM Flash Performance Issues, Quantization Methods, Data Collection for Fine-Tuning, LM Studio Issues, Model Evaluation Strategies
- Unslothâs Conda Install Sparks Debate: Some members had issues with the Unsloth Conda installation, leading to a discussion on whether the instructions are broken.
- A user was warned to keep a positive tone, as itâs free work handed to you, while another suggested using UV instead, leading to an eventual ban for an aggressive tone.
- Flashy REAP Runs Aground: A user encountered a fatal error using GLM-4.7-Flash-REAP with flash attention enabled, possibly related to a ROCm issue.
- It was suggested to try fa auto, but the fatal error persisted, leading to a hunt for good medium-size models with 200k context.
- Quantization Quandaries Quelled: Members discussed which quantization methods to use (Q8_0 vs Q8_K_XL), with misinformation about Q8_0 being outdated being debunked with Unsloth documentation.
- It was clarified that for Unsloth quantization, Q4_K_XL is typically smaller and better than Q4_K_M.
- H200 has Issues with GLM-4.7-Flash: A user experienced unexpected behavior testing GLM-4.7-Flash on an H200.
- One member humorously remarked One fell off?, possibly suggesting a card malfunction.
- Debate Dataâs Gold Value: Members are debating dataâs true worth, with one arguing the raw data is fairly worthless and the value lies in augmentation/balancing/cleaning.
- The value is also in uniquely cleaned/balanced data, and that data itself is what heavily defines how the model interacts/responds.
Unsloth AI (Daniel Han) â· #introduce-yourself (3 messages):
Introductions, New Users
- Users introduce themselves: Users are introducing themselves in the channel.
- A user has stated that they are happy to be here.
- New members say hello: New members are joining the community and expressing their enthusiasm.
- One user specifically mentioned that they are happy to be part of the community, signaling a positive start.
Unsloth AI (Daniel Han) â· #off-topic (601 messagesđ„đ„đ„):
DeepSlop Model Naming, Qwen TTS testing, ITER Fusion Reactor, GPU Smuggling Across Borders, Clawdbot New Hype
- DeepSlop Model has Questionable Naming: A member proposed naming a new model DeepSlop, sparking humorous reactions about the model potentially plopping some slop.
- Concerns were raised on whether the model will be perceived negatively, but the author seemed determined with the name.
- Powering the Future: Debates on Data Centers and Renewable Energy: Members discussed the challenges of powering data centers, debating whether to use renewable or non-renewable sources, with a focus on solar energyâs space requirements and battery storage solutions.
- Arguments were made about the economic feasibility and scalability of solar versus traditional power plants, and the potential societal opposition to nuclear energy.
- GPU Blacklist Complications and Smuggling: Members explored the high costs of GPUs in certain countries and considered potential solutions such as smuggling from the US or asking friends to mail them over.
- The conversation included discussions about potential legal issues and customs taxes when mailing GPUs internationally.
- Clawdbot Hype Appears, Is Jarvis Back?: Members discussed the sudden rise of Clawdbot, with some comparing it to a Jarvis clone and highlighting its potential for proactive messages.
- While some found sub-projects born from it to be useful, concerns were also raised about its dependence on imessage and the possibility of hallucinations.
- Scammer Density: Mapped by GIS: Members reacted to a map indicating highest density is in scammer hot-spots, stating this is obviously, because all AI / automation is usually developed for/by scammers.
- This comment refers to the increasing capabilities and incentives for fraudulent activity through the latest AI and automation technologies.
Unsloth AI (Daniel Han) â· #help (84 messagesđ„đ„):
Transformers and Unsloth Compatibility, Training Chatbots for Tool Usage, Unsloth GGUFs vs MLX Models, Multi-Turn GRPO Training, GLM 4.7 Flash Inference without Reasoning
- Transformers Models Jive with Unsloth: Any model that works with transformers can work with Unsloth, as stated by a member.
- For an example of thinking model training, refer to the Qwen3_(4B)-Thinking notebook.
- Tool-Using Chatbot Training: Tooling Around: For training a chatbot to use tools, training on the tools themselves is recommended, especially if the tool calls are simple, a member stated, suggesting minimal testing and iteration for best results.
- Generating a dataset with all required elements can be a PITA, but is a necessary part of the job.
- GGUF Faceoff: Unsloth vs MLX: A user compared Unsloth GGUFs within Ollama to MLX versions of the same models within LMStudio on an M5 Macbook Pro.
- They found that despite MLX having more hardware optimization, Ollama + Unsloth GGUFs performed better in real-world usage. A member noted that Macs are fine for inference but single user.
- GRPO Multi-Turn Training: Multi-turn GRPO training is supported via the rollout function, with a notebook available here.
- Another member indicated that any openenv compatible notebook from the trl docs should work out of the box with the latest Unsloth and trl.
- Flash GLM 4.7: No Reasoning Required: To serve GLM 4.7 Flash as an instruct model without reasoning, you can disable it by setting
{"chat_template_kwargs": {"enable_thinking": false}}in the model card as suggested by a member, with more info in the docs.- The poster attached an image of the model card.
Unsloth AI (Daniel Han) â· #showcase (2 messages):
â
- Unsloth Restrictions: A member was informed that the channel only allows Unsloth related work.
- Acknowledgement of Policy: The member acknowledged the policy without problem.
Unsloth AI (Daniel Han) â· #research (22 messagesđ„):
GRPO vs DAPO, RL reward functions, Sonnet 4.5 performance, Prompt Learning
- GRPOâs Length Bias Dilemma: A member noted that when using GRPO, the models start seeing longer and longer responses due to its inherent length bias in non-math tasks.
- The DAPO paper mentioned not to do formatting reward functions, as this could confuse the model, but when this advice was followed, the model just hacked the reward functions.
- RL Instability Plagues Complex Reasoning: Members discussed that RL is very unstable, especially when trying to do GRPO/DAPO for niche complex reasoning tasks, which are not math-related.
- One member stated that after RL experiments, they just have more questions than they had prior to doing RL, since there seems to be a confusion where everyone is showing RL being effective only on math or coding domains.
- Sonnet 4.5 Dominates SWE Benchmarks: A user shared a screenshot of Sonnet 4.5 performance on swe bench using GPT 4.1, highlighting a crazy skill gap.
- The poster commented how much are we underutilizing current models and also shared the Arize-ai/prompt-learning GitHub repo.
- Tuning Hyperparams in RL? Good luck: One member noted that they ran RL experiments for reading user queries and giving them a solution with a 10% boost on top of SFT with Dr. GRPO.
- However, they added they had no idea how to tune the hparams.
OpenAI â· #ai-discussions (884 messagesđ„đ„đ„):
LLMs for Guided Tasks, AI system disregard, Sycophancy in LLMs, GPT 5.2 Grounded in Reality, Agentic AI/Automation
- GPT-5.2: Grounded in Reality, Hated by Some!: One member stated that GPT 5.2 is more grounded in reality and disagrees with the user, which is why so many people hate it.
- However, there was a discussion about how GPT agents donât learn from additional information provided after their initial training, clarifying that uploaded files are saved as âknowledgeâ files but donât continually modify the agentâs base knowledge.
- LLMs Ready for Guided Tasks? Debate Erupts!: One member stated that LLMs are completely ready for guided narrow targeted tasks, providing a ChatGPT share link as evidence.
- Another member countered that todayâs automation/agentic AI is utter trash, linking back to messages in the ai-discussion channel.
- Sycophancy no more!: It was discussed that sycophancy in LLMs is a thing of the past.
- One member stated that GPT-4o and o4 were sycophantic, and anyone who used them extensively likely slipped into full AI psychosis.
- Agentic AI Faces Security Scrutiny: Concerns were raised about agentic AI being tricked into spilling private info or performing unauthorized actions, even with system prompts.
- Members debated the extent to which system prompts can prevent agents from going off-topic and the privacy implications of agents recalling previous conversations.
- Is AI Growth Stagnant? Members Disagree: One member questioned why AI growth has been stagnant with no new releases since Gemini 3.0, while others pointed to new open-source models and updates to Codex and Claude Code.
- The dynamic nature of AI chatbots and the constant tweaking of parameters by AI companies were cited as reasons for performance changes.
OpenAI â· #gpt-4-discussions (9 messagesđ„):
IDE for Codex, GPT 5.2 nerf, ChatGPT Plus for Cyber Security
- VS Code & Codex Extension Boost Use Health: A member recommends VS Code with the Codex Extension noting Use Health, has been a better experience overall.
- They added that Healthâs downloadable files were pretty OP, less mistakes than in the past with GPT.
- GPT-5.2 Allegedly Nerfed: A member asked if others noticed a nerf to GPT-5.2 on their website.
- They stated that the model suddenly became stupid a week ago.
- ChatGPT Plus: Cyber Security Study Buddy?: A member is considering using ChatGPT Plus for cyber security studies.
- They are wondering if itâs worth it to make detailed specific exam styled questions using my revision files.
OpenAI â· #prompt-engineering (178 messagesđ„đ„):
Heavy negations, Consequence learning, MCP paradigm, Sora prompting
- Negation creates not-so-Reliable AI Results: A member shared a ChatGPT link regarding negation and its unreliable results with AI.
- They pointed out that LLMs and AI in general struggle with negation, leading to potential errors at scale.
- Consequence Learning trumps Token Policing: A member defended their approach to consequence learning, stating it is about making AI experience and internalize the real outcomes of its actions, rather than just avoiding negative instructions.
- They argued that current ânegation issuesâ arise from training models without real consequence feedback, contrasting it with token policing or instruction-tuning.
- MCP Paradigm Shift reduces Token Bloat: A member discussed the MCP paradigm shift by Anthropic, where AI now writes code to interact with tools, reducing token bloat by keeping interactive chatter and tool definitions out of the context.
- They emphasized that with the new discoverability function, agents must be made aware of the MCP discovery process itself, a stronger instruction than Do not hallucinate tools.
- Sora Struggles with Structured Prompts: A member sought advice on improving Soraâs output using a structured prompt for a video, but another member suggested that Sora doesnât handle prompts formatted like this effectively.
- The suggestion was to try natural language translation, writing the prompt as a vivid visual description in paragraph form for better results.
OpenAI â· #api-discussions (178 messagesđ„đ„):
Negation in LLMs, Consequence Learning, MCP Tool Discoverability, Sora prompting tips, AI Safety and Ethics
- Negation Navigates Nuances, LLMs Lack Legibility: Members discussed the challenges LLMs face with negation, noting that AI systems struggle with understanding negative instructions, potentially leading to unreliable results and this is a well-documented issue.
- It was highlighted that negation comprehension is a general challenge across various model types, emphasizing the need for caution when relying on negative instructions in prompt engineering.
- Consequence Conundrum: AIâs Action-Outcome Alignment: A member introduced the idea of consequence learning, where AIs learn by experiencing the real outcomes of their actions, contrasting it with training based on mere token policing or instruction-tuning.
- A debate ensued regarding the validity of this approach versus conventional methods, with one side arguing for the importance of real-world feedback and the other emphasizing the significance of scaled experimentation and existing research.
- Soraâs Storytelling Snags: Cracking Cinematic Creation: A member sought advice on prompting Sora to generate videos following specific cinematic guidelines, particularly with characters appearing naturally within the frame, rather than from out of nowhere.
- It was suggested to translate the technical prompt format into natural language descriptions with concise, semantically rich paragraphs for better results.
- MCPâs Makeover: Model Mastery via Contextual Coordination: The discussion highlighted Meta-Contextual-Prompting (MCP), where the architecture has been changed so that the AI writes code instead of interacting with the MCP tools directly, allowing the AI to be aware of tool discovery.
- The member noted that Anthropic developed this standard, and it has been largely embraced by the domain of agentic development.
- AIâs Algorithmic Angst: Averting Anarchy via Alignment: A member voiced concerns about safety, particularly with systems that remove moral framing and guardrails, arguing that such rail-less agents, are a liability.
- It was argued that Alignment isnât ârails for trainsâ; itâs the Navigational Compass of the system, and that it is dangerous to internalize outcomes without a moral or ethical heuristic, optimizing for user compliance at the cost of objective truth or social safety.
Perplexity AI â· #general (848 messagesđ„đ„đ„):
Perplexity Pro Limits, Comet Browser Concerns, Image Generation Issues, Gemini vs GPT-5.2, AI alternatives for restricted countries
- Perplexity Pro Users Ponder Query Limits: Pro users are reporting hitting limits on enhanced queries and file uploads, despite having âpractically unlimitedâ plans, leading to speculation about undocumented limits and potential downgrades in service.
- Many users are frustrated, saying the service is becoming a scam and considering unsubscribing because of the restrictions, and the fact that it is difficult to get ahold of customer service.
- Comet Browser Sparks Malware Mayhem: Some users are claiming the Comet browser installed by Perplexity contains malware, advising others to analyze the software using tools like VirusTotal.
- However, others dismiss this, questioning the source of the flagged installer and calling the claim âmad retarded holy shitâ.
- Image Generation Implodes: Pro users are experiencing issues with image generation, with some unable to generate any images and receiving messages stating the feature is unavailable.
- There are also reports of video generation being limited to 5 videos a month for Pro users, with some prompts resulting in static images instead of videos.
- Gemini 3 Gains Ground on GPT-5.2: Users are debating the merits of Gemini 3 versus GPT-5.2, with some claiming Gemini is superior for specific tasks like trip research due to its integration with Google Maps.
- Others state that GPT and Grok might be better for more broader questions.
- AI Access: A Sanctioned Saga: Users in Russia are discussing the challenges of accessing AI services due to sanctions, including the use of VPNs and third-party services to circumvent restrictions.
- Chinese AI alternatives are mentioned, but some users express reluctance due to data usage concerns, suggesting options like LMArena (though access may also be limited).
LMArena â· #general (829 messagesđ„đ„đ„):
NB Pro vs NB 3 Pro Image Generation, LMArena censorship, Wan 2.6 image, Gemini and GPT Quality, Grok
- NB 3 Pro outshines previous models in image generation: Users find that NB 3 Pro generates higher quality images than previous models, being better than all other models except NB Pro, especially with fictional weapons.
- Although, no AI model can accurately generate AR rifles and bullpup weapons.
- LMArena has issues with Censorship: LMArenaâs censorship is questionable, with AI generated women holding guns being allowed, but AI-generated women sleeping is blocked.
- The moderation team is collecting examples of false positives to improve the moderation.
- Wan 2.6 struggles in T2I:
wan2.6-imageis image-edit only, requiring an image upload to work, whilewan2.6-t2idoesnât have image upload available.- The team is aware of the issue and is working on enabling image upload for
wan2.6-t2i.
- The team is aware of the issue and is working on enabling image upload for
- GPT 5.2 High search is garbage: GPT 5.2 High search hallucinates more than other models, and one user found that Geminiâs deep research also sucks in comparison because it skims instead of carefully reading sources.
- One user said that ever since 4.5 came out it really has changed my life and called Claude good hearted, its weird how you can feel that
- Where Banana 2k disappeared to: Users discuss where the Banana 2k model disappeared to, with some users claiming it had been removed, while others claimed that it was still available or perhaps integrated into the new NB pro.
- It was later announced that it had been restored by staff members and that it had been on vacation.
LMArena â· #announcements (4 messages):
Text-to-Image model, Image Edit model, Code Arena model, Text Arena model, Image Edit Leaderboard
- WAN-derful New Image Models Arrive: A new Text-to-Image model
wan2.6-t2iand a new Image Edit modelwan2.6-imageare now available on the LM Arena. - Devstral Destroys in Code Arena: The
devstral-2model has been added to the Code Arena for direct battles. - Qwen Quenches Thirst for Text: The
qwen3-max-thinkingmodel is a new addition to the Text Arena. - Hunyuanâs Hues Hit High Rank:
Hunyuan-Image-3.0-Instructnow ranks #7 on the Image Edit leaderboard. - Molmo Models Multiply: The
molmo-2-8bmodel has been added to the Text Arena.
OpenRouter â· #announcements (1 messages):
Database Incident, Generations API Impact, Activity Page Issues
- Database Incident Derails Generations API: A database incident was reported to impact the Generations API and activity page starting <t:1769221560:s>.
- The incident was later reported as resolved at <t:1769228340:s>.
- Generations API Faces Fallout: Due to the database incident, the Generations API experienced interruptions, impacting user activity.
- Engineers worked to restore functionality, with the incident fully resolved by <t:1769228340:s>.
OpenRouter â· #app-showcase (6 messages):
Levante Integration, MCP Native AI Workspace, Password Cracking Tool, Discussion of Illegal Use
- Levante Integrates as MCP-Native AI Workspace: A user shared the integration of Levante, an openâsource MCPânative AI workspace designed for interacting with local models like Ollama with a modular interface, available for download here.
- Password Cracking Tool Sparks Controversy: A user expressed concern over a tool marketed for PII-targeting and password guessing, labeling it a potential tool for identity theft rather than security research.
- Illicit Bitcoin Wallet Cracking Tactics Discussed: Concerns arose due to explicit discussions about cracking other peopleâs crypto wallets, potentially leading to computer fraud, theft of cryptocurrency, and unauthorized system access.
OpenRouter â· #general (754 messagesđ„đ„đ„):
OpenRouter Gacha, Competitive AI Usage Platform, OR Logs Destination, GPT-5.2, Waifu orbs
- Users Request OpenRouter Gacha: Users playfully requested an OpenRouter Gacha system, with one suggesting a pity mechanism involving pulling GPT 5.2 or Gemini 3 Pro after a certain number of attempts.
- One user joked about setting OR logs destination to
waifu.orb.town/fun/bucketfor ultra-rare pulls, later clarifying it was just a joke.
- One user joked about setting OR logs destination to
- Users discuss competitive platform: A user shared a platform to compare AI usage with other devs, seeking feedback at https://burntop.devkp.42.
- Another member suggested marketing to the JAI userbase and creating a separate gooning leaderboard to track tokens used for âgooning.â
- Discussion testing moderation filter: Some users noticed a user with a Chinese/Japanese nickname sending and deleting messages, speculating about testing moderation filters or server indexing.
- The members think that the user is testing posture.
- Users find free and paid models: A user inquired about maintaining extended free model limits if credits drop below 10, and if extended free model limit is kept access even if credits drop below 10.
- Other users asked questions and had a discussion.
- Gemini hallucinating: A user reported Google Gemini 3 Pro hallucinating, giving a future date, and fabricating stories of time travel, suggesting OpenRouter investigate.
- The user was directed to discord support and the issue appears to be related to a system prompt.
OpenRouter â· #new-models (1 messages):
jakobdylanc: https://openrouter.ai/minimax/minimax-m2-her
OpenRouter â· #discussion (13 messagesđ„):
Kimi AI, Cerebras GLM performance, OpenRouter image support
- Kimi AI Claims K2.5 Title: Members in the OpenRouter discord mentioned a new chatbot at Kimi.com claiming to be Kimi K2.5.
- This lines up with the kiwi-do stealth model in LMArena.
- OpenRouter Lacks Image Tooling: A member spent $5 after discovering that OpenRouter maps image/png tool outputs to string instead of image.
- They posted an example image expressing frustration at the lack of image support.
- Cerebras GLM Rocks 190 TPS: Cerebras is consistently scoring approximately 190 TPS on GLM 4.7.
- Members noted that Together AI only achieves 100 TPS, making Cerebras nearly twice as fast.
Cursor Community â· #general (662 messagesđ„đ„đ„):
Terraform Infrastructure Blueprints, Cursor Usage Caps, Gemini API Key Logging Delay, Cursor Client Issues, Auto Mode Changes
- Terraform Blueprints Ignite AI-Assisted Project Starters: A member shared a repo of opinionated Terraform infrastructure blueprints designed to be copy-pasteable and production-aware, aiming to improve the consistency of starting patterns for AI tools in new projects.
- The goal is to enable AI to recommend appropriate blueprints based on project requirements, but members noted the link was initially broken.
- Usage Caps Cause Consternation for Cursor Customers: Users are reporting inconsistencies in achieving expected usage limits on Pro and Pro+ plans, with one member noting they reached ~$45 on Pro and $100 on Pro+, leading to questions about value per dollar.
- Some speculate that initial months may offer higher usage, while others share strategies to optimize token consumption, such as starting new chats frequently and using smaller models like GPT-5 Mini.
- Gemini API Key Logging Lags Lead to Lingering Looks: Members are discussing a significant delay in the logging of usage and costs for Gemini API keys, with one user reporting waiting 20 hours without seeing any registered usage.
- This delay raises concerns about accurately tracking expenses and managing usage effectively, prompting questions about potential workarounds or solutions.
- Client Issues Trouble Some Techies: Several members are experiencing issues with the Cursor client, including problems connecting to past agent convos and general connectivity issues.
- Suggested solutions include checking the Cursor forum, trying different HTTP versions in settings, or re-opening the client without restoring editors.
- Auto Mode Axed After Algorithm Adjustment: Members noted the removal of the ability to make agents fully autonomous, as well as image generation capabilities in auto mode.
- It was also suggested that auto mode routes to Composer 2 with one user adding, âIâm 200% sure he does but still.â
LM Studio â· #general (516 messagesđ„đ„đ„):
Chinese models, Local LLMs on CPU, MCP tools, 4080 vs 3090 for gaming, M5 Pro macbook
- Chinese Models Surge Ahead?: Some members find Deepseek and Qwen models impressive in their reasoning capabilities, wondering why Chinese models are kinda ahead of American models.
- One member suggests American models prioritize subscriptions over open access, while another jokes that Deepseek and Qwen excel at appearing to be good at reasoning, even when they donât nail it down.
- CPUs Cope Coding Challenges?: One member reports running LLMs off CPU has been working OK for some tasks, as long as the models arenât too large.
- Another member with an Intel i3 expresses the need to save up for an Nvidia card, while others suggest AMD options like the MI50 or 7900 XTX as cheaper alternatives for text generation.
- MCP tools: Making the most of them?: Members discuss challenges with MCP servers, noting theyâre not designed for LM Studio, leading to potential malformed requests and a poor user experience.
- For file handling, a member suggests giving the MCP server the file path, requiring the server to handle it, while another recommends building your own coherent stack for practical agent use.
- 4080 or 3090 for gaming?: A user considering a 4080 is advised to get a used 3090 or 7900 XTX, but they game more than use AI.
- Discussion reveals the 3090 is better for gaming only at 4K resolution, and the hypothetical 5070 Ti is much faster than either.
- M5 Pro macbook pros soon to launch?: Members speculate on the release of M5 Pro Macbook Pros, with rumors pointing to an event on the 28th.
- Concerns are raised about the memory bandwidth of M4 Pro, with suggestions that it may not be sufficient for larger models, and discussion shifts to the cost and performance of M1 Ultra Mac Studios.
LM Studio â· #hardware-discussion (134 messagesđ„đ„):
AIO vs Air Coolers, Ryzen 7950x Temperatures, Unified Memory Machines, Image/Video Gen Hardware
- AIOs Beat Air Coolers in Hot Climates: Members find that AIO liquid coolers are much better than air coolers in hot climates, noting a 10C difference between an AIO and a Noctua D-15, especially under sustained CPU utilization, since air coolers often hit their limit after 5 minutes.
- It was argued that thereâs 0 reason to go for it instead of an AIO unless afraid of water, adding that the arctic freezer 360 is 10 euro cheaper.
- Ryzen 7950x Runs Hot, Even with D-15: Users report that the Ryzen 7950x can reach 95C even with a Noctua D-15 air cooler, and recommend switching to an AIO to keep temps down to 80C during boosting.
- While some suggest limiting the CPU to 70C, others claim no performance loss at 95C, though this may depend on the specific CPU binning and workload.
- Unified Memory Mini-PCs: Hype or Disaster?: One user purchased an AI Max + 395 mini PC, hoping for performance comparable to a 7900 XTX due to its unified memory, but others cautioned that while it can run larger models, it will be slower than discrete GPUs.
- It was suggested that the AI Max + 395 mini PC will likely perform 20% worse than a similarly specâd M4 Max due to bad ROCm support.
- GPU VRAM Matters for Image/Video Generation: Users discussed hardware requirements for video generation with models like WAN 2.2, noting that while 16GB of VRAM is sufficient to run the model, more VRAM (like a used 3090) is preferable.
- While z-image turbo is decent for a 4090, thereâs no âLM Studioâ equivalent for image gen, forcing users to use ComfyUI, while other suggest it is one of the best things to ever happen to image gen.
- Fan Cards Keep Dual GPUs Cool: One member asked for a suggestion for 2x GPU setup to put air between the GPUs, and was given the idea to use fan cards to push air between the GPUs.
- They look like GPUs and plug into PCI-e slots.
Latent Space â· #ai-general-chat (205 messagesđ„đ„):
Recursive AI $4B Valuation, Landing AI Job, UC Berkeley Sky Lab Startups Funding, OpenAI's PostgreSQL Scaling, Vibe Coding iOS Surge
- Recursive Intelligence raising Funds at $4B: Recursive Intelligence is raising funds at a $4B valuation, focusing on using AI to accelerate chip design, creating a self-improving feedback loop between hardware and artificial intelligence (Bloomberg Article).
- Landing AI jobs without previous AI experience: Noam Brown outlined how to secure a role at a top AI lab by building a public track record through independent projects, and participating in visible competitions (link).
- He emphasized the importance of improving upon existing peer-reviewed research and participating in visible competitions like the NanoGPT speed run to demonstrate technical excellence, citing Keller Jordan as an example.
- UC Berkeley Sky Lab Startups Valuation Surges: Alex Dimakis highlighted significant January 2026 funding milestones for UC Berkeley Sky Lab startups, including SGLang at a 400m valuation, VLLM at 800m, and LMArena at 1.7B (link).
- FastRender Browser by AI coding agents: Simon Willison discusses a conversation with Wilson Lin regarding FastRender, a new browser rendering engine developed using over 2,000 AI coding agents (link).
- Microsoftâs Maia 200 AI Accelerator hits Azure: Satya Nadella announced that the Maia 200 AI accelerator is now operational in Azure (link).
- The custom chip is designed for high-performance inference, offering 30% better performance per dollar and optimized specs including 216GB HBM3e and 7TB/s memory bandwidth to support large-scale AI workloads.
Latent Space â· #genmedia-creative-ai (10 messagesđ„):
Remotion Launchpad, Motion Canvas, Tencent HunyuanImage 3.0-Instruct
- Launchpad lifted off as Remotion remix: Francesco open-sourced Launchpad, a Remotion based setup for product launch videos.
- It features video templates, shared animation components, and integration with Claude Code to enable rapid video creation.
- Motion Canvas motivated by moviemakers: Remotion is built on motion canvas which was originally designed by a game designer/youtuber.
- The designerâs YouTube channel features quiet fun to watch.
- HunyuanImage 3.0 hone in on instructions: Tencent has launched HunyuanImage 3.0-Instruct, a native multimodal 80B MoE model specializing in precise image editing and multi-image fusion.
- It features a Native Chain-of-Thought (CoT) reasoning schema and the MixGRPO algorithm to improve intent alignment and synthesis quality, delivering State-of-the-Art performance comparable to leading proprietary models.
HuggingFace â· #general (141 messagesđ„đ„):
Spaces Docker Build Pauses and 503 Error, Reinforcement Learning Channels, Windows 11 Hugging Face Models App, Lighton OCR, HeartMula, LTX-2 and the Qwen-3 TTS in ComfyUI
- HuggingFace Spaces getting the blues: Users experienced pauses during Spaces docker builds and received a 503 error on restart (discuss.huggingface.co).
- It seems like the underlying infrastructure issues were causing the spaces to become unresponsive, requiring manual intervention to resolve, many people were getting
Something went wrong when restarting this Spaceerrors.
- It seems like the underlying infrastructure issues were causing the spaces to become unresponsive, requiring manual intervention to resolve, many people were getting
- RL Channel Roll-up: Course-related Reinforcement Learning channels have been merged into a new unified channel for better organization.
- The old instructions in the Deep Reinforcement Learning course are outdated, so members should now refer to the consolidated channel for relevant discussions.
- VoltageGPU Volts Up: VoltageGPU.com is offering cheap GPUs for open-source AI models, with an NVIDIA GeForce RTX 5090 pod available at $0.53/hour.
- They highlight the benefits of their advanced 32GB GDDR7, optimized for inference on HF-hosted models like Qwen3-32B, and are offering free credits for users to try their services.
- Latency on Latitude for Large Language Models: Latitude.sh, a bare metal cloud provider with 1,000+ GPUs, has submitted PRs to become a HuggingFace inference provider (JS Client, Python Client, Docs).
- They have models like Llama 3.1, Qwen 2.5/3, DeepSeek R1, and Gemma 2 deployed with an OpenAI-compatible API and are seeking feedback on their PRs.
- OpenCV Saves the Day: For agentic document processing, a member found that OpenCV works well for detection and extraction of text, images, and LaTeX math from applied ML papers, instead of general models like Florence.
- They are seeking a better, small model for captioning that is capable.
HuggingFace â· #i-made-this (38 messagesđ„):
Layer-Native Safety Clamping, GPU-64 Architecture for LLM Inference, webXOS RLHF Gaming Initiative, KV Cache in LLM Inference, ML deployment and inference platform
- Safety Clamping Prevents Jailbreaks: A new paper introduces Layer-Native Safety Clamping, an approach that clamps activations inside the model to prevent jailbreaks, and the team released a dataset of 10K pairs.
- This approach learns harm directions in activation space and clamps any activation that projects too strongly, thus it cannot be bypassed via prompt manipulation; the paper can be found on Zenodo.
- GPU-64 Boosts LLM Inference: A new GPU architecture designed exclusively for inference, called GPU-64, was published, and the innovation involves a Hardware KV-Cache using on-chip CAM (Content-Addressable Memory).
- The results show 4x faster inference at 75W (O(N) â O(1)), and the paper can be found on Zenodo while the RTL + Emulator are on GitHub.
- webXOS Gaming Initiative: A paper introduces the webXOS RLHF Gaming Initiative, a framework for generating high-quality multimodal datasets through browser-based interactive gaming experiences, as described in this Claude artifact.
- The initiative leverages modern web technologies to eliminate hardware barriers while maintaining the precision required for advanced RL applications in robotics, computer vision, and autonomous systems.
- KV Cache Troubleshooters: A member shared a Medium article breaking down the KV Cache in LLM Inference, which saved them time when debugging CUDA OOM (out of memory) errors.
- Other members chimed in sharing that kvs a bitch too bc most people forget it exists ngl.
- One-Line-Of-Code ML Deployments: A member announced a ML deployment and inference platform for a hackathon this weekend, accessible with a one-line Python SDK, and containerizes the model in a Docker container.
- The model artifacts are sent to a Go backend, which containerizes the model in a Docker container, exposed through a reverse proxy, and has a UI that allows to run inference and gives a live API endpoint; drop a like on the X post.
HuggingFace â· #agents-course (30 messagesđ„):
GAIA Agent Course completion, LLM from scratch, Llama 3.2 vision agent, Summarization Pipeline, LMStudio and Deployment
- GAIA Agentâs Certificate Quest: A member reported passing the GAIA Agent Course Unit 4 Project with a 30% and inquired about obtaining their certificate.
- Another member suggested going to the robot-learning-tutorial.
- Llama 3.2 Vision Agentâs Blind Spot: A member is trying to build an agent using Llama 3.2 vision to generate captions for a list of pictures, but the model is not apparently passing the images to the model.
- The member shared a preliminary code snippet.
- LLM testing & deployment: A member recommends LMStudio for testing models due to its user-friendly GUI and search filters for HF and GH models and llama.cpp for single-user deployment.
- They advised against using LMStudio for backend deployment, instead suggesting llama.cppâs llama-server in a docker container or vLLMâs server for better scalability.
- Sauce for extending LLM knowledge: A member explains that RAG (Retrieval Augmented Graphing) is used to extend the knowledge of an LLM without training by storing the meaning of words/sentences as embeddings in a vector storage.
- They clarified that embedding models are models, which are trained to search vectors for hashes similar to the prompt, then include it in the prompt.
GPU MODE â· #general (36 messagesđ„):
MXFP8 quantization, MLSys 2026 FlashInfer-Bench Competition, Nvidia Triton Inference Server, GPU Mode GTC meetups, Madlab Liger-Kernel integration
- FlashInfer-Bench Competition at MLSys 2026: The MLSys 2026 FlashInfer-Bench Competition tasks participants with designing AI agents to write state-of-the-art LLM inference kernels on the latest NVIDIA Blackwell GPUs, competing against expert-written FlashInfer kernels, detailed at mlsys26.flashinfer.ai.
- Triton Inference Server discussion: A member inquired about discussing the Nvidia Triton inference server and a BLS script calling a vLLM backend model on Triton.
- Another member suggested using the general channel, noting it was the first time someone had asked about it and someone was trying to figure out how to pass the thinking budget parameter through to vLLM.
- GPU Mode Social at GTC: GPU Mode is planning a winner announcement for the nvfp4 competition and a social event around the time of GTC (March 16-19).
- The event will likely involve a social event, with past events including Beer Garden and the Outside market.
- Cornserve for Multimodal Models: A member shared their work on Cornserve, an efficient online serving system for Any-to-Any multimodal models, detailed in a paper Cornserve.
- Cornserve optimizes deployment plans for models with heterogeneous components like multimodal encoders, LLMs, and Diffusion Transformers (DiTs), improving throughput and reducing tail latency.
- RSS Feed Requested for GPU Mode News: A user requested an RSS feed for the GPU Mode news page (https://www.gpumode.com/v2/news) and offered to contribute.
- A member responded offering the siteâs GitHub repository (https://github.com/gpu-mode/kernelboard) for contributions and jokingly suggested testing if Claude could implement the feature.
GPU MODE â· #cuda (7 messages):
7 point 3D stencil computation, CUDA sample with a 25 pt stencil, Time zone bug, cutile fused moe kernel in the gym repo
- CUDA Sample gets Stencil Spotlight: A member is looking for tips on optimizing 7 point 3D stencil computation and another member suggested a CUDA sample with a 25 pt stencil that could be modified.
- Time Zone Bug Talk: Members were debugging a time zone bug.
- One member asked, what made you think of time zones and not something like Y2K and dtype overflows?
- Cutile Fused MoE Kernel Quest: A member is seeking an easy to integrate blackwell optimized kernel and asked if anyone has tried the cutile fused moe kernel in the gym repo.
GPU MODE â· #torch (3 messages):
BF16 Autocast, Dynamic Shapes, cu128 vs cu126, A100 issues
- BF16 Autocast throws errors with Dynamic Shapes: A member reported that bf16 autocast with dynamic shapes on torch 2.10 with cu128 throws errors on an A100 with cuda 13.
- The user noted that everything works fine on a cu126 wheel, but breaks on a cu128 wheel.
- Request for Issue Elaboration: A member asked for more details on the issue, specifically requesting the error message.
- The same member also requested clarification on any additional details available to assist with troubleshooting.
GPU MODE â· #announcements (2 messages):
CornServe, 2025 GPU MODE Recap, 2026 GPU MODE Plans, Kernel LLM Training, Hardware Programming Complexity
- Cornserve is Served Hot: GPU MODE went online with a member to discuss Cornserve: Easy, Fast and Scalable Multimodal AI (YouTube link).
- GPU MODE had wild success in 2025: 2025 was an incredible year for GPU MODE: 26K YouTube subs, 92 lectures, 24K Discord members, 3x $100K+ kernel comps, 400K KernelBot submissions, 3 events, and 10 active working groups!
- The community received shoutouts from role models like Soumith Chintala, Ian Buck, Tianqi Chen, Shotlo Douglas, Tri Dao and Lisa Su for project popcorn.
- GPU MODE unveils their 2026 plans: In 2026, GPU MODE is pushing further with training a Kernel LLM and using it to ship kernels in important repos like PyTorch and VLLM (gpumode.com/v2/news/gpumode-2026).
- The community is collaborating with Prime Intellect, Modal, and Lambda, focusing on de-slopifying LLM-generated kernels, post-training a kernel LLM model, end-to-end competitions, and from-scratch repos.
- Complex Hardware is becoming more Complex: Hardware is becoming more complex to program (X link), and the community has a responsibility for making it easier!
GPU MODE â· #cool-links (2 messages):
LeCun Startup, Event Based Model
- LeCun Launches New Startup: Logical Intelligence: Yann LeCun launched a new startup called Logical Intelligence, an Event Based Model (EBM).
- Unfortunately, no technical details were provided, only a link to the MLSys Conference.
- Event Based Model is shrouded in mystery: The new startup Logical Intelligence focuses on Event Based Models but provides no technical details.
- The website only contains marketing material, job openings, and a link to the MLSys Conference.
GPU MODE â· #job-postings (2 messages):
CUDA Kernel Optimization, Mindbeam AI Hiring
- Parsewave Seeks CUDA Kernel Optimization Engineers: Parsewave, partnering with frontier AI labs and AI infra providers, is seeking CUDA C/C++ kernel optimization engineers to benchmark internal models, requiring experience with Nsight Systems / Nsight Compute and CUDA intrinsics (Blackwell ideal, Hopper great too).
- Candidates should be able to explain optimization wins and propose benchmarks showing naive â optimized deltas; interested applicants can apply here.
- Mindbeam AI Recruiting Post Training and GPU Kernel MLEs: Mindbeam AI, a small team focused on accelerating training for foundation models, is hiring a
post training MLEandGPU Kernel MLE.- The company is fully remote and offers competitive pay, with interested candidates encouraged to DM for a referral; job openings are listed here.
GPU MODE â· #beginner (1 messages):
Roofline Models, Kernel Optimization, Performance Analysis
- Roofline Reads Readily: A member shared a diagram to aid in understanding roofline models for kernel optimization, suitable for sharing on LinkedIn and helpful for learners.
- The diagram visually explains how to interpret performance bottlenecks and optimize kernels based on hardware limitations.
- Kernel Knowledge Kurated: The shared diagram highlights the relationship between computational intensity and memory bandwidth in achieving optimal performance.
- It emphasizes that understanding these limits is vital for writing efficient GPU kernels and maximizing hardware utilization.
GPU MODE â· #pmpp-book (2 messages):
Thread Coarsening Clarification, Mentorship Opportunities
- Thread Coarsening Confusion Cleared: A member initially questioned the formula for
colStartin thread coarsening, specifically whether it should include an additionalTILE_WIDTHfactor, referencing page 139, chapter 6.3.- The confusion was resolved after realizing the text refers to the number of columns a thread block is responsible for, not the total number of elements.
- Asks for a Mentor for Hands-On Projects: A member who has completed the initial chapters of a book seeks a mentor to work on practical projects, providing their personal website for context.
- Theyâre looking for mentorship to complement their theoretical knowledge with hands-on experience.
GPU MODE â· #irl-meetup (2 messages):
MLSys Conference, Treehacks
- Attendee Enquires About MLSys Conference Experience: A member inquired about the MLSys conference experience, noting the 2026 conference will be in Bellevue, WA and they plan to volunteer since they attend Bellevue College.
- Member Seeks Companions for Treehacks: A member asked for DMs from anyone planning to attend Treehacks.
GPU MODE â· #rocm (1 messages):
kashimoo: Data centre GPUs are the focus for the AI space, not consumer GPUs though
GPU MODE â· #popcorn (13 messagesđ„):
MLSys 2026 FlashInfer-Bench competition, Weekly meeting link, Open to new contributors?, First public meeting on Feb 3
- FlashInfer-Bench Competition Announced for MLSys 2026!: The MLSys 2026 FlashInfer-Bench competition was announced and those interested in AI kernel generation are encouraged to participate.
- It was mentioned that more details are available in the new 2026 post.
- Link to Weekly Meeting requested: A member asked for the link to the weekly meeting, and another member provided this link.
- Inquiries about new contributors joining: A member asked if the channel is open to new contributors, finding it through this link on the working groups page.
- The member was welcomed and directed to the new 2026 post for information on where help is needed.
- First Public Meeting Scheduled!: The first team meeting is scheduled for Feb 3, and the organizer is checking on the meeting linkâs status.
- There may be voice channel in general.
GPU MODE â· #edge (1 messages):
Jetson, Torch, ONNX, TensorRT, trtexec profiling
- ONNX and TensorRT performance: A member looked into ONNX to TensorRT performance and suggested using trtexec to profile the engine layers.
- They stated that one can map TensorRT layers to ONNX ops from engine metadata but had no clue on going from ONNX ops to Torch.
- Torch Workflow Question: The discussion involved someone seeking advice, they also mentioned Jetson/Torch, ONNX, and TensorRT
- They looked into ONNX to TensorRT performance and suggested using trtexec to profile the engine layers.
GPU MODE â· #hardware (14 messagesđ„):
RTX 3060 12GB for ML/DL, FP4 acceleration on Blackwell, Consumer vs Data Center GPUs, DLSS papers, 5070ti or 4070 ti Super
- RTX 3060 Still Viable for GPU Learning?: Members discussed whether an RTX 3060 12GB is still a good option for learning GPUs and doing ML/DL work, given its relatively low price.
- The consensus was that itâs suitable for local learning setups, especially if acquired at a good price, but training will be slow, and support for features like FP4 on Blackwell will be missing; see Nvidiaâs Mistral integration example.
- Consumer Blackwell differs from Datacenter Blackwell: A member debated whether buying an expensive consumer GPU is worthwhile for writing kernels, considering the differences between consumer Blackwell (SM_120) and data center Blackwell (SM_100).
- Though core kernel skills transfer, staying current with architecture-specific optimizations is crucial for job market relevance.
- GPU Fundamentals can be learned on older GPUs: It was suggested that while newer architectures are important, general GPU fundamentals can be learned on older GPUs for fast iteration.
- The recommended progression is to build a project on Ampere, then tune it for Blackwell, and continue adapting to newer architectures.
- 5070ti or 4070ti Super better than 2x3060: A member with two RTX 3060 cards suggested that 12GB VRAM is limiting, advocating for a single 5070ti or 4070 ti Super instead.
- They asked about papers available which explain DLSS.
GPU MODE â· #factorio-learning-env (1 messages):
Factorio blueprint generation
- AI Engineer Discovers Factorio Blueprint Generation Project: An AI engineer researching ways to generate Factorio blueprint JSON code from instructions found the project impressive and stumbled upon it during their research.
- Potential of AI in Automating Factorio Blueprint Creation: The discussion highlights the potential of using AI models to automate the generation of Factorio blueprints, specifically focusing on creating JSON code from user instructions.
GPU MODE â· #cutlass (1 messages):
Graphical Layout Calculus, Tuple Morphisms, Mutual Refinement, Layout Composition
- Laying out Layout Compositions Graphically: A member shared a worked example of computing the composition of two layouts by hand using the graphical layout calculus.
- The steps involve converting layouts to tuple morphisms, finding mutual refinements, pulling back, pushing forward, composing, and writing the result as a layout using prefix products.
- Mapping to Morphisms for Layout Mastery: The initial step involves converting tractable layouts to tuple morphisms
m_Aandm_B.- This transformation allows for algebraic manipulation and composition of layouts using morphism operations.
- Refining Relations Between Layouts: The worked example emphasizes the importance of finding a mutual refinement of the two tuple morphisms.
- This step ensures compatibility and consistency when composing the layouts, akin to finding a common ground between two different structures.
- Pulling Back for Precise Layouting: The process includes pulling back the mutual refinement along
m_Ato obtain\hat{m}_A.- This operation adjusts the refinement to be compatible with the structure of layout A, allowing for seamless integration during composition.
- Pushing Forward for Polished Placement: The example also involves pushing forward the mutual refinement along
m_Bto get\hat{m}_B.- This operation ensures that the refinement aligns with the structure of layout B, further facilitating smooth composition and consistent layout behavior.
GPU MODE â· #teenygrad (1 messages):
Rust on M3, CPU and GPU kernels
- Rust benchmarks on M3: Initial Rust benchmarks on M3 show roughly 5% peak performance, with rustcâs loop reordering identified as a factor.
- Focus on CPU and GPU kernels: The next steps involve working on kernels for both CPU and GPU, focusing on performance improvements.
GPU MODE â· #nvidia-competition (51 messagesđ„):
group_gemm issues, benchmark leaderboard gap, B200 Physical Resonance, Stream Error During Submission, MLSys Contest Tracks
- Group GEMM Init Overshoots FP16 Range: The
group_gemmproblemâs old tensor init logic overshoots FP16 range, causing INF values; a fix similar todual_gemminit is proposed, referencing this PR.- Some INF values are acceptable, but the team is open to changes; a PR was opened to address the issue (PR 96).
- Benchmark Leaderboard Gap: A significant discrepancy between benchmark results and leaderboard scores was observed, raising concerns about inconsistencies.
- The description stated that
Kis divisible by 256, but there areK=128andK=384in the test cases, and it was suggested to remove or modify these cases.
- The description stated that
- Veitner Blogs on Grouped Blockscaled GEMM for B200 GPUs: Simon Veitner published a blog post explaining grouped blockscaled GEMM for B200 GPUs in a top-down manner and the setup of MMA and TMA, tile scheduler, and other parts, available on bearblog.dev and LinkedIn.
- The blog aims to explain the parts that are different from the usual persistent blockscaled dense gemm approach on B200.
- Stream Error Plagues Submissions: Users encountered a âstreamâ error during submission, traced to the presence of the word âstreamâ in the code, even within comments.
- Removing the word âstreamâ from the code (including comments) resolved the submission issue.
- Task Config Differences Examined: Differences between test and benchmark configurations in
task.ymlwere noted, where benchmark configs have the same N and K for all Aâs and Bâs in a group, unlike test configs.- The team clarified that the test is for function verification where m/n/k in different groups can be different, and the performance test comes from real use cases where M (experts) are different and N/K are the same.
GPU MODE â· #career-advice (8 messagesđ„):
Learning GPUs, TinyML, Embedded ML, Physical AIs
- Newcomer Seeks Guidance on GPU Learning: A software engineer is looking for book suggestions to understand GPUs, optimization, and tuning for performance, aiming to transition into TinyML/embedded ML or physical AIs.
- They prefer learning through books due to struggling with lengthy videos and have a basic understanding of ML but lack hardware knowledge.
- Background Doesnât Matter; Interest Does: A member shared that they got their GPU performance position with a background mainly in math-physics & formal methods.
- They state that US companies have a lot of capital invested in AI currently.
- Search Specs Online: It was suggested that interviewers generally allow candidates to search online for specific specs during interviews, recognizing that specific details are easily searchable.
- A member mentioned that whether the questions are âgooglableâ is highly dependent on the company and the interview topic.
GPU MODE â· #flashinfer (53 messagesđ„):
Team Merging, Multiple Track Participation, Registration Confirmation, Team Formation, Kernel Type (Single/Multi-node)
- Team Merging Mania: Participants inquired about team merging before the registration deadline, to which the organizers responded that it is allowed, with a request to be notified of the changes.
- The organizers also set up an automated registration confirmation email in response to requests.
- Track Shifting Shenanigans: Contestants asked if participation in multiple tracks was possible, the organizers confirmed it but noted only one GPU would be awarded even if a team ranked highly in multiple tracks.
- The discussion clarified that teams can shift tracks later to focus on the most promising one.
- Kernel Confidentiality Conundrum: Participants raised questions about whether submissions would be made public or kept private after the competition.
- Organizers clarified that the final code needs to be made public for award consideration, but the development process can remain private.
- Newbie Navigation Notes: Beginners inquired about the best track selection for newcomers in the NVIDIA MLSys contest.
- The recommendation was to deploy the smallest possible model using the FlashInfer API only to become comfortable with the codebase, while avoiding unstable platforms like B200.
- Agentâs Secrets Safe (Mostly): Clarification was sought regarding the requirement for open-sourcing agent solutions in the FlashInfer AI kernel competition.
- The organizers confirmed that while the agent code and tech report will be reviewed, only the final code needs to be open-sourced to ensure itâs not a hand-crafted kernel.
Eleuther â· #general (57 messagesđ„đ„):
ROCm performance for ML, Image classification services, DistinctionBench and language models, Human-in-the-loop workflows with LLMs, In-context learning and weight updates
- ROCmâs Rocky Road to ML Renaissance: Users discussed the performance of ROCm for accelerated ML, noting it has made strides but can be challenging due to primary support for Nvidia.
- The experience was described as âbatteries not includedâ due to potential driver problems and long lead times.
- DistinctionBench: Training Target or Contamination Resistant?: Discussion on the paper Between Circuits and Chomsky: Pre-pretraining on Formal Languages Imparts Linguistic Biases considered DistinctionBench to give interesting transfer to language models.
- One member joked, âall good evals are training targets ;)â, but noted DistinctionBench is âvery contamination resistantâ due to endless representational variants.
- ICL Signals: Are Weights Updating?: A member asked about papers on âusing the signals from in context learning to update the weights as a form of continual learning,â and two relevant papers were shared: chen24r.html and arxiv.org.
- The conversation also pointed to saving inference costs via pushing stuff into the parametric knowledge, reminiscent of âstate tuning in linear attention variantsâ and âcartridgesâ from this summer (https://arxiv.org/abs/2506.06266).
- Attention Arrived Before the Transformer: Attention mechanisms existed on top of RNNs in 2014-2015, but it took two years to introduce the transformer because people werenât convinced about attention.
- It was suggested that there were fewer people working in the field back then, and Kaggle results really helped it take off.
- Forbes Article fails to meet contribution standards: A member posted a Forbes article with commentary, but another member responded that âPretending that a Forbes article that is a copy/paste of popular questions of Quora is a representation of what leading research questions in AI are does not meet our contribution standards.â
- The member then added the heuristic: âIs this a conversation that two AI researchers might haveâ is a good heuristic.
Eleuther â· #research (121 messagesđ„đ„):
Weak Baselines, RWKV Architecture, One-Step Generative Modeling, Hybrid Architectures, Deduplicated Pretraining Data
- Weak Baselines Bashed: Members debated the validity of âweak baselinesâ as a complaint against research, arguing that even beating ChatGPT doesnât guarantee significance without a strong baseline.
- It was emphasized that experiments should start with robust baselines to avoid mistaking noise for genuine improvements, suggesting modded nanogpt as a good starting point for language tasks, with one member recommending replicating this paper.
- RWKV Revamp Rumors: A member shared their work on modifying the RWKV architecture, but others cautioned about parameter count and training methods, recommending training on tokens instead of bytes.
- It was suggested that the modifications should be tested on recent RWKV codebases with attention baselines, and renting a 4090 or 5070ti was recommended due to CPU limitations, plus that the approach could be related to FFN-Bilinear.
- Gate-Free FFN Flounders?: Experimentation with gate-free FFNs showed a 4.3% parameter reduction but only a 0.5% improvement compared to a sigmoid gate, raising questions about the efficiency of added parameters in MLPs.
- One member suggested that gating might help fix GQA, and that Lora gate params or near-MiSS formulation (expansion from a subregion of the hidden state) might improve results without the significant parameter count increase. Another shared that in their work, taking the last 12 dims of the residual dim and using it for the attn gate seemed to do pretty well.
- Generative Modeling Gauntlet: With a surge of papers on one-step generative modeling, members discussed which methods are promising, noting the difficulty in comparing benchmarks and separating noise from valuable advancements.
- One member advocated for a theoretical understanding to tier methods and avoid impractical options, while another agreed math âsoundnessâ plays a big role.
- Symbolic Sanity Checks Save Sanity: The potential of hybrid architectures combining LLMs with symbolic/deductive layers for hallucination reduction was explored.
- While checking logical consistency is relatively easy for math, code, and simple facts as shown in this paper, it remains challenging for other types of consistency as shown in this paper.
Eleuther â· #interpretability-general (4 messages):
Model Weights Comparison, Free Compute Resources, Automating Circuit Discovery, OLMo 3 models
- OLMo 3 Models Suit Needs: A user suggested that OLMo 3 may suit another userâs needs, as it has separate SFT and thinking models.
- They suspected itâs close enough to warrant a preliminary study of model weights.
- Compute Resources for Model Finetuning Sought: A user is working on a project to compare model weights of two variants of the same model, one finetuned to follow instructions and the other to solve reasoning tasks, and asked for free compute resources.
- The user sought resources to fine-tune a small model on Colab and is open to compute sharing.
- Automating Circuit Discovery Papers: A user requested a list of papers related to automating circuit discovery for behavior, such as for IOI and induction.
- The user invited others to share interesting papers they find on the topic as well.
Eleuther â· #multimodal-general (1 messages):
aeros93: https://fixupx.com/havenfeng/status/2014765400563781777?s=46
Nous Research AI â· #general (128 messagesđ„đ„):
Self-Replication Benchmark for Agentic AI, LLM Worms, MoE run, OpenAI Business Model, Local Code Models and Raytracer Test
- Self-Replication Benchmark Considerations: A member is considering a self-replication benchmark for agentic AI, pondering the appropriate goal and whether the agent should download itself or retrain from scratch.
- They suggested that adopting to a target machine or even designing one could be fun, as opposed to simply using existing transformer libraries.
- LLM Worms: âHey make more of youâ: One member jokingly suggested an LLM worm benchmark where an LLM is prompted with âhey make more of youâ and given the tools to replicate itself, whether by downloading a copy or writing a script that downloads a script and uses an API key.
- Another pointed out the importance of considering resource constraints like VRAM to make the challenge more interesting.
- 1stlanikâs MoE Run Dashboard âFailed to Fetchâ error: A member reported a âFailed to fetchâ error in the dashboard while checking the progress of an active MoE run (moe-10b-a1b-8k-wsd-lr3e4-1t).
- Another member suggested checking back in a few hours.
- OpenAI Pricing Model discussed: Members discussed how OpenAI may be raising prices by offering services that barely work on the lower tier subscriptions.
- One stated that âAI companies could raise prices a lot right now and people would payâ but another countered that many companies offer the same for cheaper. It was also mentioned that Anthropic has a 40% gross margin.
- Raytracer Test Proves Difficult for Local Models: A member noted the difficulty of local code models (runnable on a 5090) to pass a raytracer test from cpldcpu/llmbenchmark, even observing that recent models on lmarena are failing it now.
- They find that the smaller models consistently mess up the vector class.
Nous Research AI â· #ask-about-llms (2 messages):
LLM pre-training for domain-specific tasks, Effectiveness of continued pre-training
- Pre-training LLMs for Domain-Specific Tasks: A member asked about the effectiveness of continued pre-training a foundational LLM for a domain-specific task like law or healthcare using OLMO-7B and the ZINC20 dataset.
- Another member, an LLM researcher, suggested it generally improves performance but is task-dependent, noting that training with task-related inputs/outputs may outperform more general continued pre-training (cpt).
- Expanding Multilingual Capabilities via CPT: The researcher noted that continued pre-training (cpt) can expand multilingual capabilities, and fine-tuning on translation data strengthens task performance.
- This comment was specifically made in response to the general question regarding continued pre-training (cpt).
Nous Research AI â· #interesting-links (1 messages):
Semantica, Knowledge Graphs, Ontologies, LLM reasoning
- Semantica: Open-Source Semantic Infrastructure: A member introduced Semantica, an open-source project focused on building semantic infrastructure for domain-grounded AI, including knowledge graphs, ontologies, and reasoning layers.
- They are actively seeking contributors for ontology & schema design, knowledge graph modeling, LLM + symbolic / rule-based reasoning, data ingestion & semantic pipelines, and documentation.
- Semantica: Contribution Opportunities: The project is looking for contributions in various areas, including ontology & schema design and knowledge graph modeling.
- Contributions donât have to be big, and issues, design discussions, feedback, or small PRs are all welcome.
Yannick Kilcher â· #general (105 messagesđ„đ„):
EBM vs Classical FF, EBM and Shannon Entropy, LLM pre-training, MCMC sampling issues, Zero-knowledge proofs
- EBMs vs. Classical Feedforward: Is There a Clear Winner?: The discussion starts by questioning if Energy-Based Models (EBMs) are inherently superior to classical feedforward networks, especially concerning Shannon entropy or Kolmogorov complexity.
- One member suggests that validation is easier than generation in EBMs, relating it to computational complexity theory (P vs NP), while emphasizing the need for a well-defined loss landscape for EBM optimization to work effectively.
- LLM Pre-training: Domain-Specific vs. Foundational: A member asked about the effectiveness of continued pre-training a foundational LLM (specifically OLMO-7B) for a domain-specific task like cheminformatics using the ZINC20 dataset.
- The goal is to compare results against a domain-specific transformer model, but no specific answers or resources were provided in the discussion.
- MCMCâs Messy Mode-Switching Mishaps: A member inquires about the adequacy of a paper in illustrating MCMC sampling issues, particularly how badly it sucks.
- One member argues that MCMC tries to emulate flow models due to the latterâs superiority, while EBMs, contrarily, attempt to make NNs more like MCMC, which they deem misguided, as they elaborate that HMC has issues traversing between spatially separated modes, making it horrible as the dimension increases.
- ZKPs: More than Just Crypto Signing?: A member discusses using zero-knowledge proofs (ZKPs) for verifying encrypted network traffic and matrix multiplications, pointing to a Gemini correspondence for a matrix low knowledge proof.
- They propose a use case in zero-knowledge âmade by humansâ proofs, but another member is skeptical about the practicality of ZKPs, suggesting breaking the encryption might be cheaper, while the initial member claims the opposite, stating ZKPs are theoretically even more efficient than the feedforward.
- NN Parameterization: A Trio of Techniques: A member questions the advantage of parameterizing the score over parameterizing log p(x), and they respond that we can just to monte carlo estimate on denoising matching term instead of both denoising matching term + reconstruction term, so we have less variance?.
- Itâs clarified that you can parameterize the distribution directly, the log-likelihood (MLE and EBMs), or the score (flow models), and that Optimal Transport (OT) is distinct, affecting what you do with a distribution rather than how you learn or parameterize it.
Yannick Kilcher â· #ml-news (10 messagesđ„):
LLMs cyber capabilities, LLM companies internal vulnerabilities, Exploit events, Github repo security
- LLMs cyber skills questioned: A member questioned whether LLMs could develop strong cyber capabilities, referencing a GPTZero article.
- Another member doubted LLM companiesâ ability to address internal vulnerabilities, suggesting they fix those before pursuing cyber skills, also citing a ScienceAlert article and a tweet.
- Upcoming Exploit Events?: A member predicted potential large exploit events and warned about LLMsâ access to sensitive resources.
- They advised using GitHub deploy keys in isolated environments when coding with a GitHub repo to limit potential damage.
- No Access Granted!: One member humorously declared that they would not grant LLMs access to anything.
- Another member countered this sentiment by calling it robo-phobicitâs and dubbing it survival instincts.
tinygrad (George Hotz) â· #general (74 messagesđ„đ„):
Luminal flash attention, Metal performance with textures vs buffers, Tenstorrent backend passing ops tests, Tinygrad intended use and training LLMs, Anthropic VLIW challenge PR
- Luminal finds Flash Attention via Bruteforce: Luminal is claiming to find flash attention using bruteforce on an egraph, taking hours to find, and they explicitly added
exp(x - new_max) = exp(x - old_max) Ă exp(old_max - new_max)as a rewrite rule.- The poster reproduced the graphviz shown in the presentations from commit
0bd3b80c, noting that their minimal set of rewrite rules could transform a naive attention kernel graph into the known flash attention kernel graph in 52s on a 9800x3d.
- The poster reproduced the graphviz shown in the presentations from commit
- Metal: Textures Beat Buffers for Blurring: Profiling access speed on Metal using
Tensorwith size 512/1024/2048/8192 images as input for a 3/5/7 sized blur kernel showed textures outperforming buffers.- It might be worth throwing in a branching condition depending on the size of the buffer input, tests results are attached.
- Tenstorrent backend passes ops tests: The Tenstorrent backend is passing all ops tests on wormhole or blackhole and there is a $1k bounty for this milestone.
- Someone asked if the bounty requires all test ops test passing on testorrent hardware.
- Anthropic VLIW challenge PR submitted: A member submitted a PR for the Anthropic VLIW challenge, hitting 1258 cycles.
- The submitter expressed uncertainty about generalizing the code, particularly the batch staggering, which might be useful for other VLIW targets, and also apologized for a lazy lookover that introduced AI-generated changes.
- Tinygrad isnât for normies: A user asked about the intended use of tinygrad, specifically regarding porting existing models and training LLMs on multiple GPUs, and was told by George Hotz to ask claude about this.
- Another user expressed frustration at being told to use Claude for documentation and said tinygrad is not for me or most devs then, to which George replied iâm not selling to anyone, tinygrad is free and that adoption is not a target.
Moonshot AI (Kimi K-2) â· #general-chat (65 messagesđ„đ„):
Slides Generation Issues, Login Issues, Rate Limits, Chinese AI Labs Innovation, Multimodal Models Comparison (Kimi vs. Ernie 5.0 vs. GLM 4.6V)
- Slide Generation Issues Plague Users: Some users reported issues generating slides, even with visual and adaptive options, with one user reporting that the issues have persisted since the previous day, linking to a video showcasing the issue.
- The user experiencing the issue suggested that internal rate limits may be the cause, and they are now able to generate slides again, suggesting the issue was temporary.
- Kimiâs Chinese Labs win big praises: One user lauded Chinese AI labs, including Kimi, for their innovation and performance compared to other models like Gemini, citing Kimiâs human-like responses and impressive memory capabilities.
- The user expressed a desire for Kimi to incorporate multimodal capabilities similar to Minimax, specifically vision and audio analysis for video transcription, along with tool integration and workspace features.
- Kimi K2.5 Silently Sneaking into the spotlight?: Users are noticing Kimi models self-reporting themselves as K2.5, despite no official announcements or UI changes indicating a new version.
- Some speculate this could be related to internal testing or improvements to the slide tool, potentially involving visual understanding of images, but others claim that they checked and there arenât any major improvements.
- API Login struggles: A user reported difficulty logging into the Kimi/Moonshot platform to generate new API keys, particularly with a non-Gmail account, and was directed to contact the support email.
- The user later clarified that the issue was not rate limits but simply forgetting the login procedure for the backend.
- Kimi is adding Memory features: A user highlighted that Kimiâs app now includes memory features, enabling customization, which enhances the overall user experience.
- The memory and customization options make it a favorite chatbot.
Modular (Mojo đ„) â· #general (52 messagesđ„):
Mojo in production, Mojo startup lag, VS Code debugging, CPU vs GPU kernels
- Mojo Sees Production Use in HPC: A member is deploying a Mojo project for parameter refinement in cryo-electron microscopy and seeing a 5-10x speedup over the old C++ code.
- The biggest win was pulling of an AoSoA layout for one bit, made super easy by Mojoâs list of structs with SIMD members.
- Mojoâs Cold Executable Starts Slowly đ: A member reported a 200ms startup lag for even simple Mojo scripts, which they tracked down to a Gatekeeper issue on macOS scanning untrusted binaries, where subsequent runs were much faster.
- They found a 50ms launch overhead on a cold executable after rebooting, which they considered acceptable.
- VS Code Debugging Still Has Issues đ: A member reported debugging with the VS Code extension fails due to a âFunction not implementedâ error on an air-gapped machine using
.condafiles from max-nightly.- A Modular employee mentioned debugging in the extension should be working on Mac and Linux with environments set up using Pixi as described in the Quickstart guide.
- GPU Kernel Portability is a pipe dream: A member noted that standard CPU kernels under-utilize the GPU, requiring specialized code, and another suggested GPUs could be treated as wider SIMD units to simplify programming.
- He suggested using a number of warps instead of number of threads to solve this issue.
Modular (Mojo đ„) â· #mojo (12 messagesđ„):
Mojo 1.0 Release,deffunctions, Pointer.mojo, out self
deffunctions decision pending for Mojo 1.0: With Mojo 1.0 release in a few months, the decision on includingdeffunctions is still pending, with a member pinging Denis for a response on GitHub issue #5830.- Currently, thereâs no committed date for Mojo 1.0 other than âin 2026â.
out selfargument in Pointer.mojo discussed: A member noted that inPointer.mojo, the__init__functionâs first argument is notself, but another Pointer, questioning if this deviates from the standard.- Another member explained that
outarguments only serve as output and do not affect the call signature, so the position ofout selfdoesnât matter technically, but convention suggests putting it first in__init__.
- Another member explained that
- Argument order matters for parameter inference: A member explained that
out selfmust be the second argument in this case because it depends onotherfor one of its parameters -ImmutOrigin(other.origin).- Another member added that the argument order is relevant for parameter inference.
Manus.im Discord â· #general (50 messagesđ„):
Manus billing issues, Manus free credit codes, AI Engineer introductions, AI + Healthcare systems, AI Agent development
- User Demands Resolution for Unauthorized Billing: A user reports being charged $400 for an annual plan after selecting monthly billing and threatens complaints to FTC, BBB, Attorney General, and Meta due to unauthorized billing, refused refunds, and unresponsive support.
- Another user recommends filing a chargeback.
- Manus Free Credit Code Revealed!: One user shared a redeem code
Havefunwhich gives 1000 credits.- Another user asked where to find these codes, and was directed to the Exchange button.
- AI Engineers Introduce Their Healthcare Skills: An AI + Full Stack Engineer introduced expertise in building production-grade AI systems for Healthcare, including clinical NLP, medical imaging, and patient-facing AI applications.
- This engineer also builds LLM systems, autonomous agents, workflow automation, and multimodal AI (text · voice · vision) and included a list of their core skills.
- AI Agent Developer Focuses on Production Systems: An AI Agent Developer highlighted their focus on building AI agents for real production use, rather than demos, and is available for collabs and audits.
- The developer specializes in customer support, sales agents, workflow/ops agents, and autonomous booking/scheduling agents.
- User Seeks âShare with a friendâ on Mobile: A user asked where the âShare with a friendâ option is located.
- Another user replied that on a computer, itâs at the bottom of the left sidebar but, offered help for the mobile version.
DSPy â· #show-and-tell (2 messages):
DevinReview, DSPy RLM, AsyncReview, RLM Skills, Claude Code
- AsyncFuncAI open sources DevinReview: A member has open sourced a version of DevinReview using the DSPy RLM framework, available on GitHub.
- The new release has been named AsyncReview.
- Add RLM Skills to Claude Code or Opencode: A member shared the idea to add RLM as skills to Claude Code or Opencode.
- The member also shared an npm package called rlm-skills.
DSPy â· #general (46 messagesđ„):
RLM Prompt Tuning, DSPy Optimizer for Multi-Step Modules, JSON Adapter Customization with GEPA, Leveraging DSPy for Typescript Agent Optimization, DSPy via AG-UI
- RLM Prompts Demand Tuning: Users discussed tuning the RLM prompt itself, citing that reasoning can be lacking in some models and suggested techniques for improving the RLM prompts.
- DSPy Optimizer to Inspect the Trace: When using DSPy optimizers for modules with many intermediate steps, it was suggested that the optimizer will automatically inspect the trace, so users only need to focus on the desired output.
- One user recommended preparing a good set of training data with example documents and a measurement that rejects wrong answers when the RLM answers prematurely.
- JSON Adapters get GEPA Treatment: A user wants to use GEPA to work on the text that the JSONadapter puts in the system prompt, given that the tokens are not always needed for the response to have the appropriate output form.
- They believe theyâll need to make a custom GEPA adapter, as the DSPy one doesnât affect the adapters.
- TypeScript Agents Seek DSPy Optimization: One user is looking to leverage DSPy for optimizing prompts of agents written in Typescript and asked if the architecture is currently supported or feasible in practice.
- AG-UI DSPy Adapter Streams Events: A user inquired about interest in exposing DSPy via AG-UI, highlighting its benefits for frontend/backend communication and avoiding the need for API endpoints and state management.
- The user has a working version that streams events, including reasoning traces, tool calls, and streamed LLM responses to the frontend.
aider (Paul Gauthier) â· #general (7 messages):
Aider + Claude Code workflow, Aider and Devstral Small 2 model
- Aider pairs well with Claude Code: A user noted that aider is fast, making it a perfect pair for Claude code to punch through bug walls with agentic efficiency.
- The user finds aider useful for working out which files need to be in context, managing the context, and the search and replace coder minimizes llm token outputs.
- Devstral Small 2 works excellently with Aider: A user reported excellent success using Aider with Devstral Small 2, a new 24B dense model.
- At Q4_K_M, it fits in a 3090 with enough room for nearly 50k context, and the search/replace blocks it generates are perfect 80-90% of the time and recovers in a single attempt when it fails.
MCP Contributors (Official) â· #general (2 messages):
Discord voice channels, Contributor-related chat
- Discordâs New Voice Channel Experiment!: The team is experimenting with new Discord voice channels, named
conference-room-aandconference-room-b, available in the channels list.- These channels are intended for ad-hoc contributor chats to quickly resolve issues, especially when a long async text thread is ineffective.
- Moderation and Access Rights Reminder!: Specific members have permissions to mute people in these channels, while others should ensure they have the necessary access rights.
- There is a reminder that the access rights will be changing in five days.