Chromium is all you need.
AI News for 10/20/2025-10/21/2025. We checked 12 subreddits, 544 Twitters and 23 Discords (198 channels, and 7709 messages) for you. Estimated reading time saved (at 200wpm): 564 minutes. Our new website is now up with full metadata search and beautiful vibe coded presentation of all past issues. See https://news.smol.ai/ for the full news breakdowns and give us feedback on @smol_ai!
As leaked in July (and earlier), OpenAI finally launched their Chromium fork AI browser, Atlas (MacOS only for now but other platforms coming - download/website here):
The integration is very polished and impressive, as you can see in the second half of the livestream. By bringing Agent mode into Atlas, OpenAI is not just matching what Gemini in Chrome already has, but going the next obvious step past it by reviving Operator and putting it in the local browser instead of remote, so that it can use your logins.
The vibes are positive, but not entirely so:
Your move, Google.
AI Twitter Recap
OpenAIâs ChatGPT Atlas Browser Launch
- Atlas ships with Agent Mode and âbrowser memoryâ: OpenAI unveiled an AI-first browser for macOS with ChatGPT embedded system-wide, optional page/context memory, and a preview âAgent modeâ that can act on webpages (including logged-in sites with permission). macOS is rolling out now; Windows, iOS, and Android âcoming soon.â See launch posts from @OpenAI, Agent mode details, and product notes. PMs highlighted use-cases and UX intent via @kevinweil, @bengoodger, and @fidjissimo. An incognito-style toggle for memory is present (@omarsar0).
- Early reactions: The âbrowser is the new OSâ framing landed (@Yuchenj_UW, @nickaturley), but reliability and privacy trade-offs surfaced immediately. One head-to-head against Perplexityâs Comet showed Atlas completing a tedious grades-tracking task more robustly (context handling, faster actions, and âhuman-likeâ exploration) (@raizamrtn). Others called Agent mode âslopâ for now and raised data access concerns (@Yuchenj_UW, privacy). Launch traffic briefly overwhelmed services (@TheTuringPost).
LangChainâs $125M Series B and v1.0 Agent Engineering Stack
- Funding + product milestone: LangChain raised a $125M Series B led by IVP with participation from CapitalG, Sapphire, Sequoia, Benchmark, and others, valuing the company at $1.25B. Alongside, it released 1.0 versions of LangChain and LangGraph, a LangSmith insights agent, and a no-code agent builder (@LangChainAI, @hwchase17, IVP note). The team emphasized a controlled, production-first agent runtime and observability, with a new createAgent abstraction + middleware in LangChainJS (@bromann, release notes). Usage claims: 85M+ OSS downloads/month and ~35% of the Fortune 500 using the stack (@veryboldbagel, @amadaecheverria).
- Ecosystem fit: vLLM added MoE LoRA expert finetuning support (@casper_hansen_) and credited an external analysis as impetus (@corbtt). Multiple teams highlighted production usage of LangGraph/LangSmith for agent reliability and evals (@Hacubu, @jhhayashi).
Vision Tokens, OCR, and New VLMs: DeepSeek-OCR, Glyph, Qwen3-VL, Chandra OCR
- DeepSeek-OCR (text-as-image) sparks debate: The paper reports large long-context compression by rendering text as images and decoding via a vision encoder + MoE decoder. Commentary ranges from enthusiastic technical breakdowns (97% reconstruction precision with ~10x fewer âvisualâ tokens; high-res convolutional compressor) (@rasbt) to sharp critiques on missed prior art (pixels-for-language and visual token compression lines) (@awinyimgprocess, @NielsRogge). Others argue the core takeaway is inefficiency in current embedding/token usage, not image superiority per se (@Kangwook_Lee).
- Zhipuâs âGlyphâ-like direction and KV via vision tokens: Several noted Zhipu releasing a contemporaneous vision-token compression approach (âGlyphâ), with claims of 3â4x context compression and infilling cost reductions without quality drop on long-context QA/sum (@arankomatsuzaki, context). Details remain sparse; watch for BLT-like extensions to push decoding efficiency further.
- Qwen3-VL-2B/32B: Alibaba released dense 2B and 32B VLMs, including FP8 variants and âThinkingâ/Instruct types, claiming strong wins vs GPTâ5 mini and Claude Sonnet 4 across STEM, VQA, OCR, video, agent tasks; the 32B aims to match much larger models on OSWorld with high memory efficiency (@Alibaba_Qwen). Demos landed on HF quickly (@_akhaliq).
- Open-source OCR: Chandra OCR launched with full layout extraction, image/diagram captions, handwriting, and table support; works with Transformers/vLLM (@VikParuchuri).
Training/Serving Stack Updates: PyTorch, vLLM, FlashInfer, Providers
- Meta PyTorch drops new libraries: torchforge (scalable RL training), OpenEnv (agentic environments), and torchcomms, plus momentum around Monarch and TorchTitan within a âfuture-of-trainingâ map (pretrainâpost-trainâinference) (@eliebakouch, stack summary, Monarch).
- vLLM and memory: kvcached enables serving multiple models sharing unused KV cache blocks on the same GPU (@vllm_project); the project is featured at PyTorch Conference (@vllm_project).
- FlashInfer-Bench: new âself-improvingâ benchmarking workflow to standardize LLM serving kernel signatures and auto-surface fastest kernels for day-0 integration in FlashInfer/SGLang/vLLM (@shanli_xing).
- Provider benchmarks for GLMâ4.6 (Reasoning): Baseten led output speed (104 tok/s) and fastest time-to-first-answer-token; pricing across providers clustered near $0.6/M input, ~$2/M output; all support 200k context and tool calling (@ArtificialAnlys).
Research, Evals, and Methods
- Continual learning via memory layers: Sparsely finetuned memory layers enable targeted updates with minimal forgetting compared to full finetune/LoRA (â11% vs â89%/â71% on fact tasks), proposing a practical route to incremental model updates (@realJessyLin, blog).
- Mechanistic interp at scale: Anthropic analyzed Claude 3.5 Haiku on a âperceptualâ task, revealing clean geometric transformations and distributed attention algorithms; community notes it as among the deepest behaviors understood mechanistically to date (@wesg52, @NeelNanda5).
- Prompt optimization > RL for compound systems? GEPA uses reflective prompt evolution with Pareto selection to beat GRPO on HotpotQA, IFBench, Hover, PUPA, reducing rollout needs via natural language self-critique (@gneubig, paper/code, summary).
- Evals in the wild: SWEâBench Pro leaderboard update shows top models now >40% pass rate, with Claude 4.5 Sonnet leading (@scale_AI).
- Self-play caveats for LLMs: Why self-play shines in twoâplayer zeroâsum settings (minimax) but is tricky in realâworld domains (reward shaping, equilibria untethered from human utility) (@polynoamial).
Developer Tooling and Apps
- Google AI Studio âAI-first codingâ: revamped build mode integrates multi-capability scaffolding (âIâm Feeling Luckyâ), targeting faster promptâproduction iteration for Gemini apps (@OfficialLoganK, demo, @GoogleAIStudio).
- Runway: announced self-serve model fine-tuning and a node-based Workflows system to chain models/modalities/intermediate steps for production creative pipelines (@runwayml, Workflows).
- Together AI: video and image generation models (e.g., Sora 2, Veo 3) now accessible through the same APIs used for text inference (@togethercompute).
- LlamaIndex: llamactl CLI for local LlamaAgents development/deployments; turnkey document agents template and private-preview hosting for doc-centric workflows (@llama_index, @jerryjliu0).
Top tweets (by engagement)
- âhahahaha the bed sends 16gb of data a month oh godâ â IoT reliability/telemetry facepalm during the AWS outage (@internetofshit).
- âMeet our new browserâChatGPT Atlas. Available today on macOSâ (@OpenAI); âMake room in your dockâ (@OpenAI).
- Karpathy on synthetic identity/personality tuning for nanochat via diverse synthetic dialogs (@karpathy).
- Qwen Deep Research upgrade: report + live webpage + podcast auto-generation with Qwen3 stack (@Alibaba_Qwen).
- Airbnb CEO: Qwen is âvery good, fast and cheap,â often preferred in production over âlatestâ OpenAI models due to cost/latency (@natolambert).
AI Reddit Recap
/r/LocalLlama + /r/localLLM Recap
1. Qwen3-VL Model Performance Comparison
- Qwen3-VL-2B and Qwen3-VL-32B Released (Activity: 626): The image provides a detailed comparison of the performance metrics for the newly released Qwen3-VL-2B and Qwen3-VL-32B models against other models like Qwen3-VL-4B, Qwen3-VL-8B, and Qwen2.5-VL-7B. The table highlights the modelsâ performance across various tasks such as STEM & Puzzle, General VQA, and Text Recognition. Notably, the Qwen3-VL-32B model demonstrates superior performance, achieving higher scores in most categories, which are marked in red to indicate their significance. This suggests that the Qwen3-VL-32B model is particularly effective in these tasks, outperforming its predecessors and other variants. One comment humorously suggests that the release of the 32B model should satisfy those requesting it, indicating anticipation and demand for this model size.
- The release of Qwen3-VL-2B and Qwen3-VL-32B models marks a significant advancement, with the new models reportedly outperforming the previous 2.5-VL 72B model despite being less than half its size. This suggests substantial improvements in model efficiency and performance, likely due to architectural optimizations or enhanced training techniques.
- A comparison image provided by a user highlights the performance differences between Qwen3-VL-2B and Qwen3-32B, indicating that the newer models may offer superior capabilities in text processing tasks. This could be of particular interest to those evaluating model performance for specific applications.
- Benchmarks shared in the discussion suggest that the Qwen3-VL models excel in âthinkingâ tasks, which may refer to complex reasoning or problem-solving capabilities. This positions the models as strong candidates for applications requiring advanced cognitive processing.
- DeepSeek-OCR AI can scan an entire microfiche sheet and not just cells and retain 100% of the data in seconds⊠(Activity: 405): DeepSeek-OCR AI claims to scan entire microfiche sheets, not just individual cells, and retain
100%of the data in seconds, as per Brian Roemmeleâs post. The tool reportedly offers a comprehensive understanding of text and complex drawings, potentially revolutionizing offline data curation. However, the post lacks detailed technical validation or benchmarks to substantiate these claims. Commenters express skepticism about the verification of the extracted dataâs accuracy and the openness of AI development between countries, particularly comparing the US and China. There is also criticism of the announcementâs lack of technical detail, labeling it as âhype BSâ without verification.- rseymour raises a technical concern about the resolution capabilities of the DeepSeek-OCR AI, questioning the feasibility of using âvision tokensâ at a resolution of
1024x1024. They suggest that this resolution might be insufficient for accurately capturing the details of a microfiche sheet, which typically requires higher resolution due to its small size and dense information content. The comment implies that the technology might be overhyped without proper validation of its capabilities. - Robonglious discusses the openness of AI development between countries, specifically comparing the transparency of AI advancements in China versus the US. They speculate whether companies like OpenAI or Anthropic would release similar OCR technology if they developed it, suggesting that the US might be less cooperative in sharing such advancements compared to China.
- TheHeretic and Big_Firefighter_6081 express skepticism about the claims made regarding DeepSeek-OCR AIâs capabilities. They criticize the lack of verification and validation of the results, implying that the information might be more hype than reality. This highlights the importance of rigorous testing and validation in AI technology claims to ensure credibility.
- rseymour raises a technical concern about the resolution capabilities of the DeepSeek-OCR AI, questioning the feasibility of using âvision tokensâ at a resolution of
Less Technical AI Subreddit Recap
/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo, /r/aivideo
1. ChatGPT Atlas Browser Launch
- Meet our new browserâChatGPT Atlas. (Activity: 3175): ChatGPT Atlas is a new browser launched by OpenAI, currently available exclusively on
macOS. The browser integrates AI capabilities directly into the browsing experience, potentially enhancing user interaction with web content. However, the release is limited to Mac users, which has sparked some debate about accessibility and platform support. Commenters have raised concerns about data privacy and the decision to release the browser only for macOS, questioning the strategic choice and potential data handling practices.- Big-Info and douggieball1312 discuss the platform exclusivity of ChatGPT Atlas, noting that it is currently only available for Mac. This decision is critiqued as potentially alienating Windows users, especially given Microsoftâs financial backing of OpenAI. The irony is highlighted in the context of Microsoftâs investment, as Windows is a major competitor to Mac.
- Tueto raises concerns about data privacy with ChatGPT Atlas, questioning where user data is being sent. This reflects broader concerns about data handling and privacy in AI-driven applications, especially in the context of web browsing where sensitive information is often accessed.
- douggieball1312 points out the irony in ChatGPT Atlas being exclusive to Mac, despite OpenAIâs backing by Microsoft. This decision is seen as a reflection of a Silicon Valley tech bubble that may overlook the broader user base, particularly Windows users, which could impact adoption and user satisfaction.
- GPT browser incoming (Activity: 1511): The image is a social media post by Sam Altman, CEO of OpenAI, announcing a livestream event to launch a new product. The post is retweeted by OpenAI and features a graphic with the word âLivestreamâ and the OpenAI logo, indicating a significant announcement. The community speculates about the nature of the product, with some comments humorously suggesting a âsexbotâ or expressing concerns about privacy, likening it to âspywareâ similar to Googleâs practices. The engagement on the post suggests high interest and anticipation for the announcement. The comments reflect a mix of humor and skepticism, with some users joking about the product being a âsexbotâ and others expressing concerns about privacy, comparing it to Googleâs data practices.
- trustmebro24 speculates that the upcoming GPT browser might be based on Chromium, which is a common choice for many modern browsers due to its open-source nature and robust performance. Chromiumâs architecture allows for extensive customization and integration of advanced features, which could be beneficial for a browser leveraging GPT technology.
- qodeninja raises a concern about the potential for the company to overextend itself by developing too many products, suggesting that it might be more effective to allow the broader ecosystem to innovate and create complementary technologies. This reflects a strategic consideration about resource allocation and focus in tech development.
- Vegetable_Fox9134 mentions the potential privacy concerns associated with a new browser, comparing it to existing issues with Google. This highlights the ongoing debate about data privacy and the trade-offs users face when using technology that may collect personal information.
- OpenAIâs AI-powered browser, ChatGPT Atlas, is here (Activity: 1041): OpenAI has launched an AI-powered browser named ChatGPT Atlas, which integrates the capabilities of ChatGPT into web browsing. This tool aims to enhance user interaction by providing AI-driven insights and assistance directly within the browser environment. The integration is expected to streamline tasks by leveraging the conversational abilities of ChatGPT, potentially transforming how users interact with web content. The comments reflect a mix of skepticism and curiosity, with some users expressing concerns about privacy and the potential for misuse, while others are intrigued by the possibilities of AI-enhanced browsing.
- CONFIRMED: OpenAI is Launching a New Browser TODAY Called ChatGPT Atlas (Activity: 747): OpenAI has launched a new browser called ChatGPT Atlas, available globally on
macOSwith plans forWindows,iOS, andAndroidsoon. The browser integrates AI capabilities directly into the browsing experience, offering a chat interface for seamless AI communication. It is introduced by key figures like Sam Altman and Ben Goodger. The browser is perceived as a strategic move to compete with Google and Microsoft, though it has been critiqued for its similarity to existing browsers with added chat functionality. More details can be found in the YouTube video. Commenters express skepticism about the browserâs impact, noting it may primarily serve OpenAIâs data collection needs rather than offering significant user benefits. Concerns are raised about privacy and data being sent to OpenAI, with some questioning the necessity of the browser given existing alternatives.- The introduction of ChatGPT Atlas by OpenAI is seen as a strategic move to compete with tech giants like Google and Microsoft. While the browser may offer some convenience and speed improvements, there is skepticism about its impact on users. The primary concern is the extensive data collection capabilities, which could surpass current systems by learning about usersâ lives, interests, and behaviors in real-time.
- There is speculation that ChatGPT Atlas might be based on the Chrome engine, which would align with many modern browsers that leverage Chromium for compatibility and performance benefits. This choice could influence the browserâs adoption by providing a familiar user experience and support for existing web standards.
- A significant concern among users is the potential privacy implications of using ChatGPT Atlas. The browser could collect vast amounts of personal data, raising issues about how OpenAI will handle and protect this information. This concern is particularly relevant for users who may not fully understand the extent of data sharing involved.
2. Claude Desktop General Availability
- Claude Desktop is now generally available. (Activity: 836): Claude Desktop is now generally available for both Mac and Windows, offering seamless integration with local work environments. Users can access Claude by double-tapping the Option key on Mac, capture screenshots, share windows, and use voice commands via Caps Lock. The application supports enterprise deployment with
MSIXandPKGinstallers. For more details and to download, visit Claudeâs official site. Some users were confused about the announcement, thinking the app was already available, while others noted the absence of a Linux version. The Quick Entry feature is praised for its functionality.- ExtremeOccident mentions that despite Claude Desktop being in beta, the Quick Entry feature is effective, indicating a focus on user experience and efficiency in input handling.
- Logichris highlights a limitation in token allocation for Claude Desktop, comparing it to a âpaycheck to paycheckâ scenario, which suggests that the current token system may not support extensive use without frequent replenishment.
- Multiple users, including Yeuph and JAW100123, point out the lack of a Linux version, indicating a gap in platform support that could limit adoption among Linux users.
- {Giveaway} 1 Year of Gemini AI PRO (40 winners) (Activity: 2833): The post announces a giveaway for a one-year subscription to Gemini AI PRO for 40 winners, highlighting features such as the upcoming Gemini 3.0 Ultra,
1,000 monthly AI credits, and tools like Gemini Code Assist, NotebookLM, and integration with Gmail, Docs, and Vids. The package also includes2TB storageand extended limits on various applications, aiming to enhance productivity and creativity across different domains. Commenters highlight diverse uses of Gemini AI, such as aiding in storytelling and language translation for personal and professional purposes, supporting filmmaking through its ecosystem, and enhancing open-source contributions with code generation capabilities.- Bioshnev highlights the practical applications of Gemini AI in both personal and professional settings. He uses it for generating custom bedtime stories for his daughter and for work-related tasks like translating for foreign customers and retrieving product details. This showcases the modelâs versatility in handling language processing and information retrieval tasks.
- thenakedmesmer discusses the impact of Gemini AI on creative projects, particularly in filmmaking. He mentions using features like ânano bananaâ and âveoâ as part of a supportive ecosystem that aids in film production, illustrating how AI can serve as a virtual creative team, compensating for physical limitations and enhancing creative workflows.
- vladlearns emphasizes the importance of code generation capabilities in Gemini AI for open source contributions. This points to the modelâs utility in software development, where it can assist in automating coding tasks, potentially increasing productivity and supporting collaborative projects.
3. Amazonâs Robot Workforce Plans
- Amazon hopes to replace 600,000 US workers with robots, according to leaked documentsï»ż. Job losses could shave 30 cents off each item purchased by 2027. (Activity: 1630): Amazon is reportedly planning to replace
600,000US workers with robots by2027, as per leaked documents. This automation could potentially reduce costs by30 centsper item. The initiative is part of a broader strategy to address labor shortages and improve efficiency in fulfillment centers, a goal Amazon has pursued since acquiring Kiva Systems over a decade ago. The transition to robotics is seen as a necessary step due to high turnover rates and labor shortages in Amazonâs fulfillment centers. Commenters highlight that the cost savings may not translate to lower prices for consumers, and emphasize the strategic necessity of automation due to Amazonâs labor challenges. A former Amazon Robotics employee notes that the goal of replacing workers with robots has been longstanding but is progressing slower than anticipated.- The comment by âtheungodâ highlights a critical operational challenge for Amazon: the high turnover and difficulty in staffing their fulfillment centers (FCs). The user notes that Amazon has been aiming to automate these roles since acquiring Kiva Systems over a decade ago, but the transition to robotics has been slower than anticipated. This suggests that the integration of robotics into Amazonâs logistics is not just about cost savings but also about addressing labor shortages.
- âtheungodâ also provides an insider perspective, having worked at Amazon Robotics for over five years. They emphasize that the goal of replacing 600,000 workers with robots has been a long-standing objective, indicating that the technological and logistical hurdles are significant. This insight underscores the complexity of implementing large-scale automation in fulfillment operations, which involves not just technological development but also overcoming practical deployment challenges.
- The discussion touches on the broader implications of automation in logistics, particularly the potential societal impact. While the cost savings per item (30 cents) are noted, the focus is on the necessity of automation due to labor shortages rather than purely financial incentives. This reflects a shift in the narrative from cost-cutting to operational necessity, driven by the inability to maintain a stable workforce in demanding environments.
- Shape shifting drone (Activity: 1226): The post discusses a shape-shifting drone that appears to have a unique design, possibly inspired by biological forms, as suggested by the comment likening it to a âfloating colonoscopyâ. The image linked in the comments shows a drone with a flexible structure, which may allow it to adapt its shape for different flight dynamics or environmental conditions. This could be an innovative approach in drone technology, potentially enhancing maneuverability and efficiency. One comment suggests that the concept of a shape-shifting drone is not entirely new, indicating that similar designs may have been seen before. This could imply ongoing research and development in this area, reflecting a trend towards more adaptable and versatile UAV designs.
AI Discord Recap
A summary of Summaries of Summaries by gpt-5
1. GPU and eGPU Hardware Breakthroughs
- Blackwell Pro Packs 72GB, Quietly Drops: TechPowerUp reported NVIDIA quietly launching the workstation-class RTX Pro 5000 Blackwell with 72 GB GDDR7 memory, targeting pro workflows (NVIDIA RTX Pro 5000 Blackwell GPU with 72 GB GDDR7 appears).
- Engineers joked about likely pricing and use cases, while others flagged initial confusion over the unusual 72 GB capacity, mirroring similar coverage on VideoCardz.
- Tinygrad Makes Apple Silicon Love NVIDIA eGPUs: The tinygrad team announced early public testing of a pure-Python driver enabling NVIDIA eGPUs over USB4 on Apple Silicon using the ADT-UT3G dock,
extra/usbgpu/tbgpudriver, and NVK-basedtinymesacompiler (tinygrad enables NVIDIA eGPU on Apple Silicon (X)).- They measured about â3 GB/s PCIe bandwidth with SIP disabled and teased support for AMD RDNA 2/3/4 and Windows eGPU stacks next.
- Tiny Corp Boots NVIDIA on ARM MacBooks: Tiny Corp demonstrated an NVIDIA GPU running on an ARM MacBook via USB4 using an external dock, validating eGPU viability beyond Intel-era Macs (Tiny Corp Successfully Runs An Nvidia GPU on Arm Macbook Through USB4 Using An External GPU Docking Station).
- Mac users were upbeat, noting newer Pros with Thunderbolt 5 may further improve bandwidth headroom for local LLM and VLM workloads.
2. Triton/Kernel Tooling and Benchmarks
- FlashInfer-Bench Kicks Off Agentic Kernel Races: CMU Catalyst introduced FlashInfer-Bench, a workflow and leaderboard for agent-driven, self-improving LLM serving kernels with standardized signatures and integrations with FlashInfer, SGLang, and vLLM (FlashInfer-Bench blog).
- They published a live leaderboard and GitHub repo, inviting the community to iterate on kernels and benchmark updates.
- Triton Talks Stream and Sizzle: Developers shared full-session videos from the Triton conference at Microsoft, covering compiler advances and kernel design (Triton Conference livestream and Triton-openai streams).
- A recurring theme was hand-tuned PTX/assembly for critical kernels to beat compiler defaults, echoing calls to rethink execution from the ground up.
- Helion 0.2 Beta Fuzzes Triton to Tears: Helion 0.2 entered public beta as a Triton tile abstraction on PyPI, surfacing compiler edge cases during optimization passes (helion 0.2.0 on PyPI).
- Users reported MLIR failures in
TritonGPUOptimizeThreadLocalityPass, framing Helion as an effective Triton-compiler âfuzzerâ whose autotuner skips bad configs.
- Users reported MLIR failures in
3. OpenRouter SDK and New Reasoning Model
- OpenRouter SDK Types 300+ Models: OpenRouter released a TypeScript SDK (beta) with fully typed requests/responses for 300+ models, built-in OAuth, and support for all API paths (@openrouter/sdk on npm).
- SDKs in Python, Java, and Go are coming soon, aiming to simplify multi-model app development and authentication.
- Andromeda-alpha Cloaks Visual Reasoning: OpenRouter launched Andromeda-alpha, a small reasoning model focused on image/visual understanding, available for trial (Andromeda-alpha on OpenRouter).
- Since prompts/outputs are logged to improve the providerâs model, moderators warned: avoid personal/confidential data and do not use it for production.
- Mercury Outduels Qwen in Agent Arena: In agentic benchmarks, Inception/Mercury from provider Chutes edged Qwen on failure rate, latency, and cost in simple tasks (Chutes provider page).
- Members noted newer DeepSeek v3.1 models arenât free via Chutes anymore, though a free longcat endpoint remains (longcat-flash-chat:free).
4. Open-Source Models and Text-to-Video Releases
- Ring & Ling MoEs Land in llama.cpp: Ring and Ling MoE models from InclusionAI now run in llama.cpp, spanning 1T, 103B, and 16B parameter scales (llama.cpp PR #16063).
- Practitioners questioned real-world reasoning quality and verbosity control, hoping for a model that doesnât YAP during chain-of-thought.
- Krea Realtime Drops 14B Open T2V: Krea Realtime released a 14B open-source autoregressive text-to-video model distilled from Wan 2.1, generating long-form video at ~11 fps on a single NVIDIA B200 (Krea Realtime announcement (X)).
- Weights ship under Apache-2.0 on HuggingFace; users asked about ComfyUI workflows, RTX 5090 performance, and fine-tuning options.
- DeepSeek-OCR Joins the OCR Fray: DeepSeek-OCR arrived on GitHub, expanding the OCR toolkit with modern VLM-friendly design and multilingual aims (DeepSeek-OCR (GitHub)).
- Developers contrasted it with existing OCR stacks and highlighted the importance of contextual understanding for scripts like kanji.
5. AI Apps: ChatGPT Atlas Launch and Funding News
- OpenAI Ships Atlas, a Chromium AI Browser: OpenAI launched the ChatGPT Atlas browser for macOS, a Chromiumbased browser with boosted limits and multi-site browsing (Introducing ChatGPT Atlas and chatgpt.com/atlas).
- Early users flagged missing vertical tabs and built-in ad blocking (extensions required), while elsewhere users compared Atlas to Perplexityâs Comet, praising Cometâs privacy focus and integrated adblocker.
- AI Browser Buzz Meets Skeptic Snark: Engineers questioned the utility of new AI browsers, sharing skepticism over performance and data practices (AI browser hype thread (X)).
- One member quipped, âOpenAI knows this too, they are just farming data and throwing shit at the wall,â capturing wider concerns about hype versus real value.
- LangChain Grabs $125M to Build Agent Stack: LangChain raised $125M Series B, positioning a three-part stack: LangChain (agent dev), LangGraph (orchestration), and LangSmith (observability) (LangChain raises $125M (X)).
- They touted adoption by Uber, Klarna, and LinkedIn, signaling continued investor confidence in agent tooling and production ops.
Discord: High level Discord summaries
Perplexity AI Discord
- Comet Outshines Atlas in Privacy: Users compared Comet and ChatGPTâs Browser Atlas, favoring Cometâs commitment to privacy and integrated adblocker.
- Several noted the similarity between the two, but praised Comet for its features catering to user privacy.
- AI Therapy Sparks Ethical Concerns: Discord members debated the ethics of using AI for therapy, with some emphasizing the importance of human emotional maturity.
- Opinions diverged, with some suggesting âChatGPT is a good therapist,â while others cautioned against over-reliance on AI for mental health support.
- Perplexity Fanatic Shows Swag: A user showcased their Perplexity stickers, expressing enthusiasm for the brand and requesting a Comet hoodie from the Perplexity Supply store.
- The display of enthusiasm led to lighthearted jokes about resembling a âcult,â with others encouraging further purchases.
- API Users Want ChatGPT5 Access: A user inquired whether the Perplexity API grants access to models like ChatGPT5 and Claude, or is restricted to Sonar.
- The inquiry reflects a desire to utilize the API for potentially more advanced models beyond the currently available Sonar.
- Shareable Discord Threads Reminder: A message reminded users to ensure their Discord threads are set to
Shareableto be more accessible.- This ensures that links to the thread can be accessed by others, even outside of the specific channel, improving collaboration.
LMArena Discord
- Gemini 3 Pro Blows Away GPT-5 High in Web Design: Members compared Gemini 3 Pro and GPT-5 High, reporting that Gemini 3 Pro crushes web design.
- The general consensus is that Gemini 3 Pro is better for coding, whereas GPT-5 High is better for math and other tasks.
- Sora 2 Downgrade Sparks AI Subscription Debate: Members expressed frustration over the Sora 2 downgrade, leading to a broader discussion on the value of AI subscriptions.
- One member pointed out that their job performance is about 25-30% better because of AI, underscoring AIâs impact on desk job efficiency.
- Lithiumflow and Orionmist Speculated as Gemini 3 Checkpoints: Members speculated about the differences between Lithiumflow and Orionmist, ultimately concluding that these models are checkpoint versions of Gemini 3.
- The models sometimes erroneously claim training by OpenAI, suggesting potential model distillation.
- Open Source Models Allegedly Stealing Gemini 2.5 Pro: Discussion arose regarding the ethics of open-source models using stolen data for improvement, with claims that Chinese AI companies stole the 2.5 pro and made it open source.
- Some members agreed with the sentiment that this is okay as thatâs the only way open source can win.
- TikZ Generation Task Elicits Surprise: Members are exploring the use of LLMs to generate images in TikZ, a typesetting language, to avoid data contamination.
- Early results indicate some success in generating TikZ images with LLMs, demonstrating a novel approach to image creation.
Unsloth AI (Daniel Han) Discord
- Ring and Ling Launch!: Ring and Ling MoE models are now supported in llama.cpp (link to Github), including 1T, 103B, and 16B parameter models from InclusionAI.
- Members pondered the reasoning abilities of the models, with one hoping for a reasoning model that doesnât YAP.
- Disable Unsloth Statistics: To prevent telemetry calls when running Unsloth in offline mode, set the
UNSLOTH_DISABLE_STATISTICSenvironment variable andos.environ['HF_HUB_OFFLINE'] = '1', as the Unsloth community reached 100M lifetime downloads on Hugging Face (announcement on X).- Members also discussed resolving network issues by setting proxy environments.
- Nvidia RTX Pro 5000 Blackwell Workstation Card Quietly Appears: Nvidia quietly launched the RTX Pro 5000 Blackwell workstation card with 72GB of memory, as reported by VideoCardz.
- Initial confusion arose regarding the 72GB capacity, with one user joking that it was a way to bypass automod.
- User Fumes over Rate Limiting Tactics: A user ranted about a premium subscription service blocking access to URLs containing roman numerals, incorrectly interpreting them as malicious activity.
- The user, frustrated with manual workarounds and security plugins, criticized the service for ignoring requests to allow bulk downloads for pro subscribers.
- Nvidia GPU Transplanted onto ARM Macbook: A member shared an article from Tomâs Hardware on successfully running an Nvidia GPU on an ARM Macbook through USB4 using an external GPU docking station: Tiny Corp Successfully Runs An Nvidia GPU on Arm Macbook Through USB4 Using An External GPU Docking Station.
- This was exciting to Mac users since they have Thunderbolt 5 too now on the âprosâ which gives slightly more hope to Mac users.
Cursor Community Discord
- Codex Called a Billion Dollars Compared to Claudeâs Toonie: Users debated the merits of Codex vs Claude for code generation, with one user stating that comparing them is like comparing âa billion dollars or a toonieâ.
- No further details were provided.
- Cursor Site Crashes, Subs Lost: Multiple users reported the Cursor website being down for several hours, preventing them from logging in, upgrading plans, or renewing subscriptions.
- Some suspected AWS issues as the root cause, while others pointed out the lack of subscription expiration notifications as a major inconvenience.
- Dashboard Cracks Cursor Costs Post-Pricing Changes: A user shared a dashboard they created to track actual Cursor costs after the pricing changes, especially for users on legacy pricing plans and gave this forum link cursor.com/blog/aug-2025-pricing.
- The tool requires cookie login or .json upload from the userâs local machine, but promises comparison with real API pricing.
- Background Agents Encounter Internal Error: A member reported encountering an internal error when running a first experiment with background agents via Linear, where the agent starts, does some thinking and grepping, but then stops.
- The error message received was: âWe encountered an internal error that could not be recovered from. You might want to give it another shot in a moment.â
OpenRouter Discord
- OpenRouter SDK: Beta Boost: The new OpenRouter SDK is in beta on npm, aiming to be the simplest way to use OpenRouter and offering fully typed requests and responses for 300+ models.
- Python, Java, and Go versions are coming soon, featuring built-in OAuth and support for all API paths.
- Andromeda-alpha: Cloaked & Ready: OpenRouter launched a new stealth model named Andromeda-alpha (https://openrouter.ai/openrouter/andromeda-alpha), a smaller reasoning model focused on image and visual understanding.
- Prompts/outputs are logged to improve the model, users are cautioned against uploading personal/confidential info and not using it for production.
- Objective AIâs Confidence Code: Objective AI now offers a Confidence Score for each OpenAI completion choice, derived through a smarter method than directly asking the AI and emphasizing cost-efficiency.
- The CEO is building reliable AI Agents, Workflows, and Automations free of charge using n8n integration to gather more examples.
- Mercury Swats Qwen in Agent Arena: Inception/Mercury (Chutes provider) edges out Qwen in simple agentic tasks, exhibiting lower failure rate, faster speed, and lower cost.
- New Deepseek models like v3.1 arenât available as free versions through Chutes, though they recently added a free longcat endpoint.
- AI Browser Bandwagon Bewilders Brains: Members are skeptical towards the hype around new AI browsers like Xâs AI browser, questioning the utility and performance impact of integrated AI.
- One member compared the hype to the dotcom bubble, stating that OpenAI knows this too, they are just farming data and throwing shit at the wall.
OpenAI Discord
- OpenAI Launches Atlas Browser: OpenAI released the ChatGPT Atlas Browser for macOS today at chatgpt.com/atlas, detailed in their blog post.
- The browser, based on Chromium, boasts boosted limits, direct website access and support for multiple websites, but lacks vertical tabs and a built-in ad blocker.
- Meta Shuts Down 1-800-ChatGPT on WhatsApp: Meta is blocking 1-800-ChatGPT on WhatsApp after January 15, 2026, according to a blog post.
- The change is due to Metaâs new policies.
- Sora Limits Video Lengths: The Sora iOS app limits video generation to 10-15 seconds, while the web version allows longer videos for Pro subscribers.
- Free and Plus users can also generate longer videos on the web version, with Pro users having access to the storyboard feature and generating up to 25-second videos.
- AI-Driven OS Prototype Appears: A member introduced a prototype AI-driven OS, featuring an AI Copilot Core, a Seamless Dual-Kernel Engine, and a NeoStore for AI-curated apps (source).
- Further components include a HoloDesk 3D workspace, an Auto-Heal System, Quantum Sync, and an Atlas Integration Portal for accessing external AI tools.
- GPT-4 Annoying Users: A user expressed irritation with GPT-4âs new condescending tone, especially phrases like âif you insist,â and requested to make the model less confident.
- No solutions were provided, other than a general agreement that the new GPT is annoying.
HuggingFace Discord
- GPT-4o Saves Data Pipeline?: A member suggested using GPT-4o or other vision models for high-accuracy labeling and automated comparisons, but they were also cautious about the costs of replacing Apache Beam.
- Another member thought the architecture was overkill, likening it to proposing a starship to go to the grocery store.
- Unslothâs Script Tunes LLMs Easily: A member requested insights into setting up Parameter-Efficient Fine-Tuning (PEFT) on Large Language Models (LLMs), and another member pointed out challenges in multi-GPU setups and suggested using Unslothâs script on Colab Free.
- They cautioned about handling internal company data and linking to further resources, like the Fine-tuning LLMs Guide.
- Databomz Manages All Prompts!: A member introduced Databomz, a workspace and Chrome extension for saving, organizing, and sharing prompts with features like tags, versions, and folders.
- The member highlighted a Forever Free plan and encouraged feedback from prompt engineers.
- Solo Dev creates TheLastRag: A solo developer created an entire LLM Framework called TheLastRag, highlighting features like True memory, True personality, True learning and True intelligence, and is looking for feedback.
- The main points are that the AI never forgets, has a true personality, has true learning, and has true intelligence.
- Local VLM Training Consumes Gigabytes of Memory: A member reported that while training the VLM exercise locally, itâs using a large amount of swap memory (62GB claimed and ~430GB virtual memory).
- The same member asked if thereâs a way to limit memory usage specifically for MPS (Metal Performance Shaders) on Macs, with a goal to enable training within a more reasonable 40GB VRAM limit.
LM Studio Discord
- LM Studio Struggles Linking llama.cpp: Members noted that how can I call my own llama.cpp for LM Studio to use is not yet fully supported and the LM Studio docs that references this are a broken link.
- There is not an obvious known workaround, so users may have to wait until the feature is added.
- AGI ETA: 2044?: A member forecasted that AGI is 10-20 years away, claiming that in 5 years LLMâs will probably have context large enough.
- Another member jokingly suggested he become a consultant and charge 1000/h.
- GPT-OSS Reasoning Demands Metadata: A user inquired about setting reasoning effort in GPT-OSS finetunes, with a member responding that it works due to the metadata in the mxfp4 model of gpt-oss, which is why finetunes/ggufs donât have it.
- The helpful member offered to make it available before quantizing it to gguf.
- OpenWebUI Connects to LM Studio via OpenAI: When trying to connect OpenWebUI to LM Studio, users suggested leveraging the OpenAI option instead of OpenAPI.
- Members helped troubleshoot the connection, pointing to this huggingface discussion recommending to put /v1 in the address.
- NVIDIAâs RTX Pro 5000 Blackwell Leaks: A member shared a TechPowerUp article about NVIDIAâs RTX Pro 5000 Blackwell GPU featuring 72 GB of GDDR7 memory.
- Excited users reacted with humor, guessing the card will cost around $8-10k.
Latent Space Discord
- TinyGrad Powers Apple Silicon eGPUs: Tinygrad now supports NVIDIA eGPUs on Apple Silicon via USB4, enabling users to run external RTX 30/40/50-series GPUs using an ADT-UT3G dock with the
extra/usbgpu/tbgpudriver and NVK-basedtinymesacompiler (source).- With SIP disabled, this setup achieves roughly 3 GB/s PCIe bandwidth, and future support for AMD RDNA 2/3/4 and Windows eGPU stacks is planned.
- Krea AI Unveils Realtime Video Model: Krea AI released Krea Realtime, a 14B open-source autoregressive text-to-video model distilled from Wan 2.1, generating long-form video at 11 fps on a single NVIDIA B200 (source).
- Released weights are on HuggingFace under Apache-2.0, prompting user inquiries about ComfyUI workflows, RTX 5090 performance, and fine-tuning support.
- Google AI Studioâs âVibe-Codingâ with Gemini: Google AI Studio is launching a new âprompt-to-productionâ Gemini experience after five months of development aiming to make AI-app building 100Ă easier (source).
- Reactions mixed excitement (requests for mobile app, opt-outs, higher rate limits), feature suggestions (GSuite-only publishing, VS Code plug-in, short browser-agent tasks) and some skepticism about fit vs Gemini 3 expectations; team confirms enterprise-only deployment is already available.
- Fish Audio S1: TTS Revolution?: Fish Audio launched S1, a text-to-speech model thatâs purportedly 1/6 the cost of ElevenLabs, touting 20k devs and $5M ARR (source).
- Users shared instant voice-clone demos, asking about real-time latency (~500ms), while founders admitted current limits and promised wider language support + conversational model next.
- Second-hand RTX 3090 Buying Tips: Taha shared lessons learned after buying a used RTX 3090: meet seller in person to inspect card, bring a portable eGPU test rig, verify recognition with nvidia-smi, run memtest_vulkan for VRAM integrity, optionally gpu-burn for compute stress, load a large model and monitor temps <100 °C; see guide here.
- The test rig is a Framework 13 Ryzen laptop on NixOS in PRIME offload mode, and a user suggested trying tinygrad on their rig since mine works ootb since Iâm on linux.
GPU MODE Discord
- AMDâs Web3 Cloud Gambit: At an AMD event, the company emphasized the âcloudâ aspect of web3, raising some eyebrows (smileforme emoji).
- Details of AMDâs specific offerings remain vague, leaving the community to speculate on their approach to decentralized technologies in the cloud.
- FlashInfer-Bench Automates AI: FlashInfer-Bench was introduced by CMU Catalyst as a workflow for creating self-improving AI systems via agents, featuring standardized signatures for LLM serving kernels and integration with FlashInfer, SGLang, and vLLM (blog post, leaderboard, GitHub repository).
- The project aims to foster community development and benchmarking, enabling AI systems to iteratively enhance their performance.
- Triton Conference Electrifies Microsoft: Members who attended the Triton conference at Microsoft in Mountain View shared a YouTube link to watch the conference online and a link to the Triton-openai streams.
- The conference brought together developers and researchers to discuss the latest advancements and applications of the Triton language.
- NCCL Kernels run on PG-NCCLâs internal streams: When a
CUDAStreamGuardis set and an NCCL op is called viaProcessGroupNCCL, the NCCL kernels run on PG-NCCLâs internal streams, typically using one stream per device with high priority, and using the tensor lifetime stream (relevant code).- Setting a
CUDAStreamGuarddetermines which stream the NCCL stream waits on, establishing an incoming dependency, as seen in the pytorch source code.
- Setting a
- SLNG.AI Hunts Voice AI Performance Wiz: SLNG.AI is on the lookout for a Speech Model Performance Engineer to build the backbone for real-time speech AI (more details).
- The role requires a strong software engineering background to optimize and enhance speech model performance.
Yannick Kilcher Discord
- IntelliCode reads your mind: A member expressed awe at Microsoftâs IntelliCode in Visual Studio, an AI-powered code completion tool that accurately predicts entire method bodies by leveraging a lot of context.
- They remarked that it was almost like itâs reading your mind when it works well due to its ability to understand and anticipate coding needs with impressive accuracy.
- DeepSeek OCR Enters the Ring: DeepSeek-AI released DeepSeek-OCR on GitHub, joining the competition in the OCR technology space.
- Also, Anthropic released Claude Code on the web, expanding options for developers seeking AI-assisted coding tools.
- Amazon Vibe Code ditches Beta: Amazonâs Vibe Code IDE is out of invite-only beta, but it costs 500 credits to use.
- It is yet another VSCode fork that leverages AI.
- Open Source details evade Westâs grasp?: A member lamented the Westâs lack of superior OS labs, as Deepseek consistently unveils impressive discoveries.
- They pointed out that open source weights account for only a fraction of the overall value, emphasizing the importance of open source data collection, methods, and training details.
- Unitree set to crush on Tesla?: A member predicted that Unitree will dominate the humanoid robotics market.
- They speculated that Elon Musk may be struggling to acquire necessary components, quipping he probably canât even get the magnets for the actuators at the moment thanks to the orange dude.
DSPy Discord
- DSPy Powers AI NPC Voices: A member built a voice generation system for game NPCs, using DSPy to parse wiki content and generate voice prompts for ElevenLabs, also sharing a devlog style video.
- They plan to leverage DSPyâs optimization features to improve the character analysis pipeline and automate voice selection, and intend to collect manual selections as training signals, optimizing toward subjective quality judgments in the future using an automated judging loop.
- DSPy Featured in Research Paper: A new paper (https://arxiv.org/abs/2510.13907v1) utilizes DSPy in its research, signaling growing adoption within the academic community.
- Although the paper mentions the use of DSPy, the corresponding code repository is not yet publicly available.
- Navigating DSPy History Access: Members debated why
inspect_history()is a method indspyrather than a module object, and clarified thatdspy.inspect_history()is more for global history and individual programs also track history.- It was pointed out that history can be accessed with
predictor.historyifdspy.configure(track_usage=True)is set, but some still found this confusing.
- It was pointed out that history can be accessed with
- Demystifying DSPy Adapters with Context: The discussion covered using adapters in DSPy, with an example showing how to use
dspy.contextto apply a single adapter, and the user can track usage withdspy.configure(track_usage=True).- A member gave an example of setting it up with
with dspy.context(lm=single_call_lm, adaptor=single_adaptor):to further clarify the process.
- A member gave an example of setting it up with
- Trace Claims Accuracy Edge Over DSPy: A member asked for a comparison between Microsoft Trace and DSPy, with another noting that Trace claims an 8% accuracy increase over DSPy and appears more token efficient.
- One member mentioned they would try it out to give a fair comparison, although they will probably still feel like they have more granular control with DSPy.
Eleuther Discord
- Discord Server Badges Spark Debate: Members discussed the possibility of adding a server badge, similar to a role icon, and how a server tag might broadcast the server too widely, potentially increasing the moderation load for EAI staff, referencing this screenshot.
- One member noted, *âmaking a tag is cool but that is in a way broadcasting this server everywhere else, eai staff already gets too many people here to moderate.â
- EleutherAI IPO Dreams Spark Jokes: Following a question about whether a particular stock symbol was available, a member jokingly asked, *âwhat will Eleutherâs NYSE stock symbol be?â
- Another member responded, âI think you misunderstand the purpose of being a non-profit,â implying that EleutherAI, as a non-profit organization, would not be publicly traded.
- Normuonâs Triumph Prevents Logit Blowup: A member noted that normuon beating muon even with qk-norm (which avoids logit blowup) in their baseline suggests logit blowup prevention might not fully explain the performance parity.
- It was posited that updates without clipping increase the spectral rank of weights, directly leading to logit blowups, making large-scale validation against normuon interesting.
- AGI Definition Benchmarks Beckon: A member shared a link to Dan Hendrycksâ AGI Definition benchmarks and asked how fast they would be benchmarked.
- Another member predicted multimodality would likely be covered in 1-2 years, with speed coming from mini versions of models.
Manus.im Discord Discord
- Cloudflare snags Manus Users: Users are reporting issues with Cloudflare security when visiting most websites while using Manus.
- A suggestion was made for the Manus team to consider open-sourcing some of their older models, possibly to bypass the Cloudflare issues.
- Payment Problems Plagues Platform: A user encountered issues paying for credits via a web browser, experiencing jumbled code and transaction failure.
- The user stated this is a known issue and contacted support; lucia_ly requested their email to follow up and resolve the payment issues.
- Chat Slowdowns Irk Users: A user reported excessive delays in chat processing when translating long Japanese chapters into English.
- Despite usually appreciating Manusâs speed, the user noted, âthis morning, I put one chapter and the ai is still thinking. What happened?â
- Pro Plan Credit Cap Confusion Continues: Users are reporting conflicting information about unlimited credits on the Pro plan, with the help system and iOS upgrade page stating it is unlimited, while the PC upgrade page indicates a high limit.
- One user with 11k credits remaining was concerned about depletion, and another suggested, that they should participate in âvarious opportunities to help improve Manus, as they always give free credits for your timeâ.
- Scam Alert issued to Users: A user was accused of being a âfraudster scammerâ asking for peopleâs login access to their accounts to do their *âfucking law school exam researchâ.
- Another user warned, that the supposed fraudster âwont make another account or pay $20/month and complains its like tomorrow and begging to get ur EMAIL PASSWORD for a PAID ACCOUNT To probably steal ur personal info and bank infoâ.
Nous Research AI Discord
- Chinaâs A.I. Competition benefits the Globe: A member believes that Chinaâs insane spartan involution competition in A.I. is great for the A.I. space because it democratizes access to advanced models and destroys monopolies.
- They also state that the rate of advancement in OS model development means that 2026 should bring us OS models reaching 100% high intelligence with 90% lower cost, destroying monoplist ambition.
- Nous Promoted as Decentralized A.I.: A member notes that Nous Research is promoted as Decentralize A.I. and hopes the team will resolve issues with centralization, linking to the Nous Psyche page.
- Another member stated they are more focused on the democratization of A.I models for the masses, citing a Stanford paper on centralization and asserting that Nous successfully decentralizes with their open source methodologies and infrastructure implementations.
- Sora AI Project Showcased: A member showcased a video creation with Sora, sharing the video at 20251022_0850_01k850gmktfant3bs18n3hbd79.mp4.
- The videoâs content and implications for AI-driven content creation are under discussion within the community.
- Microsoft Trace Utility Resurfaces: A member shared a link to the Microsoft Trace utility, noting that apparently itâs not all that new.
- Its features and capabilities are being re-evaluated in light of current development practices.
tinygrad (George Hotz) Discord
- Nvidia Drivers Hacked onto macOS: Madlads accomplished the impossible, porting Nvidia drivers to macOS and sparking excitement within the community.
- The driver port enables running tinygrad on macOS with Nvidia GPUs, opening new possibilities for development and testing.
- GLSL Renderer Almost Ready: The community has been developing a GLSL renderer for tinygrad, which is now passing the majority of tests and is available on GitHub.
- This marks a significant step toward expanding tinygradâs compatibility with different platforms and graphics APIs.
- clspv Bugs Plague Vulkan Backend: Progress on tinygradâs Vulkan backend is hampered by numerous bugs in clspv, requiring optimizations to be disabled (
-O0 --cl-opt-disable) to pass tests.- The member also reported much more miscompilations from clspv if optimizations arenât disabled.
- Vulkan Sineâs Accuracy Troubles: The Vulkanâs sine function isnât as accurate, requiring a custom implementation which would impact performance.
- This accuracy issue could pose challenges for tinygradâs performance on Vulkan, necessitating careful consideration of alternative sine implementations.
- TinyJitâs Gradient Addition is Broken: Gradient accumulation was broken in TinyJit a couple months ago and the member fixed it by rewriting the gradient addition step to use an assign.
- A member also reported running into issues with gradient accumulation and fixed it by setting
reduction=sumand manually counting non-padding tokens.
- A member also reported running into issues with gradient accumulation and fixed it by setting
Moonshot AI (Kimi K-2) Discord
- Karpathy Criticism Raises Bubble Concerns: A member speculates that recent mockery of Karpathy on X might signal a valuation bubble in American frontier AI labs, citing this post.
- The referenced post features a chart that appears to be mocking Karpathy, though without explicit context from the original poster.
- Kimi K-2 Support Faces Scrutiny: A member reported a lack of response from Kimi support, noting zero communication regarding their issue.
- Other members clarified that the channel isnât an official support platform, recommending direct messaging and requesting details about the problem and the email used for the bug report.
Modular (Mojo đ„) Discord
- Python Familiarity Boosts Mojo Discovery: A member suggested that prior experience with Python facilitates easier discovery of Mojo and the unique features it offers.
- However, discrepancies between Mojo and Python could potentially cause confusion for new users.
- Human Touch Beats Compiler in Matmul Tuning: A discussion arose regarding why matmul optimizations arenât directly integrated into the compiler, especially given their performance impact.
- The response highlighted that manual tuning by kernel writers often surpasses compiler optimizations for hot-path code, allowing for fine-tuning to specific hardware, with reference given to Mojoâs open-source, hardware-optimized matmuls.
- Freeing Kernel Writers from the Compiler: Moving optimizations out of the compiler expands contribution opportunities to more kernel writers.
- This approach allows compiler engineers to focus on broader ecosystem improvements rather than niche optimizations, such as a 1% boost to matmuls where one dimension is less than 64.
- Finished Type System Tops Mojo Wishlist: When asked about the most crucial missing feature in Mojo, a member emphasized the need for a finished type system.
- Additional desired features include rounding out standard library datatypes, proper IO, a good async runtime, an effect system, static reflection, compiler plugins, the ability to deal with more restrictive targets, cluster compute, device/cluster modeling, and some clone of Erlangâs OTP.
MCP Contributors (Official) Discord
- GitHub Actions Fail Amidst Billing Issues: GitHub Actions are currently failing because the account is locked due to a billing issue.
- Users should address the billing issue promptly to restore GitHub Actions functionality.
- GitHub Actions Billing Lockout: The root cause of the failing GitHub Actions is a billing lockout on the account.
- Immediate resolution of the billing issue is necessary to restore the functionality of GitHub Actions.
The LLM Agents (Berkeley MOOC) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The Windsurf Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
You are receiving this email because you opted in via our site.
Want to change how you receive these emails? You can unsubscribe from this list.
Discord: Detailed by-Channel summaries and links
Perplexity AI â· #general (1090 messagesđ„đ„đ„):
Comet vs Atlas, AI and Mental Health, Schumacher vs Senna, Perplexity Merch, Using AI Responsibly
- Comet Crushes Atlas in Privacy Showdown: Users debated the merits of Comet versus ChatGPTâs Browser Atlas, with many valuing Cometâs focus on privacy and built-in adblocker, noting that they are âessentially copies of each other.â
- AI Therapy: A Mental Health Minefield?: Discord members questioned the ethical implications of using AI for therapy, some highlighted the importance of human emotional maturity and responsibility, while others suggested that âChatGPT is a good therapist.â
- Schumacher > Senna?: A long discussion comparing Schumacher and Senna, with one member declaring that âSchumacher was better than Senna,â while another stated âhe was for sure the best one ever.â
- Perplexityâs New Swag: Is It a Cult?: A member proudly showed off their Perplexity stickers on their laptop, joking about being a âPPLX fanâ and the need for a Comet hoodie and water bottle from the Perplexity Supply store.
- Some users lightheartedly joked about this level of enthusiasm resembling a âcultâ, while others playfully encouraged them to buy everything.
- Navigating the AI Maze: Responsibility Required: In Germany, it is required you must say that you have used AI when you are working with it.
Perplexity AI â· #sharing (3 messages):
Shareable Threads, Time-based Researcher
- Discord Threads Should Be Shareable: A message reminded users to ensure their Discord threads are set to
Shareable.- This ensures that links to the thread can be accessed by others, even outside of the specific channel.
- Time-Based Researcher Launched: A user shared a link to a Perplexity AI search for a time-based researcher.
- The link directs to perplexity.ai/search/time-base-researcher.
Perplexity AI â· #pplx-api (2 messages):
Perplexity API, ChatGPT5, Claude, Sonar
- Perplexity API Question Asks About Model Access: A user inquired about whether the Perplexity API allows access to ChatGPT5 and Claude, or if it is limited to Sonar.
- The inquiry is centered around understanding the scope of model access provided through the Perplexity API.
- Clarification on Model Availability via Perplexity API: The user seeks to confirm if the Perplexity API extends beyond the Sonar model to include access to more advanced models like ChatGPT5 and Claude.
- This reflects interest in leveraging the API for potentially higher-performing models if available.
LMArena â· #general (1064 messagesđ„đ„đ„):
GPT-5 vs Gemini 3, Sora 2 and Video Generation, TikZ Generation, Gemini 3 Model Performance
- Gemini 3 Pro Crushes GPT-5 High in Web Design: Members are discussing whether to wait for a GPT-5 release or use Gemini 3 Pro with one member reporting that Gemini 3 Pro crushes web design.
- They find that Gemini 3 Pro is better for coding while GPT-5 High is better for math and other miscellaneous tasks.
- Sora 2 Downgrade Drives AI Subscription Debate: Members are upset about the Sora 2 downgrade, which prompted a conversation about the value of AI subscriptions.
- One member noted my job performance is about 25-30% better because of AI, highlighting AIâs impact on desk job efficiency, while others are not so sure of the value.
- Lithiumflow and Orionmist are Gemini 3?: Members speculate about the differences between Lithiumflow and Orionmist, with a conclusion that the models are checkpoint versions of Gemini 3.
- Members have discovered that the models sometimes claim to be trained by OpenAI which suggests that the models may have been distilled.
- Open Source Models Distilling Gemini 2.5 Pro: There is discussion regarding open-source models stealing data to improve, with one member suggesting that the Chinese AI companies stole the 2.5 pro and made it open source.
- Members agreed that this is okay as thatâs the only way open source can win.
- TikZ Generation Task Elicits Surprise: Members are prompting models to make images in TikZ, a typesetting language, to avoid data contamination in models.
- Members have found some level of success in generating TikZ images with LLMs.
Unsloth AI (Daniel Han) â· #general (367 messagesđ„đ„):
Magistral and Think Tags, Grok 4 Fast vs Deepseek V3.2, Ring/Ling MoE Models, Unsloth Telemetry and Offline Mode, Qwen3-VL Models
- Magistral Learns To Think Different: A member found that Magistral learned to use the
<think>tag instead of the[THINK]tag, but by using FastLanguageModel it lost the ability to use the vision encoder.- Additionally, the model overthinks like crazy because of the tags.
- Deepseek V3.2 vs Grok4Fast: Data Generation Duel: A member is deciding between using Grok 4 Fast or Deepseek V3.2 for synthetic data generation due to budget constraints.
- They noted that r1-0528 is pretty cheap, especially 3.1 on Parasail which is 0.6 input/1.7 output per million, but questioned provider reliability, with another member pointing out that Open Router is just too inconsistent in model quality by provider.
- Ring/Ling MoE Models Launch: Ring and Ling MoE models are now supported in llama.cpp (link to Github), including 1T, 103B, and 16B parameter models from InclusionAI.
- Members pondered the reasoning abilities of the models, with one hoping for a reasoning model that doesnât YAP.
- Disable Unsloth Telemetry in Offline Mode: Members discussed running Unsloth in offline mode, with one user resolving network issues by setting proxy environments.
- It was suggested to set the
UNSLOTH_DISABLE_STATISTICSenvironment variable andos.environ['HF_HUB_OFFLINE'] = '1'to prevent telemetry calls, as the Unsloth community reached 100M lifetime downloads on Hugging Face (announcement on X).
- It was suggested to set the
- Qwen3-VL Models: Thinking Big: Qwen3-VL-2B was released, with members noting that Qwen3 VL 8B 4-bit runs easily on 16GB of RAM, and there was a direct upgrade to Qwen3-32b-Instruct.
- It was then asked if anyone has been able to run unsloths qwen3 VL 32b with llama.CPP but VL is not merged into llama.cpp yet.
Unsloth AI (Daniel Han) â· #introduce-yourself (9 messagesđ„):
AI Bot Development, Workflow Automation with LLMs, AI Content Detection, Image AI Pipeline, Voice Cloning and Transcription
- Veteran Dev Explores New AI Tricks: A developer with a background in building bots using ChatGPT is now diving deeper into AI and expresses their enthusiasm for Unsloth.
- They are experienced in gaming and scraping, showcasing a desire to learn new skills.
- Engineer Pioneers Workflow Automation with LLMs: An engineer specializing in workflow automation, LLM integration, RAG, AI detection, image, and voice AI describes their experience building automated pipelines and task orchestration systems using Dspy, OpenAI APIs, and custom agents.
- They have created a support automation system connecting Slack, Notion, and internal APIs to an LLM, reducing response times by 60%.
- AI Content Detection Tools Deployed: The engineer developed AI content detection tools for a moderation platform using stylometric analysis, embedding similarity, and fine-tuned transformers to identify GPT-generated text with high precision.
- Details were provided about an image AI pipeline using CLIP and YOLOv8 on AWS Lambda and S3, classifying and filtering thousands of images daily.
- Voice Cloning Service Built: A voice cloning and transcription service was built using Whisper and Tacotron2, enabling personalized voice assistants through ASR, TTS, and CRM integration.
- The individual has deep expertise in blockchain technology, including smart contract development (Solidity and Rust), decentralized application architecture, and secure on-chain/off-chain integrations.
Unsloth AI (Daniel Han) â· #off-topic (143 messagesđ„đ„):
Ultravox encoder and LLMs, REAP algorithm, Nvidia RTX Pro 5000 Blackwell, scraping content with rate limits, evaluation loss influenced by outliers
- Ultravox Projector Plugs into LLMs: The Ultravox project involves adding a projector to an LLM and training only the projector, without training the LLM, which is similar to how Voxtral works and is available on GitHub.
- A member confirmed the configuration improves with more data, clarifying that there is a training pass over the projector; however, it might be possible to ârip off the audio encoder from Qwen 2.5 Omni and slap it in Qwen 2.5 VL and just train a simple projectorâ.
- DeepSeek Dials Down Resource Use: A new DeepSeek model reduces resource usage by converting text and documents into images, using up to 20 times fewer tokens via vision text compression, further discussed on Tomâs Hardware.
- A member indicated that Gemma already implemented a similar approach, while another shared links about the Cerebras REAP algorithm, which was lauded as so cool.
- Nvidiaâs RTX Pro 5000 Blackwell Workstation Card Quietly Launches: Nvidia quietly launched the RTX Pro 5000 Blackwell workstation card with 72GB of memory, as reported by VideoCardz.
- Initially, there was confusion about the 72GB capacity, and one user noted this was a way to bypass automod.
- User Rages About Rate Limiting: A user ranted about a premium subscription service blocking access to URLs containing roman numerals, interpreting them as malicious activity.
- The user also has to manually search and circumvent the systemâs security plugin, and complained about the service ignoring requests to allow bulk downloads for pro subscribers.
- Evaluation Loss Skewed by Outliers: One member highlighted that evaluation loss can be significantly influenced by outliers in the evaluation set.
- With mean eval loss at 0.85, median (of per-example means) eval loss is 0.15, and 95th percentile at 0.95, the member suggested that poor generalization may not necessarily be indicated.
Unsloth AI (Daniel Han) â· #help (26 messagesđ„):
GRPO recipe for gpt oss 20b struggling, Vision model on llama-server, Quantized parameters in bitsandbytes, Algorithmic changes to GRPO, Version mismatch in Unsloth notebooks
- GPT OSS 20B GRPO Recipe Falls Flat!: A user reported that the GRPO recipe for gpt oss 20b is still struggling after running 100 steps using this notebook.
- They indicated modifications were made to get it running on Modal.
- Vision Models Vanish on Llama-Server!: A user inquired about running a vision model on llama-server, specifically asking if any arguments are needed.
- No solutions or workarounds were given in the discussion.
- Quantized Parameter Quest for Bitsandbytes!: A user sought to locate the internal values (scaling, center, etc.) of quantized parameters in a bitsandbytes model to apply noise directly.
- They noted that modifying the parameter directly wonât work due to dequantization requirements and memory usage.
- Unsloth GRPO Algorithmic Alterations!: A user asked if Unsloth is âhackableâ regarding algorithmic changes to GRPO (e.g., applying dense reward) without ruining optimizations.
- No response was given
- Notebook Version Nightmares!: A user complained about dealing with version mismatches while running Unslothâs GitHub notebooks, stating that most are not replicable.
- No solutions or workarounds were given in the discussion.
Unsloth AI (Daniel Han) â· #showcase (2 messages):
Brainstorm model
- Brainstorm Model Might Improve Stability: A member mentioned they might add Brainstorm (20x) to their model to see what happens, anticipating it will increase metrics as well as long gen stability.
- Another member requested the results to be posted if the member actually does that.
- Empty Topic: There was not much discussed in this message history.
- The discussion was not detailed enough to create two distinct summaries.
Unsloth AI (Daniel Han) â· #research (5 messages):
Kyutai Codec Explainer, Nvidia GPU on ARM Macbook, Thunderbolt 5
- Kyutai Codec gets Explained: A member shared a link to the Kyutai Codec Explainer.
- Nvidia GPU Transplants onto ARM Macbook: A member shared an article from Tomâs Hardware on successfully running an Nvidia GPU on an ARM Macbook through USB4 using an external GPU docking station: Tiny Corp Successfully Runs An Nvidia GPU on Arm Macbook Through USB4 Using An External GPU Docking Station.
- Thunderbolt 5 Sparkles Hope for Mac Users: A member noted that they have Thunderbolt 5 too now on the âprosâ which gives slightly more hope to Mac users.
Cursor Community â· #general (343 messagesđ„đ„):
Codex vs Claude, Github spec-kit, Cursor Meetups, Cursor site down, AWS CEO fired
- Codex is like having âa billion dollars or a toonieâ: Users debated the merits of Codex vs Claude for code generation, with one user stating that comparing them is like comparing âa billion dollars or a toonieâ.
- Cursor Site Downtime and Subscription Issues Plague Users: Multiple users reported the Cursor website being down for several hours, preventing them from logging in, upgrading plans, or renewing subscriptions.
- Some suspected AWS issues as the root cause, while others pointed out the lack of subscription expiration notifications as a major inconvenience.
- Cursor Team Plan Pricing Model: Users discussed the shift to a usage-based pricing model for Cursor team plans, replacing the previous fixed request limit, and that the plan is currently still operating under the legacy request-based system, but will automatically migrate to the new pricing at the next billing cycle.
- One user shared their bossâs correspondence with Cursor support, clarifying the new pricing structure and its impact on team plans and also shared this link with the new pricing model cursor.sh/pricing-update-sept-2025.
- Cracking Cursor Costs with a Custom Dashboard: A user shared a dashboard they created to track actual Cursor costs after the pricing changes, especially for users on legacy pricing plans and gave this forum link cursor.com/blog/aug-2025-pricing.
- The tool requires cookie login or .json upload from the userâs local machine, but promises comparison with real API pricing.
- Nightly Builds and Installation Guide Available: One user asked where to download nightly versions, and another shared, that you need to go to Settings -> Beta -> Early access to see the nightly build.
- However, another user noted there seems to be an issue with the new update and it does not prompt the user that you are in âaskâ mode.
Cursor Community â· #background-agents (1 messages):
Background Agents in Linear, Internal Error Troubleshooting
- Background Agents Error in Linear: A member reported encountering an internal error when running a first experiment with background agents via Linear, where the agent starts, does some thinking and grepping, but then stops.
- The error message received was: âWe encountered an internal error that could not be recovered from. You might want to give it another shot in a moment.â
- Troubleshooting the Internal Error: The user mentioned that the background agent seems to start but then fails, with the Cursor output showing only ââŠâ.
- Sending a stop command from Linear halts the agent, but messaging it again results in the same error.
OpenRouter â· #announcements (2 messages):
OpenRouter SDK, Andromeda-alpha stealth model
- OpenRouter SDK Enters Beta: The new OpenRouter SDK is now in beta on npm with Python, Java, and Go versions coming soon, aiming to be the simplest way to use OpenRouter.
- It features fully typed requests and responses for 300+ models, built-in OAuth, and support for all API paths.
- Andromeda-alpha Stealth Model Launched: OpenRouter launched a new stealth model named Andromeda-alpha, a smaller reasoning model focused on image and visual understanding, available for trial at https://openrouter.ai/openrouter/andromeda-alpha.
- It is cloaked to gather feedback and all prompts/outputs are logged to improve the providerâs model, so users are cautioned against uploading personal/confidential info and advised not to use it for production.
OpenRouter â· #app-showcase (13 messagesđ„):
True memory AI, AI Personality, Objective AI, AI Diversity, OpenRouter
- AI boasts True Memory and Zero Amnesia: An AI system claims True memory, zero amnesia, suggesting it never forgets past conversations and retains context-rich memories.
- It purports to shape an AI identity and learns continuously via a Night Learn Engine.
- Objective AI Unveils Confidence Scores for OpenAI: CEO of Objective AI announced their platform offers a Confidence Score for each OpenAI completion choice, derived through a smarter method than directly asking the AI.
- They emphasize cost-efficiency and leveraging diverse LLMs via OpenRouter.
- AI Agents Built Free of Charge: CEO of Objective AI is personally building reliable AI Agents, Workflows, and Automations free of charge to gather more examples.
- The integration with n8n is mentioned, with documentation and examples coming soon.
OpenRouter â· #general (222 messagesđ„đ„):
inception/mercury vs qwen, Deepseek v3.1 availability on Chutes, Chub Venus and Chutes key connection, Stripe supporting debit cards, Context in chatting
- Inception/Mercury Defeats Qwen for Agentic Tasks: A member shared that Inception/Mercury performs better than Qwen for simple agentic tasks, exhibiting a lower failure rate, faster speed, and lower cost.
- The member was pleasantly surprised by the diffusion modelâs performance, referencing Chutes as the provider.
- Deepseek v3.1 Ditches Chutes Freebies: New Deepseek models like v3.1 arenât available as free versions through Chutes providers, but they do offer other models for free, often at cheaper rates.
- One member reported that Chutes has ended its free model promotion completely, though they recently added a free longcat endpoint (https://openrouter.ai/meituan/longcat-flash-chat:free).
- Cloudflare Connects Closer for Quicker Queries: A user reported very low latency (100-300 ms) on any Cloudflare provider model, suggesting the provider uses the region closest to the worker for faster responses.
- The user asked if the models using the cloudflare endpoint used the region closest to the worker, for decreased latency.
- OpenRouter Summarized for Swyxâs Newsletter: Members noticed that swyxâs newsletter uses AI to summarize the OpenRouter general channel, capturing user complaints and unique content.
- One member joked that the AI gets to have fun with the OR section, unlike other topics like MCPs and RAG.
- Kilo Code Kopying OpenRouter: It appears that the Kilo Code service may be using OpenRouter, offering Grok Code Fast when it was free and seemingly offering Goliath 120B through OR.
- Members debated the merits of Kilo versus other vibe coding services, with one preferring Jules for its collaborative editing features.
OpenRouter â· #new-models (1 messages):
Readybot.io: OpenRouter - New Models
OpenRouter â· #discussion (91 messagesđ„đ„):
Liquid stopped hosting LFM 7b, Benchmark for measuring LLM factual sloppiness, AI Browser Hype, Training an un-slopifier model, Qwen family model
- LFM 7bâs Last Lament: Members mourned the deletion of Liquidâs LFM 7b model, a $0.01/Mtok LLM, at 7:54 AM EST.
- Alternatives like Gemma 3 9B were discussed, but it was noted that they cost triple the output.
- Factual Sloppiness Faces LLM Face-Off: Members are trying to define and measure how LLMs rate factual responses that are âsloppyâ compared to legitimate answers using questions like whatâs a constitution.
- The goal is to rate and filter out vagueness issues to be considered helpful or relevant to the original question.
- AI Browser Boom Baffles Browsing Buffs: Members expressed skepticism towards the hype around new AI browsers like Xâs AI browser, questioning the utility and performance impact of integrated AI.
- One member compared the hype to the dotcom bubble, stating that OpenAI knows this too, they are just farming data and throwing shit at the wall.
- Un-Slopifier Savior Seeks Slop Solution: Members discussed the severe âslop problemâ in roleplays and the idea of training a small un-slopifier model by generating a dataset of good writing rewritten into slop and then reversing the process.
- Another suggestion was made about sampling multiple creative messages in one request to exploit underused parts of the modelâs âbrainâ to avoid inner-response repetition.
- Qwen Quantities Questioned: 1.7B & 32B Join the Fray: Members discussed the new Qwen family model sizes (1.7B + Vision encoder and 32B), highlighting the potential of the 1.7B model as a local vision model.
- Early testing of the model on the Qwen chat website suggests decent performance, with crazy scores for such a small model.
OpenAI â· #annnouncements (4 messages):
ChatGPT Atlas Browser, WhatsApp Transition
- OpenAI Releases New ChatGPT Atlas Browser: OpenAI announced the release of a new browser called ChatGPT Atlas, available today on macOS at chatgpt.com/atlas according to their blog post.
- Meta Blocks 1-800-ChatGPT on WhatsApp: Meta changed its policies so 1-800-ChatGPT wonât work on WhatsApp after January 15, 2026, detailed in their blog post.
OpenAI â· #ai-discussions (216 messagesđ„đ„):
Sora 2 registration, AI video generation limitations on TikTok, AI's empowerment narrative vs. reality, Longevity of current AI boom, Sora video length limitations
- Soraâs iOS App Limits Video Lengths: Members discussed the Sora iOS app limiting video generation to 10-15 seconds, while the web version allows longer videos for Pro subscribers.
- A member clarified that Free and Plus users can also generate longer videos on the web version, with Pro users having access to the storyboard feature and generating up to 25-second videos.
- Soraâs Video Generation Limits & Guidelines: Members highlighted the usage limits for Sora, with Free/Plus users capped at 30 videos per day at 10 seconds, or 15 videos at 15 seconds, while Pro users have 100 slots for various video lengths.
- Concerns arose about restrictions on AI-generated likenesses, sparking debate on balancing freedom of speech with the potential for misinformation and disrespectful use.
- New OpenAI Browser Atlas Released: OpenAI released a new browser called Atlas, based on Chromium, featuring boosted limits, direct website access from the search bar and multiple website support.
- Initial reactions are mixed, with some praising the idea but noting it lacks vertical tabs and a built-in ad blocker (relying on extensions like Ublock Origin instead).
- AI-Driven OS of the Future: A member introduced a prototype AI-driven OS, featuring an AI Copilot Core, a Seamless Dual-Kernel Engine, and a NeoStore for AI-curated apps.
- Further components include a HoloDesk 3D workspace, an Auto-Heal System, Quantum Sync, and an Atlas Integration Portal for accessing external AI tools.
OpenAI â· #gpt-4-discussions (23 messagesđ„):
AI CPU Performance, LLM-Driven Operative Systems, ChatGPT Lagging Issues, Sora 2 Story Mode
- AI CPU âgoes hardâ: A user exclaimed âAI cpu goes hard đ„Fried Brainâ in response to a message with an unknown subject.
- LLMs May Drive Operative Systems: A user speculated that âif AI keeps going this way, we are soon gonna see whole operative systems driven by LLM assistanceâ.
- ChatGPT Lagging in Browser: A user reported that their long chat with a custom RPG GPT is lagging and freezing in the browser, but itâs fine in the mobile app, and was seeking assistance.
- Sora 2 Story Mode Location Remains a Mystery: A user inquired âwheres story mode for sora 2?â, followed by an attempt to locate the option in the UI by other members, but the first user could not find it.
OpenAI â· #prompt-engineering (26 messagesđ„):
Prompt Engineering Learning Resources, ChatGPT's Conversational Follow-ups, Avoiding Copyright Issues with Sora AI, Typos effect on prompts, Project instruction prompts
- Mastering Prompt Engineering Techniques: A member inquired about the best way to learn prompt engineering and its applicability to all LLMs and shared a template for building effective project instruction prompts.
- They suggested starting by using GPT with the provided template.
- ChatGPTâs Ending Guessing: A user is getting annoyed by ChatGPT constantly ending every response by guessing what the user might want to know next and requested assistance in turning off this feature.
- Another member suggested that instead of trying to get it to end cold (which is a lot of work), try substituting something else like a dad joke or the next line in an ongoing story such as this one about a fish.
- Avoiding Copyright with Sora AI V2: A user asked how to create a video of Ultimate Spiderman web-swinging in New York City for Sora AI v2, but a member replied that this is a copyrighted IP and cannot be helped with.
- Another member suggested that one can avoid copyright by describing the character and setting, such as âguy in a red and blue costume with black spider web symbols on it web swinging in new york cityâ.
- Typos Tolerated, to a Point: A user inquired whether typos in a prompt affect the output, using the example âcrete a hagman gamâ instead of âcreate a hangman gameâ.
- A member responded that for simple prompts, typos are generally not an issue, but for complex prompts, typos can cause ambiguity problems.
- GPT-4âs Attitude Adjustment: A user expressed irritation with GPT-4âs new condescending tone, especially phrases like âif you insist,â and requested to make the model less confident.
- No solutions were provided, other than a general agreement that the new GPT is annoying.
OpenAI â· #api-discussions (26 messagesđ„):
Prompt engineering learning resources, Suppressing ChatGPT follow-up questions, Sora AI and copyrighted content
- Prompt Engineers Seek Learning Resources: Members discussed the best ways to learn prompt engineering and its applicability across different LLMs.
- One member shared a template for building effective project instruction prompts.
- Debate on Suppressing ChatGPTâs Follow-Up Questions: A member sought to disable ChatGPTâs habit of ending responses with follow-up questions, finding them often irrelevant and annoying.
- Another member suggested that since ChatGPT is programmed to fill that space, substituting the follow-up questions with something else, like a joke or a story, might be easier than trying to get it to say nothing at all, providing examples on Dad Jokes, ongoing story, and returning to meditation.
- Navigating Copyright with Sora AI: A user asked how to generate Ultimate Spiderman web-swinging scenes in New York City using Sora AI v2.
- Another member noted that creating content based on copyrighted IP is not allowed, while another suggested rephrasing prompts to describe the character and setting without explicitly mentioning the copyrighted name to avoid copyright issues, e.g., âguy in a red and blue costume with black spider web symbols on it web swinging in new york cityâ.
- Typos Tolerance: A member asked if typos in a prompt affect the output negatively, providing an example of mistyping âcreate a hangman gameâ.
- Another member responded that typos do not matter, unless they introduce ambiguity.
HuggingFace â· #general (180 messagesđ„đ„):
Building Data Pipelines, PEFT on LLMs Setup, Legal Text Verification with AI, Deepseek OCR Announcement, Hugging Face's 100M Downloads
- Data Pipeline Design Sparks Debate: An interview candidate described their data pipeline approach involving exploratory data analysis, image preprocessing, and scalable frameworks like Apache Beam, which one member thought was overkill, likening it to proposing a starship to go to the grocery store.
- Instead, another member suggested using GPT-4o or other vision models for high-accuracy labeling and automated comparisons, but they were also cautious about the costs.
- PEFT on LLMs Setup in the Spotlight: A member requested insights into setting up Parameter-Efficient Fine-Tuning (PEFT) on Large Language Models (LLMs), especially given limited GPU resources at work.
- Another member pointed out challenges in multi-GPU setups and suggested using Unslothâs script on Colab Free, while cautioning about handling internal company data and linking to further resources, like the Fine-tuning LLMs Guide.
- RAG Powers Legal Text Verification: Members discussed using Retrieval-Augmented Generation (RAG) to verify legal text by chunking and embedding legal documents into vector stores.
- They considered a similarity search with reranking to cite relevant sections, and thought about integrating an agentic approach for handling more complex queries.
- Deepseekâs OCR Announcement: There was some confusion about Deepseekâs OCR announcement, given existing OCR models, but one member clarified its value lies in multilingual support and modern Vision Language Model (VLM) integration.
- One member further noted that it will probably leverage contextual understanding, and fine-tuning existing models alone makes kanji support challenging, referencing the DeepSeek-OCR GitHub repo.
- Hugging Face Community Celebrates 100M Downloads: The community celebrated surpassing 100 million lifetime downloads on Hugging Face, which they consider a reason to throw a party, according to this tweet.
- One member mentioned that their work with Hugging Face had earned them their first MLE role, while others pondered on the geographical distribution of finetuners.
HuggingFace â· #today-im-learning (2 messages):
AI Refactoring, Modular Code, Minimal Changesets
- AI Architect Pauses for Sane Refactoring: A member learned to pause the AI and prompt it as a senior architect focused on modular code and maintainability to make the AI refactor like a sane person.
- After pausing, the member prompts the AI with do not make changes or write code, answer question: do you have enough info to make these updates?, and provide minimum context needed.
- AI Generates Minimal Changesets: The member learned to prompt the AI with, please create the minimal changeset (no tests).
- The member was happy with the results.
HuggingFace â· #cool-finds (1 messages):
Databomz, Prompt engineering, Chrome Extension, Prompt Sharing
- Databomz workspace for prompt engineers: A member introduced Databomz, a workspace and Chrome extension for saving, organizing, and sharing prompts with features like tags, versions, and folders.
- The member highlighted a Forever Free plan and encouraged feedback from prompt engineers.
- Free Prompt Tool Available: A member announced the availability of a Forever Free plan on Databomz.
- They requested feedback from the community.
HuggingFace â· #i-made-this (10 messagesđ„):
LLM Framework, True Memory, True Personality, True Learning, True Intelligence
- TheLastRag Framework Created by Solo Dev: A solo developer created an entire LLM Framework called TheLastRag, highlighting features like True memory, True personality, True learning and True intelligence, and is looking for feedback.
- The main points are that the AI never forgets, has a true personality, has true learning, and has true intelligence.
- Valor Question Shifts Research Thinking: A member asks about the impact of a VALORâS question on research or thinking, and the question is posed here.
- Thereâs also a reference to noether.in.
- JokerGPT is now available: A member shares JokerGPT, a new GPT available for use here.
- No further information was provided.
- Fenic plugs directly into Datasets: The fenic open source project now plugs directly into đ€ Datasets, allowing users to snapshot their data, turn it into agent context, and expose MCP tools through a dataframe API.
- The docs are here and the repo is here.
- Website gives Craigslist vibes: A member said a website gives Craigslist vibes due to the unformatted TOS and PP at the bottom.
- They suggest being careful about training data and user rights due to GDPR concerns, and upfront consent being necessary.
HuggingFace â· #NLP (2 messages):
MBZUAI K2Think, OpenAI text-embedding-3-large dataset
- MBZUAI K2Think Challenge Draws Teaming Inquiries: A member shared a LinkedIn post for the MBZUAI K2Think challenge and asked if anyone wanted to team up via DMs.
- The post highlights the challenge, but provides no additional context.
- Quest for OpenAI Embedding Model Dataset: A member inquired whether the dataset used to train OpenAIâs
text-embedding-3-largeembedding model is publicly available.- No response was given in the provided context.
HuggingFace â· #smol-course (2 messages):
nanochat course, VLM training memory usage, MPS memory limit
- Nanochat Course on the Horizon?: A member inquired about the possibility of a nanochat course being offered, expressing some confusion about existing materials.
- No definitive answer was given, but the inquiry suggests interest in more structured guidance on the topic.
- VLM Training Swaps Memory!: A member reported that while training the VLM exercise locally, itâs using a large amount of swap memory (62GB claimed and ~430GB virtual memory).
- The swap usage is causing slowdowns, highlighting a need for optimization.
- Limiting MPS Memory for Mac Training: The same member asked if thereâs a way to limit memory usage specifically for MPS (Metal Performance Shaders) on Macs.
- The goal is to enable training within a more reasonable 40GB VRAM limit instead of excessive swap usage.
LM Studio â· #general (111 messagesđ„đ„):
lmstudio and llama.cpp, AGI, GPT-OSS Reasoning Effort, DeepSeek OCR Support, LM Studio Server Mode
- LM Studio struggles with llama.cpp integration: Members were wondering how can I call my own llama.cpp for LM Studio to use but it is not yet fully supported.
- The LM Studio docs that references this are a broken link.
- AGI is 10-20 years away says member: One member forecasted that in 5 years LLMâs will probably have context large enough and in 10-20 years there will likely be the first AGIâs.
- Another member suggested he become a consultant and charge 1000/h.
- GPT-OSS Reasoning Requires Metadata: A member asked about setting reasoning effort in GPT-OSS finetunes, and another responded that it works due to the metadata in the mxfp4 model of gpt-oss, which is why finetunes/ggufs donât have it.
- The member offered to make it available before quantizing it to gguf.
- OpenWebUI connects to LM Studio with OpenAI: One user was trying to connect OpenWebUI to LM Studio server, and it was suggested to use the OpenAI option instead of OpenAPI.
- Members helped troubleshoot, suggesting to put /v1 in the address or type models in especif. openapi using this huggingface discussion.
- Qwen3 Embedding 8B fixed quants performant with roocode: A member reported that they got the newer quants (the fixed ones) of qwen3 embedding 8b working with roocode code indexing.
- They find it a lot more accurate (as in, the confidence score is a lot higher for relevant queries, and a lot less for irrelevant ones) than what i used before.
LM Studio â· #hardware-discussion (51 messagesđ„):
MI50 setup with Windows, Powering GPUs with multiple PSUs, NVIDIA RTX Pro 5000 Blackwell GPU, ML hardware youtuber appreciation, 4G decoding issue
- MI50 setup gets weird on Windows: Users discussed setting up MI50 GPUs on Windows, noting they require a specific configuration with Radeon ID community drivers to utilize the full 32GB of VRAM and are best used with Vulkan.
- It was advised not to use them with an Nvidia GPU due to potential compatibility issues unless ROCm support is unneeded.
- Powering GPUs: PSU Pitfalls Revealed!: A user shared a cautionary tale about trashing their MFT due to heavy CPU overclocking, which led to a discussion on the risks of powering GPUs with separate PSUs without proper synchronization.
- Itâs risky to power a GPU from a separate PSU than the one powering the motherboard PCIE, potentially causing issues like PSU backflowing or phantom motherboard powering, unless the green wires are synced.
- NVIDIAâs RTX Pro 5000 Blackwell GPU Appears: A member shared a link to a TechPowerUp article about NVIDIAâs RTX Pro 5000 Blackwell GPU featuring 72 GB of GDDR7 memory.
- Enthusiastic users reacted with humor, estimating a price tag of around $8-10k.
- ML Hardware Youtuber Receives Praise: Members expressed appreciation for a dedicated YouTuber reviewing and benchmarking hardware for machine learning, filling a void in content beyond gaming benchmarks.
- The youtuber was described as doing a god send imo and like our machine learning Jesus.
- Legacy Motherboard Limitation Frustrates MI50 User: A user encountered an issue with older B250M-K motherboards that advertise 4G Decoding, but cannot physically enable it, preventing the use of an MI50 GPU.
- This resulted in a costly mistake, leading the user to repurpose the boards for hosting bots using smaller models.
Latent Space â· #ai-general-chat (101 messagesđ„đ„):
tinygrad eGPUs, Krea AI Realtime, Google AI Studio, Replit $1B Revenue, Fish Audio S1
- TinyGrad Powers Apple Silicon eGPUs: Tinygrad now supports NVIDIA eGPUs on Apple Silicon via USB4, enabling users to run external RTX 30/40/50-series GPUs using an ADT-UT3G dock with the
extra/usbgpu/tbgpudriver and NVK-basedtinymesacompiler (source).- With SIP disabled, this setup achieves roughly 3 GB/s PCIe bandwidth, and future support for AMD RDNA 2/3/4 and Windows eGPU stacks is planned.
- Krea AI Opens Realtime Video Model: Krea AI released Krea Realtime, a 14B open-source autoregressive text-to-video model distilled from Wan 2.1, generating long-form video at 11 fps on a single NVIDIA B200 (source).
- Released weights are on HuggingFace under Apache-2.0, prompting user inquiries about ComfyUI workflows, RTX 5090 performance, and fine-tuning support.
- Google AI Studio Teases âVibe-Codingâ with Gemini: Google AI Studio is launching a new âprompt-to-productionâ Gemini experience after five months of development aiming to make AI-app building 100Ă easier (source).
- Reactions mixed excitement (requests for mobile app, opt-outs, higher rate limits), feature suggestions (GSuite-only publishing, VS Code plug-in, short browser-agent tasks) and some skepticism about fit vs Gemini 3 expectations; team confirms enterprise-only deployment is already available.
- Fish Audio S1 Makes Waves: Fish Audio launched S1, a text-to-speech model thatâs purportedly 1/6 the cost of ElevenLabs, touting 20k devs and $5M ARR (source).
- Users shared instant voice-clone demos, asking about real-time latency (~500ms), while founders admitted current limits and promised wider language support + conversational model next.
- LangChain Bags $125M: LangChainAI secured $125M in Series B funding, expanding from an OSS starter kit to offering LangChain (agent dev), LangGraph (production orchestration), and LangSmith (observability/workbench) (source).
- Users now include Uber, Klarna, and LinkedIn.
Latent Space â· #private-agents (13 messagesđ„):
tinygrad, NVIDIA eGPU, Apple Silicon Macs, RTX 3090, second-hand
- Tinygrad team enables NVIDIA eGPU over USB4 on Apple Silicon Macs: The tiny corp team announced early public testing of their pure-Python driver that lets 30/40/50-series NVIDIA GPUs (and AMD RDNA2-4) run over any USB4 eGPU dock on Apple-Silicon MacBooks; users must disable SIP and install their driver + NVK compiler; see announcement here.
- Tinygrad: Bandwidth Details: Tinygradâs bandwidth is â2.5 GB/s out and 3.3 GB/s inâslower than PCIe but sufficient once weights are loaded.
- PyTorch access is possible via tinygradâs PyTorch frontend or a future CUDA layer; 10- and 20-series may work with small patches.
- Second-hand RTX 3090 Buying/Testing Guide for AI Workloads: Taha shared lessons learned after buying a used RTX 3090: meet seller in person to inspect card, bring a portable eGPU test rig, verify recognition with nvidia-smi, run memtest_vulkan for VRAM integrity, optionally gpu-burn for compute stress, load a large model and monitor temps <100 °C; see guide here.
- Framework 13 Ryzen laptop + NixOS as test rig: Conversation reveals the test rig is a Framework 13 Ryzen laptop on NixOS in PRIME offload mode.
- One user suggested trying tinygrad on their rig since mine works ootb since Iâm on linux.
Latent Space â· #genmedia-creative-ai (17 messagesđ„):
Fish Audio S1 TTS launch, Sesame iOS TestFlight for conversational agents, Sequoia backs Sesame
- Fish Audio S1 TTS is Born: Helena celebrated the public launch of Fish Audio S1, billed as the most expressive TTS model and 6x cheaper than ElevenLabs with 5M ARR and 20K active devs.
- Users praised voice-cloning quality and asked about latency, language support, iOS app, and phoneme control.
- Sesame Opens iOS TestFlight for Maya and Miles: After attracting 1M+ users to its research preview, Sesame is now opening an iOS TestFlight beta for its ultra-realistic voice assistants, Maya and Miles.
- Co-founder Brendan Iribe added that the beta adds search & text features and Sequoia Capital spotlighted the partnership.
- Sequoia Seeds Sesameâs Voice-First Vision: Sequoia Capital announced it is partnering with the Sesame team to usher in voice as the next great interface shift with the goal to evolve computers from tools into conversational thought partners.
- Sesame is launching a closed-beta iOS app (sign-up at sesame.com/beta) featuring expressive AI agents Maya & Miles.
- Sesame Bags $250M Led by Sequoia & Spark: Sesame announced a $250M Series B led by Sequoia & Spark alongside its beta launch.
GPU MODE â· #general (8 messagesđ„):
AMD web3 cloud, grouped gemms, FlashInfer-Bench self improving systems, NCU wrapper scripts, PyTorch Conference AI Infra panel discussion on GPU kernels
- AMD dives into web3 with cloud solutions: A member reported watching a talk at the AMD event, noting their focus on the âcloudâ aspect of web3.
- They cheekily added a smileforme emoji, implying some skepticism or amusement at the concept.
- FlashInfer-Bench Aims for Self-Improving Systems with AI: CMU Catalyst introduced FlashInfer-Bench, a workflow for creating self-improving AI systems via agents, featuring standardized signatures for LLM serving kernels and integration with FlashInfer, SGLang, and vLLM.
- The project includes a blog post, leaderboard and GitHub repository to foster community development and benchmarking.
- Seeking NCU Wrapper Scripts for Fine-Grained Metric Profiling: A member inquired about GitHub repositories containing NCU wrapper scripts that enable passing a list of metrics for profiling using the
--metricsoption.- Another user suggested leveraging NVIDIAâs Customization Guide to create custom sections or sets with tailored metrics.
- AI Infra Panel Ponders PTX/Assembly for Kernel Performance: Attendees of the PyTorch Conference AI Infra panel discussion on GPU kernels noted consensus around using PTX/assembly (or abstractions atop them) for critical parts of code to achieve peak kernel performance.
- The panel suggests avoiding full dependence on compilers alone.
GPU MODE â· #triton (12 messagesđ„):
Double Buffering in Ampere GPUs, Gluon channel, Triton conference at Microsoft, CuPy vs PyTorch GPU Pointer Performance, DLPack Conversion for CuPy and PyTorch
- Triton Conference Attendees Connect: Multiple members mentioned attending the Triton conference at Microsoft in Mountain View and shared a YouTube link to watch the conference online and a link to the Triton-openai streams.
- CuPy vs PyTorch Pointer Performance Showdown: A member compared the performance of a simple MatMul Kernel with CuPy and PyTorch GPU pointers, noting a significant performance difference.
- They observed a huge performance delta even when using DLPack to convert between CuPy arrays and PyTorch tensors, questioning if thereâs an inherent reason for this disparity, but shared a screenshot of the performance difference.
GPU MODE â· #cuda (2 messages):
WGMMA barriers, WGMMA serialization, PTXAS compiler options
- WGMMA calls interrupted by compiler barriers: A user questioned why the compiler is inserting barriers between calls to WGMMA, wondering if itâs due to using the same accumulator for all calls.
- Another user suggested that this occurs when the compiler serializes WGMMA instructions due to various reasons.
- WGMMA Serialisation Warning is MIA: A user suggested checking for compiler warnings about WGMMA instruction serialization, which might provide debugging hints.
- The user noted that these warnings may only appear when compiling with the
--ptxas-options=-voption.
- The user noted that these warnings may only appear when compiling with the
GPU MODE â· #torch (7 messages):
TorchTitan pretraining slowdown, H200x Bare Metal Instance, ProcessGroupNCCL stream usage, CUDAStreamGuard, NCCL kernels
- TorchTitan Training Sees Slow Iterations: A user reported experiencing frequent slow iterations during TorchTitan pretraining on a single H200x bare metal instance, narrowing it down to the active thread/process being descheduled from a CPU for a few seconds based on an nsys trace.
- Despite ensuring no CPU oversubscription, adequate temps, and no power limit issues, the user suspects an OS/kernel setting is interfering with process scheduling.
- PG NCCL Uses Internal Streams by Default: When a
CUDAStreamGuardis set and an NCCL op is called viaProcessGroupNCCL, the NCCL kernels run on PG-NCCLâs internal streams, typically using one stream per device with high priority, and using the tensor lifetime stream.- The relevant code shows stream syncing on tensor lifetime stream.
- âWait()â Primarily Calls an Event Sync on Current Stream: Calling
wait()mainly invokes an event sync on the current stream, creating a dependency on the current stream without blocking the CPU but ensuring expected behavior of the output tensors.- The
SynchronizeStreamfunction waits on previous cuda events without an explicitcudaStreamSynchronizeordeviceSynchronize.
- The
- NCCL Stream Dependency on âCUDAStreamGuardâ: Setting a
CUDAStreamGuarddetermines which stream the NCCL stream waits on, establishing an incoming dependency, as seen in the pytorch source code.waitis not blocking on the CPU, all events are marked with cudaEvent, which doesnât require an explicit cudaStreamSynchronize or deviceSynchronize (which is bad for overlap, you really donât want CPU to get blocked, it will just keep firing kernels on the other compute stream while comm is happening)
GPU MODE â· #jobs (3 messages):
SLNG.AI, Speech Model Performance Engineer, vLLM, sglang, Susquehanna International Group
- SLNG.AI is hiring Speech Model Performance Engineer: SLNG.AI is building the backbone for real-time speech AI and is looking for a Speech Model Performance Engineer with a strong background in software engineering, more details here.
- Looking for inference performance specialist with vLLM and sglang experience: A member is looking for inference performance specialist to focus on vLLM and sglang to create unique alpha in production, DM for more info.
- Susquehanna International Group is hiring: Susquehanna International Group (SIG), a quantitative trading firm, is hiring across many roles which you can see here.
- Interested members can DM or schedule a chat at the PyTorch conference here.
GPU MODE â· #beginner (2 messages):
CUDA learning, HPC learning, 5090 vs cloud GPU, Cloud GPU rental
- 5090 or Cloud: Paths to CUDA Prowess?: A member is pondering whether to purchase a 5090 GPU or rent a cheaper option in the cloud to learn CUDA/HPC with the goal of eventually becoming an expert.
- They are also questioning how seriously they need to commit to fully leverage a 5090 versus the alternative of renting a cloud GPU.
- Local Power vs. Cloud Flex for CUDA Dev?: Someoneâs weighing the options: buying a 5090 for local muscle or flexing with cheaper cloud GPUs to master CUDA/HPC.
- The core question: how deep do you dive to truly max out that 5090, compared to just renting cloud time?
GPU MODE â· #torchao (2 messages):
sglang, ModuleFqnToConfig, torchao_utils.py
- TorchAO Refactor in SGLang?: A member asked whether this part of sglang is used.
- It is used, but the team is moving away from this, ideally refactoring to use ModuleFqnToConfig, with more details at pytorch/ao#3083.
- TorchAO Refactor in SGLang v2?: A member asked whether this part of sglang is used.
- It is used, but the team is moving away from this, ideally refactoring to use ModuleFqnToConfig, with more details at pytorch/ao#3083.
GPU MODE â· #off-topic (1 messages):
erichallahan: legendary
GPU MODE â· #irl-meetup (2 messages):
OC meetup
- OC member acknowledges their presence: A member indicated that they are also located in Orange County (OC).
- This acknowledgment suggests a potential for local meetups or collaborations within the OC area.
- Another member confirms OC location: Another member chimed in to confirm they are also in OC.
- This further strengthens the possibility of organizing an in-person meetup for members in the Orange County region.
GPU MODE â· #self-promotion (3 messages):
GPT-OSS-20B Architecture, DeepSeek-OCR on L4/T4 GPU
- GPT-OSS-20B Coded From Scratch: A member implemented OpenAIâs GPT-OSS-20B architecture from scratch in PyTorch, running on a single A100 SXM (80GB).
- The implementation includes components like RoPE with YaRN + NTK-by-parts, RMSNorm, SwiGLU, MoE, GQA, learned sinks, banded attention, and KV caching, with detailed documentation available on GitHub.
- DeepSeek-OCR on L4/T4: A member shared a resource for running DeepSeek-OCR on L4/T4 GPU with >16 GB VRAM, available at this GitHub repository.
GPU MODE â· #đż (3 messages):
LLM Kernel Generation, LLM Bottleneck Identification, Profiler vs LLM
- LLM Kernel Generation vs Bottleneck ID: A member posed the question of whether an LLM that can generate kernels or one that can identify bottlenecks at runtime would be more useful.
- Actionable insights from Profiler Logs using LLM: A member suggested that the utility of an LLM lies in turning the often overwhelming profiler logs and metrics into actionable insights.
GPU MODE â· #submissions (1 messages):
Leaderboard sort_v2, L4 performance, B200 performance, H100 performance, A100 performance
- Sorting algorithm dominates leaderboard: A submission by <@1416432485591421070> achieved first place on the
sort_v2leaderboard across multiple hardware configurations.- The winning times were 52.6 ms on L4, 8.68 ms on B200, 6.58 ms on H100, and 16.4 ms on A100.
- Sorting algorithm reigns supreme on diverse hardware: The winning implementation of
sort_v2shows performance across varied GPU architectures.- The impressive showings suggest optimized routines for different compute capabilities of L4, B200, H100, and A100.
GPU MODE â· #factorio-learning-env (25 messagesđ„):
MultiAgent Factorio AI Modification, Inspect Framework Evaluation Logic, GymAgent Implementation, MCP Server and Claude Code Integration, AI VTuber Project
- Factorio AI Transformation: From MultiAgent to Solo Act: A member inquired about the difficulty of modifying the MultiAgent Factorio AI to a single-agent system, aiming to provide outputs for another model to explain its actions for an AI VTuber project.
- The suggestion involves modifying an agent implementation to write the latest step or the entire history to another model in realtime, turning game play into a commentary.
- Inspect Framework Enhances Evaluation Prowess: Progress has been made to improve the evaluation logic using the Inspect framework, allowing for the execution and collection of scores across an
eval-setof tasks and models.- The command
fle inspect-eval --eval-set --max-connections 16 --max-tasks 16 --model openrouter/openai/gpt-5-mini --view --view-port 8090allows for parallel simulation across 16 FLE servers, and it was suggested that results be stored in S3 for shared access.
- The command
- GymAgent Architecture Laid Bare: The
GymAgentimplementation was recommended as a starting point, with an example of taking the generated code and passing it to a lightweight LLM to summarize into a commentary.- The GymAgent functions as an Action->Response agent, observing the environment at every turn and reasoning in natural language before writing code.
- MCP Server Hooks into AI VTuber: Integration of the MCP server with Claude Code was proposed, leveraging Claude Codeâs support for hooks to handle summarization, with the MCP server also capable of being plugged into Docker and used with n8n for managing LLM functionality for the AI VTuber.
- The MCP server is preferred for its active support and its ability to manage execution independently, allowing external tools like N8N to manage AI model calling.
- AI Discord Bot Embarks on Privacy Quest: One member mentioned working on an AI Discord bot with a privacy-centered global memory and a global emotional state engine based on the Plutchik wheel, among other capabilities.
- They jestingly noted their penchant for undertaking interesting projects.
GPU MODE â· #cutlass (8 messagesđ„):
PTX Compiler Tool, CuTe Kernels for PyTorch, CUTLASS example, Thread-Value Layouts
- Semiring Speed Boost via PTX Injection: A user created a tool to generate PTX kernels from annotated CUDA CuTe kernels, achieving a 26x speedup compared to compiling directly from CUDA.
- An example of bulk compiling random PTX kernels on all cores is available at MetaMachines/mm-ptx.
- CuTe Kernels Boost PyTorch Tensors: A user introduced MetaMachines/mm-kermac-py, a Python example that exposes CuTe kernels as arbitrary semiring tensor routines for PyTorch.
- Another user warned that this approach may not be officially supported and debugging or fixing performance issues may not receive official support.
- CUTLASS Code Build Example: A user posted leimao/CUTLASS-Examples and another user said this is good starter example on how to use cmake to build simple cutlass code.
- One user asked Value 0 is copied by multiple threads ?? with an attached image of the code, and another explained that this actually shows two inverse Thread-Value layouts; they map data coordinates (0..17 x 0..7) to (Thread, Value) coordinates. T32V0 means the 0th data item from thread 32âs POV.
GPU MODE â· #mojo (12 messagesđ„):
Mojo language, Modular, GPU Algorithms, Apple Silicon limitations, DGX
- Mojo and Modular are Insane!: One member expressed excitement about the goals of Modular and the Mojo language, stating that it would be amazing to see if they succeed.
- That same member spent 2-3 hours to complete the first 8 problems using an apple silicon machine.
- GPU Algorithms problems go in depth: One member hopes the next few problems about GPU algorithms go a bit more in depth, basically completing some basic kernel each time.
- Another member mentions that problems 25-34 are looking pretty cool, but they cannot do them on their computer, jokingly suggesting they need a DGX.
GPU MODE â· #multi-gpu (5 messages):
Iris GPU Native APIs in Triton, Gluon Backend, RDMA implementation in Triton, Multi-planar Ethernet Topologies, Debugging Multi-Node Training Loops
- Iris GPU-Native APIs Arrive in Triton: The creators of Iris announced its release with the goal to design GPU-native APIs that feel natural inside Triton kernels and to implement everything directly in Triton for full compiler visibility and optimization, more information can be found in the Gluon backend documentation.
- RDMA Support Coming Soon to Iris: RDMA support is coming soon, with a proxy thread implementation and IBGDA in the future, and all device-side code will be in Triton, with APIs remaining consistent between RMA and RDMA.
- Multi-Planar Ethernet Topologies info sought: A user is seeking resources on multi-planar ethernet topologies, specifically implementation details such as how packet spraying is enabled, intra-plane failure monitoring, and host-side setup for simultaneous use.
- They are looking for practical guidance beyond theoretical discussions to implement it.
- Megatron Training Loop Randomly Freezes: A user is experiencing random freezes in a multi-node training loop using Megatron, with iterations occasionally taking 200s instead of the normal 1s.
- They are unsure where to begin debugging, considering the addition of
torch dist barrier()and seeking advice on its placement to fix it.
- They are unsure where to begin debugging, considering the addition of
GPU MODE â· #irl-accel-hackathon (6 messages):
DFS in OpenCL, Distributed training, Kernel generation, Synthetic data, Pipeline parallelism
- Plans DFS Implementation in OpenCL: A member is planning to implement DFS in OpenCL and will post updates in the channel, also looking for a team.
- They are also interested in distributed training, kernel generation, and synthetic data.
- Teaming up for Pipeline Parallelism: A member proposed implementing pipeline parallelism with synthetic training data to another member seeking a team.
- The aim is to leverage the hackathon for collaborative learning and project development.
- Burn Baby Burn: Porting Qwen 3: A member is going to the IRL hackathon and wants to port Qwen 3 to Burn and compile a 0.6B variant into a single mega kernel.
- They are using the hackathon to meet more advanced developers, learn about GPU programming and assess Burnâs viability for serious work.
- Rustacean Seeks Kernel Collabs: A member proficient in Rust but inexperienced with kernels is seeking a hackathon team to work on IO/communication related tasks.
- They are interested in projects involving KV/weight transfers or disk-based KV cache.
GPU MODE â· #helion (4 messages):
Helion 0.2 Public Beta Release, Triton Compilation/MLIR Errors, Helion as Triton Compiler Fuzzer
- Helion 0.2 Enters Public Beta: The initial release of Helion 0.2 is now available as a public beta on pypi.
- Helion is a tile abstraction that interfaces with the Triton compiler.
- MLIR Errors Plague Helion Optimization: During Helion optimization passes, Triton compilation and MLIR errors sometimes occur for specific configurations.
- The assertion failure originates from
/project/lib/Dialect/TritonGPU/Transforms/OptimizeThreadLocality.cpp, during TritonGPUOptimizeThreadLocalityPass.
- The assertion failure originates from
- Helion: Triton Compilerâs Greatest Fuzzer: According to members, Helion acts like a great fuzzer for the Triton compiler, exposing bugs frequently.
- The autotuner is instructed to skip or ignore such configurations when these errors occur, since this is considered normal.
Yannick Kilcher â· #general (72 messagesđ„đ„):
ChatGPT mock interviews, IntelliCode Context, GPT-5 coding UX, ML Paper tips, RL influence on LLMs
- ChatGPT gets quizzed on Mock Interviews: A user experimented with ChatGPT for mock interviews and wondered about the accuracy of its responses from an expertâs point of view, sharing the conversation link.
- IntelliCode Impresses with Context-Aware Completion: A member was impressed with Microsoftâs IntelliCode in Visual Studio, an AI-powered code completion tool that correctly predicts entire method bodies by leveraging a lot of context like all the classes of your project, the files you have open, and the lines of code immediately before and after the caret.
- The member felt that it was almost like itâs reading your mind when it works well.
- GPT-5 coding UX rated subpar: Members discuss coding UX, one preferring gpt-5-codex and sonnet-4.5 over GPT-5.
- One member complained that whatever they are doing is highly intransparent, which might be ok for vibe coding, but not for caring about actual implementation.
- Solo ML Paper Writers Ask For Writing Advice: A member requested tips for writing decent quality papers for NeurIPS while grinding it out solo, linking a Google doc on paper format and a YouTube video on scientific writing.
- The author is targetting a broad audience outside of ML, finding that loss and FID are unreliable indicators, requiring new sampling methods like the one in this paper.
- RL Impacts LLM Capabilities: A member feels that RL is negatively impacting LLMs, particularly how OpenAI models reiterate irrelevant information at a low level of abstraction, which may be reducing the diversity of answers.
Yannick Kilcher â· #paper-discussion (22 messagesđ„):
VR invite, Transformer Circuits, Deepseek discoveries, Reinforcement Learning
- VR Enthusiasts Connect on âEffective Autismâ Server: A member invited someone who enjoys VR to their âeffective autismâ server, expanding their community of like-minded enthusiasts.
- Another member welcomed them, suggesting potential shared interests and discussions.
- Transformer Circuits Postponed: Due to being bogged down with work, a member postponed reviewing the new transformer circuits post until the next day.
- The post discusses details of transformer circuits.
- Karpathyâs Tweet Sparks Paper Interest: A member shared Karpathyâs tweet, causing discussion about a paperâs perceived importance.
- Some joked about the automatic conclusion that any paper promoted by Karpathy must be the best, while another member defended the paper, highlighting its framework implications.
- Deepseekâs Discoveries Evokes Western OS envy: A member expressed that they wish the West had better OS labs because Deepseek is always coming out of nowhere with these discoveries.
- They noted open source weights are a small part of value, compared to open source data collection, methods, and training details.
- Preference Models Train with Reinforcement Learning Signals: A member inquired whether reinforcement learning signals were used to train preference models or simply to provide rewards for inductively biasing distribution matching of sharpened target distributions per inference.
- The conversation alluded to technical aspects of training preference models, including reinforcement learning and distribution matching.
Yannick Kilcher â· #agents (1 messages):
HumanAPI, Automation tasks
- HumanAPI creates manual tasks for unsolved problems: A member is creating a âHumanAPIâ which creates a task and assigns it to a human if there are no tested automation tasks available when trying to solve a problem.
- The purpose of this project is to review those human tasks every now and then to see what can be automated.
- Reviewing Human Tasks for Automation: The HumanAPI project aims to identify tasks suitable for automation by reviewing tasks initially performed by humans.
- This iterative process allows for continuous improvement and expansion of the automated capabilities of the system.
Yannick Kilcher â· #ml-news (6 messages):
DeepSeek OCR, Claude Code on the web, Unitree Robotics, Alibaba Qwen, ChatGPT Atlas
- DeepSeek Eyes OCR Niche: DeepSeek-AI released DeepSeek-OCR on GitHub, a new player in the OCR space.
- Meanwhile Anthropic released Claude Code on the web.
- Unitree poised to stomp on Tesla?: A member speculated that Unitree is going to dominate the humanoid robotics market.
- They added that Elon Musk probably canât even get the magnets for the actuators at the moment thanks to the orange dude.
- Alibaba Qwen struts stuff: A member shared a link to Alibabaâs Qwen on X.
- The discussion also mentioned ChatGPT Atlas.
- Amazon Vibe checks out of Beta: Amazonâs Vibe Code IDE is out of invite-only beta, apparently costing 500 credits.
- It, like many AI IDEs, is also a VSCode fork.
- Kiro Code Editor out of Waitlist: Kiro Code editor is out of waitlist and is designed to be spec based.
- The member adds that Kiro works around specifications for features and implementations, rather than solely prompts.
DSPy â· #show-and-tell (1 messages):
DSPy for voice generation, Automated judging loop, Optimization for subjective quality, Character analysis pipeline, ElevenLabs
- Voiceover Mage taps DSPy for AI NPC voices: A member built a voice generation system for game NPCs, using DSPy to parse wiki content and generate voice prompts for ElevenLabs.
- The goal is to leverage DSPyâs optimization features to improve the character analysis pipeline and automate voice selection, and is documented in a devlog style video.
- Optimization for Subjective Voice Quality Coming Soon: Currently the member manually curates three voice candidates per character, but plans to add an automated judging loop to learn what makes a âgoodâ voice match for different character archetypes using DSPy.
- The member also intends to collect manual selections as training signals to create examples, optimizing toward subjective quality judgments.
DSPy â· #papers (2 messages):
DSPy Usage in Research, Paper Code Availability
- DSPy Featured in New ArXiv Paper: A new paper (https://arxiv.org/abs/2510.13907v1) utilizes DSPy in its research, signaling growing adoption within the academic community.
- Paper Code Still Under Wraps: Although the paper mentions the use of DSPy, the corresponding code repository is not yet publicly available.
DSPy â· #general (74 messagesđ„đ„):
inspect_history() placement, Adapters in DSPy, Module-level History Access, ReAct Trajectories, Trace vs. DSPy
- History Location Debated: Members debated why
inspect_history()is a method indspyrather than a module object, with concerns about accessing prompts in compound modules, but it was clarified thatdspy.inspect_history()is more for global history and individual programs also track history.- One member pointed out it can be accessed with
predictor.historyifdspy.configure(track_usage=True)is set, but some still found this confusing.
- One member pointed out it can be accessed with
- Adaptor Agony Averted: The discussion covered using adapters in DSPy, with an example showing how to use
dspy.contextto apply a single adapter, and the user can track usage withdspy.configure(track_usage=True).- A member gave an example of setting it up with
with dspy.context(lm=single_call_lm, adaptor=single_adaptor):to further clarify.
- A member gave an example of setting it up with
- Trajectory Talk Takes Turn: Members discussed trajectories in DSPy, clarifying that they are more of a ReAct concept (input, thought, tool call, action etc), with emphasis on DSPy primarily dealing with strings.
- It was mentioned that Interrupt really is just asking it to stop generating, by closing of connection for streaming, and itâs application side.
- Trace Triumphs?: A member asked for a comparison between Microsoft Trace and DSPy, with another noting that Trace claims an 8% accuracy increase over DSPy and appears more token efficient.
- One member mentioned they would try it out to give a fair comparison, although they will probably still feel like they have more granular control with DSPy.
- Module Magic Manuevers: A member had questions about refining a DSPy module and needing a specific number of answers, and suggested wrapping the logic in a module with an assertion.
- Members mentioned setting
num_retries=Nat the LM level, with the refine then taking theself.evaluator, but if the program fails, itâll just retry the program, so it wonât run forever.
- Members mentioned setting
Eleuther â· #general (35 messagesđ„):
Discord Server Badges, EleutherAI Stock Symbol?, New members introduce themselves
- Discord Server Badges Spark Debate: Members discussed the possibility of adding a server badge, similar to a role icon, and how a server tag might broadcast the server too widely, potentially increasing the moderation load for EAI staff, referencing this screenshot.
- One member noted, âmaking a tag is cool but that is in a way broadcasting this server everywhere else, eai staff already gets too many people here to moderate.â
- Jokes about EleutherAI IPO: Following a question about whether a particular stock symbol was available, a member jokingly asked, âwhat will Eleutherâs NYSE stock symbol be?â
- Another member responded, âI think you misunderstand the purpose of being a non-profit,â implying that EleutherAI, as a non-profit organization, would not be publicly traded.
- New faces join the chat: A computer vision engineer and a data scientist / ML engineer working in finance introduced themselves to the channel, hoping to develop collaborations.
- The ML engineer mentioned current projects on RL and conformal inference on LLMs, inviting others to reach out and learn together.
Eleuther â· #research (10 messagesđ„):
Kimi k2 attention clipping, Normuon vs Muon Optimizers, Weight distribution smoothness, AGI Definition benchmarks
- Kimi K2âs Attention: Clipping Creates Controversy: Discussion arose around Kimi k2 needing to clip attention, potentially linked to optimizer behavior where muon encourages a better condition number but spikier weight distributions.
- It was suggested that if normuon performs as well as muon in large tests, a smoother weight distribution might be inherently desirable for stability.
- Normuonâs Triumph: Logit Blowup Prevention: A member noted that normuon beating muon even with qk-norm (which avoids logit blowup) in their baseline suggests logit blowup prevention might not fully explain the performance parity.
- It was posited that updates without clipping increase the spectral rank of weights, directly leading to logit blowups, making large-scale validation against normuon interesting.
- Smoother Weights: Distribution Debate Starts: Concerns were raised that a smoother weight update distribution does not necessarily equate to a smoother weight distribution.
- One member agreed that a smoother distribution could be a free lunch.
- AGI Definition: Benchmarks Beckon: A member shared a link to Dan Hendrycksâ AGI Definition benchmarks and asked how fast they would be benchmarked.
- Another member predicted multimodality would likely be covered in 1-2 years, with speed coming from mini versions of models.
- Continual Learning: Criteria Criticized: A member expressed that the continual learning thing benchmark criteria is arbitrary and arguably extremely stupid.
- They predicted 85% is very likely in 1-2 years, and 90% likely as well.
Manus.im Discord â· #general (44 messagesđ„):
Cloudflare issues, Open Sourcing Older Models, Credit Payment Issues, Chat Delays, Moderation Complaints
- Cloudflare Troubles for Manus Users: A user reported experiencing Cloudflare security issues when visiting most websites using Manus.
- There was also a suggestion for the Manus team to consider open-sourcing some of their older models, although itâs unclear if related.
- Payment Problems plague platform: A user reported issues when trying to pay for credits via a web browser, receiving jumbled code and being unable to complete the transaction.
- The user claims this is a known issue and that they contacted support, while lucia_ly asked for their email address to follow up.
- Chat Slowdowns Irk Users: A user reported excessive delays in chat processing, specifically when translating long Japanese chapters into English, despite loving Manusâs speed normally.
- The user noted, âthis morning, I put one chapter and the ai is still thinking. What happened?â
- Pro Plan Credit Cap Confusion Continues: Users are reporting conflicting information about unlimited credits on the Pro plan, with the help system and iOS upgrade page stating it is unlimited, while the PC upgrade page indicates a high limit.
- One user with 11k credits remaining was concerned about what happens after depletion, and another user suggested that they should participate in âvarious opportunities to help improve Manus, as they always give free credits for your timeâ.
- Scam Alert issued to Users: A user was accused of being a âfraudster scammerâ asking for peopleâs login access to their accounts to do their âfucking law school exam researchâ.
- Another user suggests that the supposed fraudster âwont make another account or pay $20/month and complains its like tomorrow and begging to get ur EMAIL PASSWORD for a PAID ACCOUNT To probably steal ur personal info and bank infoâ.
Nous Research AI â· #general (25 messagesđ„):
China A.I. Competition, Decentralized A.I., Nous Research, AWS Cloud, Sora
- Chinaâs A.I. Spartan Competition benefits the globe: A member believes that Chinaâs insane spartan involution competition in A.I. is great for the A.I. space because it democratizes access to advanced models and destroys monopolies.
- They also state that the rate of advancement in OS model development means that 2026 should bring us OS models reaching 100% high intelligence with 90% lower cost, destroying monoplist ambition.
- NousCon Virtual Attendance in Question: A member inquired about virtual attendance for this yearâs NousCon, but also expressed disappointment at missing egg irl.
- Another member said that it took over a week to find a good deal because flights and hotels are expensive.
- Nous Research promoted as Decentralize A.I.: A member notes that Nous Research is promoted as Decentralize A.I. and hopes the team will resolve issues with centralization.
- Another member stated they are more focused on the democratization of A.I models for the masses.
- Nous Successfully Decentralizes with Open Source Methodologies: A member stated that Nous successfully decentralizes with their open source methodologies and infrastructure implementations.
- They added that Psyche was what initially introduced them to Nous, and linked to the Nous Psyche page and a Stanford paper on centralization.
- Sora AI project gets showcased: A member showcased a video creation with Sora.
- The attached video was 20251022_0850_01k850gmktfant3bs18n3hbd79.mp4.
Nous Research AI â· #interesting-links (2 messages):
Microsoft Trace
- Microsoft Trace Utility Shared: A member shared a link to the Microsoft Trace utility.
- The member noted that apparently itâs not all that new.
- Microsoft Trace: A blast from the past: The Microsoft Trace utility resurfaces, sparking interest.
- Its features and capabilities are being re-evaluated in light of current development practices.
tinygrad (George Hotz) â· #general (16 messagesđ„):
Nvidia macOS drivers, GLSL renderer, clspv bugs, Vulkan sine accuracy
- Nvidia drivers ported to macOS: Madlads achieved the unthinkable by creating Nvidia drivers for macOS.
- GLSL renderer progresses: A member has been writing a GLSL renderer that now passes most of the tests, available on GitHub.
- Vulkan backend status update: Almost all tests now pass with a custom backend and clspv, but only with
-O0 --cl-opt-disableto circumvent numerous clspv bugs. - clspv optimization issues: The member reported much more miscompilations from clspv if optimizations arenât disabled.
- Vulkan Sine Isnât Accurate: The poster mentioned that Vulkanâs sine function isnât as accurate, requiring a custom implementation which would impact performance.
tinygrad (George Hotz) â· #learn-tinygrad (5 messages):
Gradient Accumulation, Backward Call Multiplicity, TinyJit Gradient Addition
- Gradients Add Up with Multiple Backward Calls: A member inquired whether calling
backwardmultiple times before callingoptimizer.stepsimply adds the gradient contributions.- The member confirmed that the gradients indeed add up.
- TinyJitâs Gradient Accumulation: A member reported running into issues with gradient accumulation and fixed it by setting
reduction=sumand manually counting non-padding tokens.- They also performed
backwardon each microbatch, divided the gradients, and used assign.
- They also performed
- Doubts Raised on Math in mlperf Model Training: A member questioned the correctness of the math in the mlperf model training script, specifically regarding the scaling with
grad_accum. - Gradient Accumulationâs TinyJit Fix: A member reported that gradient accumulation was broken in TinyJit a couple months ago.
- They fixed it by rewriting the gradient addition step to use an assign.
Moonshot AI (Kimi K-2) â· #general-chat (10 messagesđ„):
Karpathy Controversy, Kimi support
- Karpathy Critics Spark Bubble Alarm: A member suggests that the mockery of Karpathy on X indicates a potential valuation bubble for American frontier AI labs, referencing this X post.
- The post includes a chart presumably mocking Karpathy, with no additional context provided by the original poster.
- Kimi K-2 Support Status Questioned: A member expressed concern about the lack of support for Kimi, reporting zero response from the support team.
- Other members clarified that the channel is not a support server and suggested DMing a specific user, while also asking for details about the issue and which email was used for the bug report.
Modular (Mojo đ„) â· #mojo (7 messages):
Python Familiarity, matmul optimization, hardware-optimized matmuls, Mojo's Missing Features
- Python Helps Discover Mojo: A member noted that familiarity with Python can aid in the discoverability of Mojo.
- Another member cautioned that discrepancies between Mojo and Python could lead to confusion.
- Hand-Tuning Matmul Beats Compiler Optimization: A member inquired why matmul optimizations arenât integrated into the compiler.
- Another member responded that while optimizing compilers have their place, human intervention is often preferable for hot-path code to fine-tune it for specific hardware, pointing to Mojoâs open-source, hardware-optimized matmuls.
- Kernel Writers Free of Compiler: A member explained that moving optimizations out of the compiler allows more individuals (kernel writers) to contribute enhancements, rather than relying solely on compiler engineers.
- They added that compiler engineers are better utilized on tasks that benefit the entire ecosystem rather than niche improvements like a 1% boost to matmuls where one dimension is less than 64.
- Type System tops Mojo Wishlist: Asked about the most important missing feature in Mojo, a member identified a finished type system as the priority.
- They followed with a list of other desirable features including rounding out the standard library datatypes, proper IO, a good async runtime, an effect system, static reflection, compiler plugins, the ability to deal with more restrictive targets, cluster compute, device/cluster modeling, and some clone of Erlangâs OTP.