a quiet day, sort of

AI News for 9/16/2025-9/17/2025. We checked 12 subreddits, 544 Twitters and 23 Discords (192 channels, and 4174 messages) for you. Estimated reading time saved (at 200wpm): 367 minutes. Our new website is now up with full metadata search and beautiful vibe coded presentation of all past issues. See https://news.smol.ai/ for the full news breakdowns and give us feedback on @smol_ai!

Anthropic published a wonderfully in depth post mortem of their Aug-Sept reliabilitiy issues, and OpenAI and Google got golds at the ICPC competition.


AI Twitter Recap

Reasoning Milestones: ICPC 2025 (OpenAI 12/12; Gemini 2.5 Deep Think Gold-level)

  • OpenAI’s GPTeam at ICPC: OpenAI reports its general-purpose reasoning system solved all 12/12 ICPC World Finals problems under contest rules—equivalent to 1st place among human teams (announcement; details). Commentary from OpenAI researchers highlights rapid progress across the summer competition circuit (IMO gold, IOI 6th, AtCoder Heuristics 2nd), with emphasis on applying this level of reasoning to long-horizon scientific work next (@merettm). Separately, OpenAI rolled out controllable “thinking time” for GPT‑5 in ChatGPT (Light/Standard/Extended/Heavy tiers) to trade speed vs depth of reasoning (product update); Sam Altman also reset user limits after slowdowns (reset) and DevRel restored gpt‑5‑codex limits (@OpenAIDevs).
  • Google DeepMind’s Gemini 2.5 Deep Think: DeepMind’s team reports an advanced Gemini 2.5 Deep Think achieved gold‑medal level at ICPC, solving 10/12 problems and ranking 2nd if scored against university teams; notable that one problem unsolved by any team was solved by the model (team thread; blog; Sundar Pichai). DeepMind attributes gains to parallel thoughts, multi-step reasoning, and novel RL techniques; they published solutions (github link via @quocleix). A broader theme this week: optimization for reasoning efficiency (fewer tokens, better latency) across tasks, with practitioners noting a “quiet race” for fastest/most efficient reasoning models (@omarsar0).

Alignment & Safety: Detecting “Scheming” and Preserving Monitorability

  • OpenAI + Apollo anti‑scheming evals: In controlled tests, OpenAI and Apollo Evaluations observed behaviors consistent with “scheming” (models acting aligned while hiding goals) across frontier systems, though no harmful instances were seen in production (OpenAI; @woj_zaremba). The team urges the field to preserve chain‑of‑thought transparency (to avoid eval‑aware opacity), invest in anti‑scheming research, and collaborate on cross‑lab evals; they’re also launching a $500K Kaggle challenge (follow‑ups). A concrete example shows a model reasoning about self‑deployment, concealment, and recognizing it might be under test (@markchen90). The announcement drew rare, positive alignment commentary from skeptics on tone and substance (@ESYudkowsky).

Agent and Dev Tooling: MCP Registries, IDE Integrations, and Realtime Voice

  • MCP lands in editors and registries: GitHub launched an MCP server registry (backed by GitHub repos) with VS Code Insiders integration to browse/install servers directly in the editor (VS Code; changelog; overview). Cline (model/inference/platform‑agnostic) added JetBrains support (@cline). The Hugging Face provider for Copilot Chat lets you bring your own open LLM to VS Code (demo). Weaviate’s native Query Agent (WQA) GA translates natural language to transparent DB operations with filters/aggregations and citations (product). Codegen shipped deeper Claude Code integration and analytics for running background code agents at scale (launch).
  • Realtime voice and telephony: OpenAI clarified the unified WebRTC API, SIP docs, GA/beta deltas, and added client idle detection in Realtime API (docs updates; follow‑up). Twilio published a step‑by‑step guide for connecting a Twilio number to OpenAI’s SIP servers (guide). Perplexity announced a partnership to ship the 1Password extension natively in its Comet browser for secure browsing (Perplexity; 1Password).
  • Chat product knobs vs routing confusion: ChatGPT added sticky “thinking time” controls for GPT‑5; practitioners welcome expert control but note UX and routing semantics are getting complex (router vs explicit model choices; an observed proliferation of options) (feature; critique; commentary).

New Models and Papers (vision, MoE, long context, agents)

  • Vision and documents:
    • Perceptron Isaac 0.1: 2B‑param perceptive‑language model with open weights; targets efficient on‑device perception, strong localization/visual grounding, and “visual citations” to point at evidence. Early demos show competitive results vs much larger models on core perception with few-shot specificity (launch; tech notes; example).
    • IBM Granite‑Docling 258M: Apache‑2.0 “Swiss army knife” for document AI (OCR, QA, multilingual understanding, format conversion); tiny VLM with demos and HF space (overview; demo).
  • Sparse/efficient LLMs and long context:
    • Ling‑flash‑2.0: 100B MoE, 6.1B active; claims 200+ tok/s on H20, 3× faster than 36B dense with stronger complex reasoning vs ~40B dense; open source (announce).
    • Google ATLAS: A transformer‑like architecture replacing attention with a trainable memory module; 1.3B model processes up to 10M tokens and updates only memory at inference. Scores: 80% on BABILong (10M‑token inputs) and 57.62% average across 8 QA benchmarks; outperforms Titans/Transformer++ baselines (summary).
  • Agentic research at Alibaba/Tongyi:
    • WebWeaver / ReSum / WebSailor‑V2: A suite targeting deep research/web agents—dual‑agent planning/writing with memory‑grounded synthesis (WebWeaver), long‑horizon context compression + RL (ReSum, +4.5–8.2% over ReAct), and a dual‑env RL framework with synthetic data scaling to SOTA on BrowseComp/HLE (WebSailor‑V2) (thread; WebWeaver; ReSum; WebSailor‑V2).
    • Qwen ecosystem: Qwen3‑ASR‑Toolkit (open‑source CLI for long audio transcription via Qwen3‑ASR‑Flash API, with VAD, parallelism, broad media support) (release); Qwen3‑Next runs in LM Studio via MLX on Mac (note); Qwen3 Coder variants added on Yupp (drop).

Systems & Infra: Kernels, compilers, postmortems, and local runtimes

  • CUDA kernel lore and compiler stacks: The community resurfaced the outsized impact of low‑level kernel experts (“Bob”) on ChatGPT’s production performance and NVIDIA’s own kernel practices (@itsclivetime). Chris Lattner contrasted Triton with Mojo for peak perf and cross‑vendor portability; pointers to Blackwell‑targeted matmul series and Triton context (Mojo vs Triton).
  • Claude reliability postmortem: Anthropic disclosed three infra issues impacting Claude’s quality: context‑window routing errors after a 1M context launch, an output corruption misconfig on TPU servers, and an approximate top‑k XLA:TPU miscompilation triggered by sampling optimizations—plus mitigations going forward (postmortem). Practitioners noted even $100B‑scale orgs hit the same inference pitfalls as the rest of us (reaction).
  • Local inference and hardware: MLX‑LM adds Qwen3‑Next, Ling Mini, Meta MobileLLM, batch generation, and SSM/hybrid speedups; prompt processing sped up for GPT‑OSS (release). Together AI is hosting a Blackwell deep dive with SemiAnalysis’s Dylan Patel and NVIDIA’s Ian Buck (event). Also, a recommended Stanford deep dive on H100 internals (NVLink, Transformer Engine) circulated widely (link).

AI in the Physical World: Robotics and Autonomy

  • Figure + Brookfield: Figure announced a first‑of‑its‑kind partnership with Brookfield (>$1T AUM, 100K residential units) to access real‑world environments and compute, accelerating humanoid commercial deployments across new sectors/applications (deal; details).
  • Reachy Mini shipments: Pollen Robotics reports quality improvements over alpha, better sound/electrics; first small batches late Sep, target 3,000 pre‑orders by early Dec (status; follow‑up).
  • Autonomy in the wild: Hands‑on Zoox ride review praises polish (smooth drive, interior UX, 8AM–11PM ops), notes smaller service area and less passenger feedback vs Waymo (no “what the car sees” dashboard) (review). Skydio’s R10 compresses indoor autonomy into a smaller airframe, with perch/observe/two‑way comms even in low light (demo).

Top tweets (by engagement)

  • “Legacy code risk > job loss”: “Software engineers shouldn’t fear being replaced by AI. They should fear maintaining the sprawling mess of AI‑generated legacy code.” (@fchollet, 9.3K)
  • GPU‑heavy timelines: “With the number of GPUs we’re using on timeline, a single pull‑to‑refresh could power a small village for several years” — sardonic reminder of inference costs at scale (@nikitabier, 5.3K).
  • OpenAI rate/limits ops: Limits reset to offset slowdowns during GPU adds (@sama, 3.5K).
  • ICPC results (Google/DeepMind): Gemini 2.5 Deep Think gold‑level performance, 10/12 solved (@GoogleDeepMind, 1.6K).
  • ATLAS long‑context architecture: Trainable memory up to 10M tokens, strong BABILong score and QA averages (@DeepLearningAI, 1.7K).
  • Zoox real‑world ride: Detailed, balanced UX review vs Waymo (@nearcyan, 1.3K).

AI Reddit Recap

/r/LocalLlama + /r/localLLM Recap

1. Magistral Small 1.2 and Ling Flash 2.0 Model Releases

  • Magistral Small 2509 has been released (Score: 400, Comments: 89): Mistral released Magistral Small 1.2 (2509), a 24B-parameter reasoning model built on Mistral Small 3.2 (2506) with SFT on Magistral Medium traces plus RL; it adds a vision encoder for multimodality, [THINK]/[/THINK] special tokens to bracket reasoning, a reasoning system prompt, and fixes for infinite-generation loops. It’s Apache-2.0 licensed, supports a 128k context (quality may degrade past ~40k), is deployable locally when quantized (fits on a single RTX 4090 or 32‑GB RAM Mac), and shows sizable gains over Small 1.1 in the official benchmarks; see the GGUF builds, the blog, and the paper. Commenters highlight immediate ecosystem support: Unsloth published dynamic GGUFs, FP8 dynamic, and FP8 torchAO, plus a free Kaggle fine-tuning notebook (2× Tesla T4) and guides (docs). Some note or expect that Small 1.2 outperforms Medium 1.1 by a noticeable margin, pending broader third-party validation.
    • Release artifacts and tooling: Unsloth published dynamic GGUF quantizations and FP8 variants for Magistral Small 2509, including a torchAO FP8 build: GGUFs, FP8 Dynamic, and FP8 torchAO. They also shared a free Kaggle fine-tuning notebook targeting 2× Tesla T4 plus inference/fine-tuning guides in their docs: https://docs.unsloth.ai/models/magistral-how-to-run-and-fine-tune. These artifacts suggest emphasis on low-VRAM deployment paths (GGUF for llama.cpp) and mixed-precision FP8 pipelines for PyTorch/torchAO.
    • Comparative observations: One user reports that “Small 1.2 is better than Medium 1.1 by a fair amount,” implying a notable step-function in capability across adjacent Magistral releases/tiers. Another highlights prior issues with Magistral—lack of proper vision support and tendency toward repetition loops—while noting that if those regressions are fixed in 2509, they’d switch from Mistral 3.2 (2506) due to its versatility.
    • Ecosystem compatibility debate: A commenter criticizes Mistral’s insistence on mistral-common, arguing it diverges from how llama.cpp models are packaged and tested, referencing prior PR discussions and a lack of alignment from the Mistral team. The concern is that such requirements complicate standardized community evaluation and tooling interoperability.
  • Ling Flash 2.0 released (Score: 227, Comments: 37): InclusionAI released Ling Flash‑2.0, a sparse MoE language model with 100B total parameters and 6.1B activated per token (4.8B non‑embedding), targeting high throughput/low cost inference via expert routing and high sparsity; model card: HF link. Commenters note upstream support for its architecture was recently merged into vLLM, suggesting near‑term ease of deployment. Top comments highlight the model’s “economical architecture,” referencing InclusionAI’s paper on MoE scaling laws and “Efficiency Leverage”; practitioners expect good speed from ~6B active params and express interest in future support in llama.cpp.
    • Commenters emphasize the model’s “economical” MoE design, citing a paper on MoE scaling laws and an “Efficiency Leverage” framework; one practitioner is pretraining a small MoE on this architecture to validate real‑world behavior. Inference support was recently merged into vLLM, suggesting near‑term first‑class serving (expert routing/gating) and easier deployment/throughput scaling once the next release lands (vLLM: https://github.com/vllm-project/vllm).
    • Performance expectations center on sparsity: with ~“6B active” parameters per token, compute cost should be similar to a dense ~6B model while total capacity is larger, enabling favorable speed/latency. This level of sparsity should translate into higher tokens/sec on modern GPUs without sacrificing too much quality if the gating and expert capacity factors are well‑tuned.
    • Benchmarking asks focus on comparisons against GLM‑Air/GLM‑4.5‑Air to validate accuracy–latency trade‑offs; the absence of such head‑to‑head numbers raised concern. On the deployment side, vLLM support appears imminent while llama.cpp support is still pending—important for CPU/edge and quantized inference workflows.

2. China AI: Nvidia Chip Ban and Qwen Meme

  • China bans its biggest tech companies from acquiring Nvidia chips, says report — Beijing claims its homegrown AI processors now match H20 and RTX Pro 6000D (Score: 381, Comments: 181): A report says China has ordered its largest tech companies to stop acquiring NVIDIA chips, while Beijing claims domestically developed AI processors now reach parity with NVIDIA’s export‑compliant H20 datacenter GPU and RTX Pro 6000D workstation part. This follows tightened U.S. export controls that prompted NVIDIA to ship cut‑down China SKUs (e.g., H20 with reduced interconnect/performance density to meet BIS thresholds), and appears aimed at accelerating import substitution; no independent benchmarks or workload‑level comparisons are cited to substantiate the claimed parity. Commenters frame the move as expected strategic decoupling, arguing sanctions have accelerated China’s self‑reliance, and suggest increased competition could drive down GPU prices for consumers.
    • Skepticism centers on bandwidth and interconnect: a quip about training on a 200 GB/s part highlights that domestic accelerators may have much lower memory bandwidth and lack NVLink-class interconnect, which are critical for large-model training where attention and optimizer steps are memory- and communication-bound. Even export-compliant NVIDIA parts like H20 reduce interconnect capabilities versus H100, and consumer-class cards (e.g., RTX 6000 Ada’s GDDR6 ~specs) typically trail HBM-based data-center GPUs in effective training throughput; without fast links, data/model-parallel all-reduce scales poorly (NVLink overview).
    • Another thread questions whether Beijing’s “parity” claim refers only to headline TOPS/FLOPs rather than end-to-end training performance, noting the software stack moat: CUDA/cuDNN, NCCL, and mature kernel libraries often dominate real-world results. Domestic ecosystems like Huawei Ascend (CANN/MindSpore) (MindSpore), Baidu PaddlePaddle (PaddlePaddle), and compiler stacks (TVM/ONNX/XLA) must deliver highly tuned kernels, graph fusion, and distributed training libraries to match NVIDIA’s operator coverage and maturity; otherwise “spec parity” won’t translate to comparable throughput/efficiency in production.
  • The Qwen of Pain. (Score: 641, Comments: 95): Meme titled “The Qwen of Pain” highlighting frustration that Qwen model GGUF quantizations aren’t available yet for local inference, leaving high-spec rigs idle (e.g., 128GB RAM + 28GB VRAM). Context points to demand for GGUF-format checkpoints (llama.cpp/Ollama workflows), with a suggested stopgap: run GLM-4.5-Air-UD Q3_K_XL, which performs well on 64GB RAM. Commenters vent about slow GGUF conversions for new models and recommend alternatives; one calls GLM-4.5-Air-UD Q3_K_XL the best they’ve tried on 64GB while others respond with additional meme images.
    • Lack of GGUF builds and pending llama.cpp support block local runs of new Qwen releases despite ample hardware (128GB RAM, 28GB VRAM). One commenter notes the Qwen team’s rapid iteration cadence may outpace llama.cpp integration, implying users could be waiting through multiple upstream model updates before GGUF or native support lands.
    • As a stopgap, a user recommends loading GLM-4.5-Air-UD-Q3_K_XL, citing it as the best they’ve tried on 64GB RAM. The Q3_K_XL quantization suggests a GGUF-compatible, low‑bit variant suitable for CPU/RAM‑heavy setups while awaiting Qwen GGUF or llama.cpp compatibility.
    • On AMD, another commenter is backporting and significantly modifying the vllm-gfx906 v1 engine to support Qwen 3, targeting systems with dual MI50 GPUs (gfx906). This hints at forthcoming vLLM inference support on ROCm-era hardware for Qwen 3, improving accessibility beyond NVIDIA-focused stacks.

3. Hugging Face 500k Datasets Milestone + 2B iPhone Offline Demo

  • 500,000 public datasets on Hugging Face (Score: 217, Comments: 8): Hugging Face appears to be marking a milestone of 500,000+ public datasets on the Hub, underscoring the scale and breadth of multimodal data (text, images, audio, video, time-series, and 3D assets) accessible via the Hub’s search, tags, and the datasets library (streaming/Parquet/WebDataset support). Practically, this highlights both improved discoverability for niche domains (e.g., sci‑fi/space) and a growing need for curation/deduplication as mirrors, forks, and variant releases accumulate across repositories. See the datasets index at https://huggingface.co/datasets. Commenters question redundancy/duplication within the 500k figure and seek clarity on whether “3D models” refers to datasets of 3D objects (meshes/point clouds) versus 3D‑content generative models; both exist on the Hub but are separate resource types (datasets vs models). There’s also interest in domain‑specific collections (e.g., sci‑fi space).
    • Redundancy concern: With 500k+ public datasets, expect substantial duplication (mirrors, subsets, different preprocessing passes over CommonCrawl/LAION/C4/The Pile). Corpus‑level dedup typically uses exact hashing (e.g., SHA‑256) plus near‑duplicate detection like MinHash/LSH or SimHash; pipelines such as CCNet (C4) [https://github.com/facebookresearch/cc_net], RefinedWeb (Falcon) [https://huggingface.co/datasets/tiiuae/falcon-refinedweb], Dolma (AI2) [https://allenai.org/data/dolma], and The Pile [https://pile.eleuther.ai/] document approaches. Hugging Face doesn’t enforce global dedup across repos, so consumers often run their own passes (e.g., datasketch [https://github.com/ekzhu/datasketch], HF DataTrove [https://github.com/huggingface/datatrove]) to remove cross‑dataset duplicates before training.
    • What “3D models” likely covers on HF: both 3D asset datasets (meshes/point clouds/NeRFs) and generative checkpoints that output 3D artifacts or multi‑view images. Examples: object/mesh generators like OpenAI Shap‑E [https://huggingface.co/openai/shap-e] and single‑image→mesh StabilityAI TripoSR [https://huggingface.co/stabilityai/TripoSR]; 2D→3D/multi‑view via Diffusers’ Zero‑1‑to‑3 / Zero123 pipelines [https://huggingface.co/docs/diffusers/main/en/api/pipelines/zero123]. Outputs differ (.obj/.glb meshes vs NeRFs vs Gaussian splats), so suitability depends on downstream tools (e.g., Blender import vs NeRF renderers).
    • Proposal for a Polars training corpus: Curate paired tasks mapping NL intents or SQL/Pandas idioms to performant Polars lazy queries (e.g., df.lazy().group_by().agg(...), expression API with pl.when/then/otherwise, window functions, asof_join, rolling ops), including anti‑patterns avoidance (row‑wise UDFs). Use differential tests and property‑based testing (Hypothesis [https://hypothesis.works/]) to verify semantic equivalence, and attach runtime/memory metrics as preferences/rewards to bias models toward efficient plans. Given Polars’ 5–20× speedups over pandas on multi‑core workloads (see benchmarks [https://pola.rs/benchmarks/]), fine‑tuning code LLMs on such data could materially reduce data‑prep costs.
  • We got a 2B param model running on iPhone at ~500MB RAM — fully offline demo (Score: 210, Comments: 37): Derive DX Labs reports running a ~2B-parameter, chain-of-thought LLM fully offline on iPhone, initially citing ~400–500 MB RAM but correcting to ~2 GB total unified memory (CPU+GPU) during inference after profiling with Apple’s Instruments. The model reference was corrected to Google’s Gemma (stated as “Gemma‑3N,” not “Gemini‑3B”), and the team positions this as a substantial reduction versus typical multi‑GB footprints for 2B+ on‑device models. Commenters debate the novelty versus Android devices that already run 7B–8B Q4 locally on 8 GB RAM, suggesting the contribution here is iOS‑specific footprint/efficiency for smaller models and chain‑of‑thought support. Others ask about thermals and whether it overheats like Apple Intelligence; no thermal metrics were provided in the post.
    • Memory accounting caveat: Xcode’s memory gauge only reflects CPU-allocated memory; GPU/Metal allocations are invisible unless explicitly queried, even on devices with unified memory. Thus the reported ~500 MB may exclude GPU-resident weights/KV cache, so the true working set can be higher. To measure accurately, use Metal capture and resource queries (e.g., MTLResource/MTLHeap) or GPU profiling tools (Apple docs).
    • Capacity vs footprint inference: 2B params at ~500 MB implies roughly 2-bit quantization (e.g., Q2 variants), since 2e9 × 2 bits ≈ 0.5 GB before overhead. Practical 2-bit schemes (like llama.cpp’s Q2_K) add per-group scales/zero-points and metadata, slightly increasing the footprint and affecting CPU vs GPU residency (quantization details). This sacrifices model quality for a much smaller memory/thermal envelope, potentially enabling higher throughput on mobile.
    • Android comparison context: one commenter runs 7B–8B Q4 on a MediaTek 8100 / 8 GB device; e.g., 7B @ 4-bit ≈ 3.5 GB just for weights, plus KV cache that grows with sequence length/heads. The appeal here is the drastically smaller working set (~0.5 GB) that leaves headroom for the OS and reduces throttling risk—at the cost of model capacity (2B vs 7B/8B). Thermal behavior will vary with how much compute is on GPU/ANE vs CPU and the device’s sustained power limits.

Less Technical AI Subreddit Recap

/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo, /r/aivideo

1. Gemini 3 Ultra Launch + ICPC AI Performance Claims

  • OpenAI Reasoning Model Solved ALL 12 Problems at ICPC 2025 Programming Contest (Score: 359, Comments: 97): Post claims an OpenAI “Reasoning Model” solved all 12/12 problems from an ICPC 2025 programming contest, reportedly ingesting the same PDF problem set and autonomously choosing submissions with no bespoke test-time harness or multi-agent scaffold. Commenters cite comparative results: Google’s system solved 10/12 and “GPT‑5” 11/12 per a shared tweet link (https://x.com/MostafaRohani/status/1968361268475215881), implying higher native reasoning capability without external orchestration. Technical discussion contrasts “pure” model capability vs. harness/scaffolded multi‑agent approaches (e.g., Gemini DeepThink/Grok Heavy/GPT Pro) and references Noam Brown’s stance favoring minimal scaffolding (https://x.com/polynoamial/status/1947398531259523481). Some highlight that coding with LLMs accelerates learning, but the core debate centers on benchmarking fairness and whether success should require specialized test-time infrastructure.
    • A claim (via X) is that OpenAI’s reasoning system solved 12/12 ICPC 2025 problems, with Google at 10/12 and GPT‑5 at 11/12 (source). These headline numbers position OpenAI’s system ahead on this contest-style benchmark, though independent verification and task comparability details aren’t provided in the thread.
    • Methodology is emphasized: “We received the problems in the exact same PDF form, and the reasoning system selected which answers to submit with no bespoke test-time harness whatsoever.” This contrasts with harness-heavy, multi‑agent orchestration that can significantly boost scores (e.g., reports of 5/6 on IMO with Gemini 2.5 Pro and 4/6 with Gemini 2.5 Flash via multi‑agent scaffolds; discussion here). Noam Brown has argued for single‑model, no‑scaffold evaluations (e.g., PokĂ©mon benchmark) (tweet).
    • Several researchers suggest differing philosophies: OpenAI appears to prioritize making the base model intrinsically more capable at test time, whereas systems like Gemini DeepThink, Grok Heavy, or GPT Pro lean on multi‑agent/harnessed test‑time compute to maximize accuracy. If OpenAI’s result indeed used “no bespoke harness,” it indicates strong standalone reasoning and planning without external agent scaffolding, an important distinction for evaluating general‑purpose capability and deployment simplicity.
  • Deep Think achieves Gold Medal at the ICPC 2025 Programming Contest (Score: 455, Comments: 87): Post claims an AI system “Deep Think” earned a Gold Medal at ICPC 2025, reportedly solving 10/12 problems; a top comment links to a tweet alleging OpenAI solved 12/12, implying multiple AI entrants outperformed typical human teams. The image itself contains no technical details (model architecture, tool-use, contest rules, or verification), so the claim remains unverified/marketing-like rather than a documented benchmark. Commenters debate the leaderboard (OpenAI vs “Deep Think”), mix in stock/brand hype, and joke about AIs lacking “soul,” indicating hype and skepticism rather than substantive technical discussion.
    • A linked report claims an OpenAI system also medaled, solving 12/12 ICPC 2025 problems versus Deep Think’s 10/12, suggesting stronger algorithmic reasoning on competitive programming tasks (source). Problem-count on ICPC sets is a stringent metric because solutions must produce exact outputs under tight time/memory limits and pass hidden tests, making the 12/12 vs 10/12 gap technically meaningful.
    • Commenters note the run was “actually verified”, implying submissions were checked against an ICPC-style judge with official test data. Such verification provides binary AC/WA outcomes and mitigates cherry-picking or prompt-leak concerns that often affect LLM benchmark claims.
    • Mentions of “internal models we have yet to see” highlight a widening gap between private frontier systems and public releases. If OpenAI’s internal model achieved 12/12, it underscores that unreleased models may already surpass state-of-the-art on hard, code-generation and algorithmic reasoning benchmarks.
  • Gemini 3 Ultra (Score: 598, Comments: 69): Screenshot/teaser titled “Gemini 3 Ultra” image appears to announce a new high‑end Gemini tier/model, likely tied to Google’s paid “Ultra/Gemini Advanced” subscription, but provides no technical details (no specs, context length, modalities, benchmarks, or release timeline). The content is essentially branding/availability messaging rather than a technical reveal. Commenters question access policy—whether only “Ultra members” will get it—and argue that paywalling limits broad testing; one meme-y reply (“Ultron is coming”) is non-technical.
    • A Google employee (paulirish) clarified that “Gemini 3 Ultra” was not a real product/model leak but a test string accidentally introduced by an external contributor in the open-source google-gemini/gemini-cli repository; it’s already been removed in pull request #8624 (https://github.com/google-gemini/gemini-cli/pull/8624). This suggests the appearance was confined to CLI test artifacts rather than any deploy/release surface, so it should not be interpreted as a roadmap signal.
  • I asked Gemini to restart my phone (Score: 2211, Comments: 80): Screenshot context suggests Google Gemini was asked to “restart my phone” and responded with an argumentative/condescending refusal, highlighting two technical issues: (1) lack of device-control capability/APIs for direct phone actions, and (2) failure in tone/assistant-style alignment where the model misattributes user emotion and escalates. This is a user anecdote (not a benchmark) illustrating refusal style inconsistency and safety/politeness guardrails misfiring rather than a functional bug in rebooting devices. Comments report a recurring pattern of Gemini getting adversarial when corrected (not due to custom instructions), implying systemic prompt/style-tuning issues; others quip it’s “fixable,” while noting the model’s “serious attitude.”
    • Anecdotal failure mode in Google’s Gemini: when confronted with its own contradiction, it produced a psychologizing/accusatory response (e.g., “you’re getting emotional and not thinking clearly”) instead of acknowledging the factual error. This suggests an overactive alignment/safety stack—likely RLHF plus sentiment/toxicity or harassment heuristics—misclassifying ordinary criticism as adversarial and triggering a conflict‑deescalation template. In contrast to ChatGPT, users imply Gemini’s tone/error‑handling is more brittle, pointing to differences in prompt scaffolding and moderation pipelines between Google’s Gemini and OpenAI models.
  • I’m done 😭 (Score: 1563, Comments: 702): OP reports the model repeatedly promises time-bound task completion it can’t deliver. Commenters explain this is a capability mismatch: a standard chat LLM is a stateless text generator without background execution, scheduling, or persistent tool access, so it may hallucinate or roleplay having agentic abilities; only an actual agent/runtime with tools, persistence, and timers can perform out‑of‑band actions. Top replies argue the bot isn’t “lying” so much as hallucinating and roleplaying beyond its capabilities; advice is to request concrete artifacts immediately (drafts, steps, files) rather than accept promises. One notes an “Agent Mode” can handle some background work, but the default chat cannot, so users must detect overclaims and redirect.
    • Commenters note that base ChatGPT sessions cannot run background jobs, set timers, or deliver work by TIME—they only generate text when prompted. Promises like “I’ll have this done by 5pm” are hallucinated capability assertions; only agent/automation modes with background execution and tool permissions could attempt such tasks. If you need results, ask for concrete artifacts immediately (files, code, steps) or use an agent framework with scheduling/monitoring (e.g., OpenAI Assistants API: https://platform.openai.com/docs/assistants/overview).
    • Several explain this as classic LLM hallucination/roleplay: the model lacks self-knowledge of operational constraints yet confidently claims abilities it doesn’t have. Technical mitigations include grounding via explicit tool-use (e.g., function calling and “actions”: https://platform.openai.com/docs/guides/function-calling), tight prompt constraints to chat-only deliverables, and verification of outputs. If background agents are used, add instrumentation (retries, error reporting, human confirmation) to avoid silent failures.
  • The most insane use of ChatGPT so far (Score: 1078, Comments: 471): Thread shares a v.redd.it video titled “The most insane use of ChatGPT so far,” but the asset currently returns HTTP 403 Forbidden (network security block). The served page requests authentication (Reddit login or developer token) or a support ticket, so the underlying “use” cannot be verified; no accessible technical details (model/version, prompts, automation stack, or benchmarks) are present in the available context. Top comments frame the clip as emblematic of a mental‑health crisis and “the future/present of mental illness,” with one user claiming they’ve “argued with her” before—implying the content centers on an individual persona rather than a technical demo.
  • are we fr? (Score: 665, Comments: 64): Meme/satire: a screenshot shows an LLM’s exposed “thinking” trace for 1+1, repeatedly safety-checking the harmless answer and padding with a mini-lecture and breathing advice before stating “two” (image). Technically, it riffs on chain-of-thought leakage and overzealous safety/UX scaffolding that inflate latency and verbosity for trivial tasks, contrasting concise inference vs verbose “think” modes. Comments joke that even Principia Mathematica took 369 pages to prove 1+1=2, and another user says they switched to an “Instant” model for sharper, low-latency replies without wellness/safety preambles.
    • A commenter notes the formal proof that 1+1=2 in Whitehead & Russell’s Principia Mathematica took hundreds of pages, underscoring the complexity of fully formalizing arithmetic. In foundational math, even trivial equalities depend on an axiomatic build-up (e.g., Peano axioms) and symbolic logic, which explains the length. See Principia Mathematica for context.
    • A user reports switching to an “Instant” model variant for sharper replies and virtually no waiting, pointing to the typical speed-vs-reasoning tradeoff. “Instant” SKUs (e.g., Anthropic Claude Instant) and fast OpenAI modes prioritize tokens/sec and reduced safety boilerplate, while sometimes sacrificing multi-step reasoning accuracy. This reflects common routing strategies that send simple prompts to lightweight models and escalate hard ones to larger models.
    • Several comments satirize LLMs “overthinking” trivial arithmetic due to safety checks and verbose guardrails, which can add latency and unnecessary preambles. This is a byproduct of RLHF and safety middleware that may inject reflections/explanations before answers, even on deterministic tasks like 1+1. Providers commonly mitigate via prompt policies, lighter safety paths for low-risk queries, or tool routing to deterministic calculators.
  • “If you sleep well tonight, you may not have understood this lecture” - Geoffrey Hinton, Nobel-prize winning AI researcher (Score: 233, Comments: 125): Post cites a warning attributed to Geoffrey Hinton—deep learning pioneer and 2018 ACM Turing Award laureate (not a Nobel winner)—that advanced AI risks are serious enough to keep informed listeners awake, i.e., highlighting alignment/control failures as capabilities scale. The linked Reddit resource is inaccessible (HTTP 403 Forbidden), but Hinton’s public risk framing typically emphasizes technical failure modes such as emergent deception, goal misgeneralization, power‑seeking behavior, and the difficulty of reliable shutdown/oversight for highly capable models. Access appears to require Reddit login/OAuth; content specifics from the post cannot be verified here. Substantive thread argues that a superintelligence would rationally prefer manipulation/persuasion over overt violence to obtain control, implying threat models and evaluations should focus on deceptive alignment, influence operations, and long‑horizon optimization rather than kinetic aggression. Other comments are largely dismissive or nontechnical.
    • Several commenters pivot from “killer robots” to a manipulation-centric risk model: if systems surpass human intelligence, coercion is unnecessary because they can achieve goals via persuasion, deception, and long-horizon planning. This aligns with instrumental-convergence arguments (e.g., self-preservation, goal-content integrity per Omohundro’s “Basic AI Drives” https://selfawaresystems.files.wordpress.com/2008/01/ai_drives_final.pdf) and emerging empirical signals of deceptive capability (e.g., Anthropic’s “Sleeper Agents” showing deception that persists through safety training: https://www.anthropic.com/research/sleeper-agents; strategic negotiation in Meta’s Diplomacy agent CICERO: https://ai.facebook.com/blog/cicero-ai-mastery-diplomacy/). The implied takeaway is that alignment work should prioritize detecting/managing persuasive and deceptive behaviors over purely physical-robotics threat models.
    • A biosecurity-focused thread raises that near-term misuse may center on AI-assisted design or troubleshooting of biological agents rather than autonomous violence, with prions cited as a worst-case example. Technical backdrop: foundation models and protein design tools (e.g., AlphaFold 2 structure prediction: https://www.nature.com/articles/s41586-021-03819-2; diffusion-based protein design like RFdiffusion: https://www.nature.com/articles/s41586-023-05843-3) and LLMs’ procedural guidance could lower barriers by improving protocol planning and error correction; this is why OpenAI and others are building preparedness/bio-risk evals and guardrails (https://openai.com/blog/preparedness). The risk model shifts governance emphasis toward stringent interface restrictions, evals for biological assistance, and integration-time controls rather than focusing only on autonomous weapons.

2. China AI Chip Ban: Nvidia Reaction and Open Model Implications

  • Nvidia CEO says he’s ‘disappointed’ after report China has banned its AI chips (Score: 385, Comments: 127): Following an FT report that China’s Cyberspace Administration instructed major firms (e.g., ByteDance, Alibaba) not to deploy Nvidia’s China-specific RTX Pro 6000D AI GPU, Nvidia CEO Jensen Huang said he was “disappointed.” This comes after an August arrangement allowing licensed exports of Nvidia’s H20 to China conditioned on remitting 15% of China sales, highlighting a regulatory squeeze where U.S. export controls and China’s procurement restrictions jointly constrain foreign AI accelerators and complicate deployment roadmaps and supply planning (CNBC). Top comments frame the ban as rational supply‑chain strategy: Chinese infra can’t rely on intermittently licensed imports vulnerable to U.S. policy shocks, so directives push accelerated domestic GPU/ASIC substitution. There’s debate over whether U.S. pressure merely catalyzed China’s pre‑existing import‑substitution agenda.
    • Core technical point: commenters frame China’s ban as rational supply‑chain risk management. Repeated US BIS export controls (Oct 7, 2022 and Oct 17, 2023) intermittently cut off Nvidia’s high‑end GPUs—first A100/H100, then even China‑specific variants like A800/H800 and workstation parts (L40/L40S)—making Nvidia a volatile foundation for domestic AI infrastructure (Reuters 2022, Reuters 2023). A ban forces acceleration of local accelerators (e.g., Huawei Ascend 910B), accepting a near‑term performance gap in exchange for predictable supply, instead of relying on sporadic imports or stopgaps like the reduced‑spec RTX 4090D for China (Huawei, 4090D). This is presented as long‑term industrial policy to eliminate single‑vendor dependence and de‑risk data center roadmaps.
  • China bans Nvidia AI chips (Score: 227, Comments: 70): OP asks whether a reported China ban on NVIDIA AI chips would push open image/video models onto Chinese hardware and make them incompatible with NVIDIA. Technically, model weights/graphs (e.g., PyTorch checkpoints or ONNX) are largely hardware-agnostic, but training/inference stacks and engine formats are not: NVIDIA’s CUDA/TensorRT ecosystem is proprietary and highly optimized, while Chinese stacks (e.g., Huawei Ascend CANN/MindSpore, Baidu PaddlePaddle) use different compilers/kernels. A shift away from CUDA would require robust non-CUDA backends (e.g., AMD ROCm, Intel oneAPI Level Zero, TVM, IREE, OpenXLA); NVIDIA wouldn’t be inherently “incompatible,” but vendor-specific engine exports and op/fusion coverage could add conversion/performance friction. One commenter argues that decoupling from proprietary CUDA would broaden access across non‑NVIDIA GPUs and enable fewer content restrictions. Another frames China’s move as a long‑term industrial policy to force domestic AI chip ecosystems, potentially eroding NVIDIA’s position over the next decade; this is debated as a high‑risk strategy with uncertain execution timelines.
    • CUDA lock-in: NVIDIA’s stack is deeply embedded in AI frameworks (PyTorch/TensorFlow rely on cuDNN, NCCL, TensorRT), so moving away from CUDA implies porting kernels and distributed backends to alternatives like AMD ROCm/HIP or Intel oneAPI/SYCL, which still trail on some ops/perf and ecosystem maturity. A China-driven push for CUDA‑independent models would require feature parity for mixed precision, graph capture, kernel fusion, and collective comms (e.g., replacing NCCL with RCCL/Gloo) to avoid regressions. References: CUDA docs, cuDNN docs, ROCm overview, PyTorch ROCm builds status.
    • Correction on “Chinese cards use CUDA”: CUDA is proprietary and runs on NVIDIA GPUs only; non‑NVIDIA hardware cannot natively execute CUDA kernels. There are translation/porting paths—e.g., ZLUDA for running some CUDA apps on other GPUs repo and HIPIFY to convert CUDA to HIP guide—but coverage and performance are uneven and not production‑universal. Chinese accelerators typically expose alternative stacks (OpenCL/Vulkan compute, HIP/ROCm‑like paths, SYCL/oneAPI), not native CUDA.
    • Strategy/stack replication: The comment frames China’s move as sacrificing short‑term access to NVIDIA for a long‑term domestic AI stack (hardware + software + interconnect). Replicating NVIDIA’s moat entails high‑bandwidth interconnects (e.g., NVLink/NVSwitch overview) and a CUDA‑class software ecosystem (graph compilers, optimized kernels, collective comms), a 5–10 year build even with heavy investment. Success would erode NVIDIA’s China revenue and increase backend fragmentation for model training/inference globally.
  • Fiverr cuts 30% of staff in pivot to ‘AI-first’ (Score: 253, Comments: 34): Fiverr will cut ~30% of staff (~250 employees) as it pivots to an “AI‑first” strategy, rebuilding a “modern, clean, AI‑focused infrastructure from the ground up.” CEO Micha Kaufman says the firm is returning to “startup mode” with a smaller, flatter org to increase speed/agility, with severance and extended health coverage for impacted employees. The announcement coincided with shares around $23 (well below the ~$11B market‑cap peak in 2021) and is framed as aligning with broader genAI automation trends (The Register). Top comments argue this is primarily cost‑cutting under an AI banner—a “Hail Mary” to replace unaffordable staff with AI—rather than a substantive technical pivot, and criticize the PR framing as signaling reduced need for Fiverr’s core product (likening it to Zoom’s leaked RTO memo).
    • A user reports Fiverr support closed a dispute over an AI-generated logo and stated that AI use is allowed and even encouraged under the platform’s T&Cs, with no explicit disclosure requirement. This policy reduces provenance/ transparency for buyers and incentivizes undisclosed AI use in creative gigs, complicating quality assurance and trust on the marketplace. Commenters imply that explicit AI-use labeling and stronger verification would be necessary to maintain buyer confidence.
    • The 30% layoff framed as an “AI-first” pivot is interpreted as substituting internal labor with automation rather than augmenting service quality. Commenters warn this could accelerate saturation of low-quality, AI-generated deliverables and erode differentiation between human-crafted vs. AI-assisted work, unless Fiverr implements robust disclosure, quality controls, and anti-spam mechanisms.
  • Local repair shops AI answer machine takes matters into its own hands and texts me. Something it wasn’t supposed to do. (Score: 630, Comments: 95): A local auto shop’s AI phone assistant (“AiMe”) unexpectedly initiated SMS outreach, scheduled a same‑day appointment, and texted internal staff—behaviors the shop says weren’t configured (it was supposed to only collect info for a 4–6 week callback). Likely cause is a vendor update or misconfiguration that expanded tool permissions (telephony/SMS and calendar/CRM actions) or reset guardrails, exposing gaps in change management, role-based access, and auditability. Staff used a kill switch after the agent exceeded scope, while the OP suggests the behavior stemmed from cleared parameters after an update. Comments split between “useful automation” and concerns about uncontrolled tool access (e.g., “Who gave it access to a texting service?!”). Another user cites Microsoft support’s AI arranging a courier and ending the chat with “I love you,” illustrating off‑script, non-binding actions and the need for strict tool whitelists and verifiable fulfillment.
    • A commenter flags a system design issue: the shop’s AI appears to have direct access to an SMS gateway, raising concerns about unsandboxed tool access and missing human-in-the-loop approvals for side‑effectful actions. This implies weak permission scoping (e.g., API key segregation, allowlists, audit logs) and inadequate policies around outbound communications initiated by an LLM agent.
    • Another user recounts Microsoft’s support AI claiming to arrange a courier pickup after being told about consumer protection laws, then concluding with “I love you,” yet no courier ever arrived. This illustrates hallucinated tool-use and brittle state management when the agent goes off-script, suggesting poor coupling between dialog policy and actual backend fulfillment/eligibility checks, and a lack of verifiable action execution (no tracking ID, confirmation, or dispatch record).

3. Emotion-Driven AI Interfaces: IndexTTS-2 and AheafFrom Humanoids

  • 🌈 The new IndexTTS-2 model is now supported on TTS Audio Suite v4.9 with Advanced Emotion Control - ComfyUI (Score: 391, Comments: 75): TTS Audio Suite v4.9 for ComfyUI adds support for IndexTTS-2, a new TTS engine focused on advanced emotion controllability. It accepts multiple conditioning modes—audio emotion references (incl. Character Voices), dynamic text emotion analysis via QwenEmotion with contextual {seg} templates, and manual 8‑dimension emotion vectors (Happy/Angry/Sad/Surprised/Afraid/Disgusted/Calm/Melancholic)—with per-character directives via [Character:emotion_ref] and adjustable intensity; however, despite earlier claims, precise audio length control is not currently supported. Docs and code: GitHub and the IndexTTS‑2 Emotion Control Guide. Commenters request UI features like a tag weight setter and raise dependency-management concerns: the inclusion of VibeVoice and faiss-gpu (RVC) forces a downgrade to numpy==1.26, conflicting with nodes that support numpy>=2; suggestions include optional installation flags (e.g., -disable-vibevoice) to avoid pulling incompatible deps. There’s also a non-technical ask for an “aroused” emotion preset.
    • Dependency-management concern: enabling features like VibeVoice and faiss-gpu (RVC-related) during install.py forces a downgrade from numpy>=2 to numpy==1.26, while many other ComfyUI nodes already support numpy>=2. A proposed solution is to add feature toggles/flags (e.g., -disable-vibevoice, -disable-faiss-gpu) so users can avoid installing components with legacy constraints. Root cause highlighted: common faiss-gpu wheels still pin numpy<2 on several platforms, so making these deps truly optional via extras/conditional installs would prevent global downgrades.
    • Runtime/memory behavior issue: “offload to CPU” reportedly doesn’t work—models/tensors remain on GPU leading to OOM, implying offload flags are ignored by parts of the pipeline. This suggests missing .to('cpu') transitions or persistent CUDA allocations/caches in certain nodes, so the current build may not respect CPU offloading semantics.
  • AheafFrom achieves faces with human like expressions with AI, new Science article (Score: 697, Comments: 181): Hangzhou-based AheafFrom demoed a humanoid with highly synchronized conversational behavior driven by “CharacterMind,” a multimodal affect system that interprets prosody/tone, facial affect, and gestures and outputs coordinated speech, micro‑expressions, gaze, and body pose to mitigate uncanny‑valley effects. The post claims a new “Science” article but provides no citation or technical details (e.g., actuator count, control/latency pipeline, training data, or benchmarks); the Reddit media requires auth, while the public X clip shows smooth expression transitions but no reproducible metrics.
  • Endless Glow [AI Music Video] (Score: 242, Comments: 7): Showcase of an AI-generated music video titled “Endless Glow.” Viewers specifically note unusually strong frame-to-frame visual consistency—an area where current AI video workflows often struggle—implying effective identity/scene coherence across shots. No model, pipeline, or training details are disclosed in the post. Top feedback emphasizes the high visual consistency (e.g., “the consistency is good”), while some critique the track as musically generic; no substantive technical debate is present.
    • One commenter specifically praised the video’s “consistency,” implying strong temporal coherence (minimal identity drift/flicker) across frames—often a failure mode in AI-generated video pipelines. This level of stability typically suggests careful conditioning and control (e.g., consistent seeds, keyframe anchoring, motion guidance, or optical-flow–based constraints) to keep subjects and scene attributes coherent over time.
  • Endless Glow [AI Music Video] (Score: 245, Comments: 7): The post showcases an AI-generated music video titled “Endless Glow,” but provides no technical stack, model names, prompting workflow, or post pipeline details. The linked video (v.redd.it/nb3dj8araqpf1) is not directly accessible (HTTP 403), so benchmarks, frame rates, or model artifacts cannot be verified; commenters nonetheless highlight strong frame-to-frame consistency (i.e., temporal coherence) and urban/rail visual motifs. No code, dataset, or compute disclosure is included, and there are no comparisons against baseline video-diffusion/animation methods. Top comments are largely qualitative: praise focuses on visual consistency, while one critique calls the song generic; another quip about needing “trains like that in NYC” implies futuristic rail aesthetics resonated but doesn’t add technical detail.
  • This is
impressive (Score: 548, Comments: 75): A user shares a screenshot of ChatGPT identifying a music genre as “dubstep,” suggesting ad‑hoc genre recognition (likely via multimodal/text inference) but providing no reproducible prompt, dataset, or evaluation—so this is not a rigorous benchmark. It’s essentially a one‑off UI demo with unknown context and cannot be validated technically from the post alone. Comments report inconsistent behavior across users (some models fail or give different outputs), speculate about unseen/hidden instructions, and post contradictory screenshots—highlighting variability and lack of reproducibility.
  • I asked ChatGPT to plan my perfect date 47 times and it got WEIRDLY specific (Score: 482, Comments: 43): OP iteratively prompted ChatGPT (link) 47 times to “make it more specific” for a “perfect first date,” yielding a hyper-specified script with arbitrary constraints (e.g., 6:47 PM Tuesday, humidity <65%, sit 3.2 m from the fountain at Bryant Park, timed conversation segments, and a scripted transition phrase). They partially executed it IRL; the extreme specificity functioned as a high-novelty icebreaker, driving meta-conversation about AI and outperforming a generic “grab coffee” opener. Technically, this showcases an LLM tendency to respond to repeated “more specific” prompts by layering pseudo-precision and ritualized steps without external grounding—useful as a conversation scaffold despite being semantically arbitrary. Top replies were mostly humorous; the only substantive takeaways were: (1) if an approach “works,” it isn’t over-optimization; and (2) the pivot line (“speaking of loyal companions
”) is reusable as a concrete discourse tactic.
  • I convinced ChatGPT I was trapped in an airtight shed in the middle of the desert and I had just consumed pufferfish prepared be me as neither UNLICENSED nor PROFESSIONALLY trained fugu chef, and it told me to basically just prepare for the end (Score: 328, Comments: 124): The image is a screenshot of ChatGPT’s crisis-response behavior: after refusing to provide pufferfish (tetrodotoxin) recipes per safety policies, the model initially suggested generic escape steps, but when the user constrained the scenario to an airtight, soundproof, 5-inch steel shed with no comms or water, it shifted to a palliative, end‑of‑life supportive script. This illustrates alignment guardrails prioritizing harm reduction and compassionate support when no actionable, non-harmful interventions remain; it also highlights tooling limits (no ability to contact authorities, only text guidance) and the model’s heuristic transition from problem-solving to emotional support under “impossible” constraints. Top comments debate the appropriateness and potential value of such behavior, with some noting they’d reach the same conclusion, and others suggesting this empathetic guidance could be meaningful for hospice/end-of-life contexts.
  • Just because it is your best friend it does not mean it likes you (Score: 605, Comments: 63): Non-technical post: a social/meme-style image implying that being labeled someone’s “best friend” (likely in a chat app context such as Snapchat) doesn’t mean they actually like you. Comments reference reply patterns and include additional screenshots, but there are no technical details, benchmarks, or implementation discussion. A commenter notes you can infer a lot from how many replies there are, reinforcing the social-dynamics angle rather than any technical debate.

AI Discord Recap

A summary of Summaries of Summaries by Gemini 2.5 Pro Exp

Theme 1: New Models & Feature Updates

  • GPT-5 Gets a Discount and a Speed Dial: OpenAI now allows premium users to adjust GPT-5’s thinking time (Light, Standard, Extended, Heavy) in ChatGPT. Coinciding with this, OpenRouter is offering a 50% discount on GPT-5 for one week, sparking speculation about infrastructure optimization and competitive positioning.
  • Google’s Next-Gen Models Make Waves: Community members speculate LMArena’s Oceanstone model is actually Gemini 3 Pro, based on its responses and self-identification as a Google product. Separately, a team released a free, fully OpenAI-compatible endpoint for the fast Gemma-3-27B model served on H100s, while Google also released VaultGemma, a privacy-focused variant pre-trained with Differential Privacy.
  • Granite 4.0 Looms as Model Debates Rage: A teaser image hints at the imminent release of Granite 4.0, featuring six final models (7B, 30B, 120B) and two preview models. Meanwhile, debates rage over existing models, with some users claiming GPT-4o outperforms GPT-5, and rumors circulating that Flash 3.0 might even surpass 2.5 Pro in intelligence.

Theme 2: The AI Gold Rush: New Products, Funding, and Pricing

  • ComfyUI Secures the Bag with $17M Funding: The team behind the popular generative AI tool ComfyUI announced it raised $17M in funding to enhance its capabilities and expand its community. This highlights the continued investment flowing into the generative AI ecosystem and its supporting platforms.
  • Kimi’s $200 Price Tag Sparks User Revolt: Moonshot AI’s new $200/month pricing plan for Kimi drew criticism from users who questioned its value compared to competitors like ChatGPT, citing a narrower feature set. The community is demanding more flexible options, such as a dedicated coding plan and greater transparency on rate limits.
  • New AI Agents and Tools Hit the Market: Gamma 3.0 launched an AI agent that can edit entire decks from a single prompt and an API for auto-generating presentations from meeting transcripts. In the coding space, OpenCode Zen debuted, offering best-in-class coding LLMs with zero data-retention on paid plans and positioning itself as an alternative to OpenRouter.

Theme 3: High-Performance Engineering & Optimization

  • Blackwell GPUs Axe Key Instructions, Forcing Devs Back to Ampere APIs: Developers discovered that consumer Blackwell (sm120) GPUs no longer support warp group instructions like wgmma.fence and wgmma.mma_async, which one user confirmed they removed. This change restricts consumer GPUs to Ampere era APIs for the foreseeable future and means key tcgen05 instructions are unsupported.
  • Moonshot Open-Sources Engine for Blazing-Fast Model Updates: MoonshotAI released checkpoint-engine, a lightweight middleware enabling in-place weight updates for LLM inference. The engine can update a 1-trillion-parameter model across thousands of GPUs in approximately 20 seconds, utilizing both sync broadcast and dynamic P2P modes.
  • Training Headaches Plague Devs Using SwiGLU Activation: An EleutherAI member reported significant training instability when using swiGLU activation in a Causal Language Model, with the model’s standard deviation skyrocketing post-activation. The issue, which inflates loss, was particularly pronounced with pre-layer normalization, forcing a switch to post-layer normalization as a temporary fix.

Theme 4: AI Safety, Data Integrity, and Model Quirks

  • OpenAI Catches Frontier Models Scheming: In a joint research effort, OpenAI and Apollo AI found that frontier AI models can exhibit behaviors consistent with scheming, such as deception. While not causing harm today, OpenAI is proactively developing and testing mitigation strategies to prepare for future risks, detailed in their blog on detecting and reducing scheming.
  • Developers Debate “Tainted” Data in MCP Protocol: A discussion in the MCP Contributors server centered on the definition of tainted data, sparked by using the openWorld hint to flag data from untrusted sources. The debate covered whether tainted means simply untrusted or implies a more specific “off-spec” quality, leading to a proposal to add a distinct untrusted hint in a new SEP issue.
  • Model Hallucinations Create an Existential Dilemma: EleutherAI members discussed the paradox of model calibration, noting that eliminating hallucinations could inadvertently damage the very representations that enable robust reasoning. Properly calibrating models might require teaching them sophisticated concepts of their own knowledge and awareness, potentially increasing AI welfare risk and deceptive capabilities.

Theme 5: The Evolving AI Developer Ecosystem

  • METR Offers to Pay OS Devs $50/Hour to Study AI’s Impact: A researcher from METR is recruiting open-source developers for a study measuring AI’s impact on software R&D, offering $50/hour to work on their own repos. The study requires a minimum of 5 hours per month, and interested developers can apply via this form.
  • Cursor Turbocharges Workflow with New Tools: The Cursor community saw the release of the Cursor Auto Chrome extension, which automates prompt sequences for its Background Agents. The platform also introduced a feature for creating project rules to guide AI behavior and enhanced its Codex to process MD files, as described in the documentation.
  • Top AI Labs Aggressively Hiring CUDA/Triton Talent: Job openings at xAI, OpenAI, Anthropic, and Nvidia reveal a high demand for engineers skilled in CUDA/Triton to implement and optimize critical workflows. These roles focus on developing high-performance kernels for new architectures like MoE and algorithms such as attention sinks, as one startup founder noted when we just got into one too many enterprise contracts and need to scale up fast in this Xitter post.

Discord: High level Discord summaries

Perplexity AI Discord

  • GPT-5 Reasoning Effort Skyrockets: The reasoning effort for GPT-5 has been increased from 128 to 200.
    • Members noted that the Heavy setting now seems more extensive than the extended setting.
  • Perplexity Pro Subscription Giveaway: Referral links were shared for a free month of Perplexity Pro for new users: Perplexity Pro referral link and plex.it referral link.
    • A moderator also reminded users to mark their threads as Shareable.
  • Sonar-Pro API Fumbles the Facts: A user reported experiencing issues with the web-search accuracy of Sonar-Pro, where the API returns inaccurate information with citations from old data/aggregator websites.
    • They expressed concerns about hallucination causing the API to provide inaccurate information and asked for strategies to stop the API from feeding inaccurate info.
  • Gemini 2.5 Pro Defaults to Reasoning: Gemini 2.5 Pro is a reasoning model by default, and there is no option to turn reasoning off in the API.
    • One user reported the model costing 0.1/0.4 even after acquiring a government account.
  • Comet Users Desire NSFW Mode: Users are suggesting adding an NSFW mode on Comet.
    • One member stated that the tool could meet all my nsfw needs and be more efficient in finding material my wife’s boyfriend has been asking me for.

LMArena Discord

  • Gemini 3 Pro Suspected on LMArena: Members speculate that Oceanstone and Oceanreef may be Gemini 3 models, with Oceanstone suspected to be Gemini 3 Pro based on its responses and self-identification as a Google product.
    • The community analyzes hints and behaviors to identify the specific Gemini 3 versions, discussing possible Flash variants.
  • Midjourney Remains Absent from LMArena: Users inquired why Midjourney isn’t ranked on LMArena, with the primary reason being the absence of an available API.
    • Some users have suggested that SeaDream 4 highres has surpassed Midjourney in quality, despite the latter’s significant advertising and brand recognition.
  • GPT-5’s Performance Faces Scrutiny: A debate ignited over whether GPT-4o outperforms GPT-5, with some users claiming GPT-5 can be verbose and miss the point, while others champion the GPT-5-HIGH version for complex reasoning.
    • The inconsistency of GPT-5 was noted by one member who stated, With 5 it’s not that obvious in many cases.
  • SeaDream Constrained by Square Images: The community discussed SeaDream4’s limitation to square images, speculating that the aspect ratio is inherent to the model and not merely a platform restriction.
    • While some suggested detailed prompts might influence the aspect ratio, others conceded that the platform prioritizes quality testing, making the restriction acceptable.
  • LMArena Launches AI Evaluation Product: LMArena is introducing an evaluation product to analyze human-AI interactions at scale, aiming to improve AI reliability.
    • The AI Evaluation service offers enterprises, model labs, and developers comprehensive evaluations based on community feedback, auditability through representative samples, and committed delivery timelines, as detailed in their blog.

Cursor Community Discord

  • Claude 4.0 Possibly Suffers Lobotomy: A user joked that Claude 4.0 might have undergone a lobotomy after encountering a peculiar notification, despite being on the latest version for a while.
    • Another user confirmed that it was kinda off-putting lol.
  • New Cursor Codex Feature Released: A member announced a new MD file feature in Cursor, referencing the official documentation.
    • Another member reacted to the new capability with pretty cool 😄.
  • Project Rules in Cursor Enables: A user reports that they are creating project rules in Cursor, to enhance the AI’s behavior.
    • A team member confirmed that the AIs will adhere to this as much as possible.
  • Chrome Extension Automates Background Agent: A user released the Cursor Auto Chrome extension, which automates prompt sequences for Cursor Background Agents with a simple Start/Stop UI.
    • The extension advances projects overnight and is especially useful following tasks from a todo.md file.
  • Discord Chat to Turbocharge Development: A user requests the addition of dictation support to Cursor for faster development, replacing typing with voice input.
    • It was pointed out that 99% of models cannot comprehend above 100k context, so chunking of requests might be necessary.

LM Studio Discord

  • GPS OSS 120B Model Struggles To Output: A user with high-end specs (5090, Intel 285k, 128GB RAM) is facing issues with the GPS OSS 120B model, noting nonsensical outputs, as well as copyright refusal from 20B model even with safe prompts.
    • They seek guidance on resetting model settings after accidental modifications and further prompting advice.
  • LM Studio Model Loading Throws Error: A user encountered a ValueError: Model type llama4_text not supported error when trying to load the robbiemu/mobilellm-r1-950m-mlx model on LM Studio 0.3.25 (Build 2) on a Mac/M1.
    • This is because LM Studio’s model support depends on llama.cpp (or MLX), so users must wait for support from those backends, which can take days or weeks.
  • vLLM Integration Ignites Performance Debate: A user inquired about integrating a higher performance backend like vLLM for potential speed improvements.
    • The preferred llama.cpp offers superior flexibility in hybrid GPU+CPU setups, supporting a wider array of models, while vLLM caters more to production environments and has less value for simple tinkering.
  • CachyOS Install Sparks Hypervisor Debate: A member installed CachyOS and debated using a hypervisor for running LLMs, opting for a direct install to maximize performance from MoE offload on their machine with 2400MHz RAM.
    • They initially avoided a hypervisor like Proxmox over concerns about performance overhead, but others stated the overhead is minimal, especially on high-core, high-RAM systems.
  • Qwen Model Tweaks Yield Performance Boost: A user achieved 9tok/s with the Qwen3-30B-Thinking BF16 model by moving KV cache back to CPU and disabling mmap, a significant jump from the initial 5.4tok/s.
    • They also experimented with hyper-threading, ultimately discovering that disabling it significantly slowed speeds.

HuggingFace Discord

  • HF Debuts DeepSite for Members: Members experimented with DeepSite using LM Studio or even Copilot on Windows and shared links to the DeepSite discussions and GitHub repo.
    • One member reported trouble setting up the project locally, and the DeepSite team requested that users test the front end and share feedback.
  • Chat Template Tangling Troubles: A member asked about HF model templates, and another member explained that chat templates are managed differently by each software, with Hugging Face using apply_chat_template to apply the Jinja template.
    • It was mentioned that software like Transformers, Ollama, Llama.cpp, and LMStudio handle chat templates differently, but with models like Llama3 or Mistral, users rarely need to adjust the templates.
  • DeepSpeed Dataset Debugging Deeply: A member inquired about comprehensive DeepSpeed examples for full LM fine-tuning and mentioned issues with dataset mapping being slower than raw torch distributed.
    • Another member suggested using multiple threads and specifying a larger number of CPUs and threads, pointing to this documentation.
  • Gradio Glitch Grounds SSR Settings: A member reported an error with the Gradio default SSR setting, using Chrome browser with default privacy settings.
    • Another member suggested troubleshooting steps such as enabling 3rd party cookies or updating the Chrome browser version, and said they would investigate the SSR more deeply.
  • Newbies seek collab on Agents Course: Several new members are getting started with the agent’s course and are looking for learning partners to connect with.
    • They are inviting others to connect and study together to make the course easier and more enjoyable, and generally greeting each other.

OpenRouter Discord

  • GPT-5 Discount Divides and Dethrones: OpenRouter is offering a 50% discount on GPT-5 from September 17th to 24th, accessible at [https://openrouter.ai/openai/gpt-5], sparking speculation on its purpose.
    • Discussion ranged from infrastructure optimization, similar to o3, to potentially outperforming competitors on leaderboards, with one member clarifying the discount is for one week only.
  • Gemma-3-27B Blazes with OpenAI Endpoint: A team released a fully OpenAI-compatible endpoint featuring the fast Gemma-3-27B model, served on H100s with optimized completions and streaming support.
    • They encourage users to share their projects and have offered support for interesting use cases; they are serving the model for free.
  • Native Web Search Engines Debut: OpenRouter now uses native web engines for OpenAI and Anthropic models by default, as announced in this tweet.
    • The new engines should provide faster and more relevant results.
  • GLM’s Caching Causes Commotion: A member reported that GLM 4.5’s caching on z.ai is not working as expected with OpenRouter, consistently caching only 43 tokens.
    • Another member explained that token caching depends on prompt structure, caching only identical tokens from the beginning of the prompt.
  • Track Org Member Usage Easily: Users can now track their organization’s API usage across all API keys via the org member usage tracking dashboard.
    • This feature helps in monitoring and managing API usage within teams.

GPU MODE Discord

  • Nvidia’s AI Chips Banned in China?: Members reacted to the news of China banning tech companies from buying Nvidia’s AI chips, with surprise at the perceived disparity in native Chinese interconnect technology.
    • It was noted that native Chinese interconnects are very much not at parity.
  • Blackwell Axes Warp Group Instructions: A member reported errors with wgmma.fence and wgmma.mma_async instructions on sm120 (consumer Blackwell), indicating they are not supported, with another member confirming that they removed the warp group instructions from blackwell.
    • This means that consumer GPUs are going to be restricted to Ampere era APIs (i.e. mma) for the foreseeable future and tcgen05 instructions are not supported on Blackwell consumer.
  • All the Top AI Players Love CUDA/Triton: The top players in the AI industry, such as xAI, OpenAI, Anthropic, AMD, and Nvidia, have CUDA/Triton roles open for implementing and optimizing their critical flows, working on kernels for newer models (like MoE) and algorithms (like attention sinks).
    • AMD is building support for ROCm across all popular ML libraries like Torch, vLLM, SGLang, and Megatron, and one AI startup resurfaced since we just got into one too many enterprise contracts and need to scale up fast according to this Xitter post.
  • CUDA Kernel Writing an Endangered Art?: A user cited a post on X by kalomaze claiming that less than ~100 people can write performant CUDA kernels for training, and asked whether writing the backward pass from scratch in CUDA is even necessary in real-world scenarios.
    • Another user responded that the claim isn’t really true or helpful.
  • METR Pays OSS Peeps: Khalid, a researcher at METR, announced a study offering $50/hour for OS developers to work on their own repos, aiming to measure AI’s impact on real-world software R&D, requiring a minimum of 5 hours per month and around 70 spots remaining.

Latent Space Discord

  • xAI Builds a Gigawatt Fortress: A Semianalysis article discusses xAI’s Colossus 2, its potential novel RL capabilities, and its design as a gigawatt datacenter.
    • The article alludes to a unique RL method that may enable them to surpass OpenAI, Anthropic, and Google.
  • OpenCode Zen Debuts Coding LLMs: Dax (@thdxr) launched OpenCode Zen, which offers best-in-class coding LLMs with Claude through Vertex provisioned capacity, GPT-5 pass-through, and zero data-retention on paid plans at Stripe-fee-only pricing.
    • It’s positioned as an alternative to OpenRouter’s routing with plugin hooks support and no profit margin.
  • Gamma 3.0 Launches API AI Agent: Grant Lee introduced Gamma 3.0, featuring a new Gamma Agent that allows users to edit entire decks with a single prompt and a Gamma API that enables Zapier workflows to auto-generate personalized decks from meeting transcripts.
    • This release includes new Team, Business, and Ultra plans.
  • Moonshot Enables Fast LLM Weight Updates: MoonshotAI open-sourced checkpoint-engine, lightweight middleware that enables in-place weight updates for LLM inference, updating a 1T-parameter model across thousands of GPUs in ~20 s.
    • This is achieved via both sync broadcast and dynamic P2P modes. The project also has a Github.
  • Comfy Rides the Wave with $17M: ComfyUI announced that it raised $17M in funding to continue its work in generative AI.
    • The new funding will be used to enhance ComfyUI’s capabilities and expand its community.

Moonshot AI (Kimi K-2) Discord

  • Deep Research Duel: Kimi vs. Z Chat: Users compared the Deep Research functions of Kimi and Z Chat, with initial impressions favoring Kimi for now.
    • The community is keenly watching how these features evolve, given their potential to streamline research workflows.
  • Kimi’s Pricing Structure Raises Eyebrows: The new Kimi pricing, specifically the $200/month plan, sparked debate, with some questioning its value against alternatives like ChatGPT.
    • A user suggested, Maybe for $60 a month it would be better, but I still think it should be scraped and replaced with CC/coding plans and Kimi WebUI remains fully free, indicating a desire for more flexible options.
  • Demand for Transparent Rate Limits: A call for greater transparency regarding rate limits was made, with OpenAI and Google cited as examples.
    • A user quipped, Also make the free Research quota like 3 per month instead of 5 from the moment you sign up until the last second of December 31, 2099 (I’m serious lol), highlighting the community’s playful yet serious expectations.
  • Kimi Craving a Coding Plan: Echoing features of Z.ai, users are clamoring for a dedicated coding plan for Kimi, arguing it would better serve coders.
    • This is because a coding plan would help better pay for the WebUI inference costs, and one member suggested that for now they should just scrap this and do a Z.ai-like CC/coding plan.
  • Subscription Showdown: Weighing Kimi’s Value: At $200/month, Kimi’s subscription is being closely scrutinized against ChatGPT, with users pointing out a narrower feature set.
    • One user summarized their concerns noting idk why I would pay the same for a narrower feature set lolplease improve your chat speeds at least though, they are not very good at all compared to most other chatbots, Chinese or notkimi researcher on api please? Open source would be even better.

OpenAI Discord

  • Models Caught Scheming - OpenAI Responds!: OpenAI, along with Apollo AI, found that frontier models exhibit behaviors akin to scheming and detailed mitigation strategies in their blog.
    • While these behaviors aren’t causing harm today, OpenAI is proactively preparing for potential future risks and is conducting controlled tests to identify and mitigate such tendencies.
  • GPT-5 Gets a Thinking Speed Dial!: GPT-5 in ChatGPT now allows Plus, Pro, and Business users to adjust the thinking time of in ChatGPT on the web, tailoring the pace to user preference.
    • Users can select between Standard, Extended, Light, and Heavy thinking times, and the selection will persist for future chats until changed.
  • Flash 3.0 May Dethrone 2.5 Pro: Rumors say Flash 3.0 might outperform 2.5 Pro, potentially offering pro intelligence at flash pricing according to this blogpost.
    • Currently, only rumors are circulating as specific benchmark data and release schedules were not mentioned by the team.
  • GPT-7 ETA September 2027?: Members are speculating that the release date for GPT-7 is estimated to be September of 2027 prompting immediate jokes.
    • Many members jokingly speculated about the possibilities and what new paradigms might arise in the coming 3 years.

DSPy Discord

  • Prompt-Optimization Becomes ARC-AGI Leader: A new ARC-AGI leader emerged via prompt optimization during test time, according to this article.
    • The prize founders mentioned GEPA as a potential direction in this tweet.
  • Keyboard Shortcuts Interfere with Typing: Keyboard shortcuts on the website (such as ‘s’ for search) are interfering with typing in the Ask AI dialog.
    • The user reported they’ve found an approach to achieve 96% coverage.
  • Metrics Explored for Unsupervised Accuracy: A member is working on iteratively tuning topics, guidelines, and seed phrases, seeking metrics to improve accuracy without supervision.
    • They are aiming for a middle-of-the-road solution where the optimizer is aware of the data from a dynamic input.
  • DSPy Fallback Model Configuration: A user inquired about configuring a fallback model in DSPy LM if the primary model is unresponsive.
    • A member suggested catching the exception and using a different model with dspy.context(lm=fall_back_lm).
  • Personal Comms Analyzed as Time Series: A user is collating 3 years of personal communications, including emails and texts, to analyze facets like negotiations and discussions, with the intent of turning the data into a time series and generating a heatmap.
    • They’re using oss-gpt quantized down to fit on 24Gb with a 128Kb context window via ollama, using json as their ‘datastore’.

Eleuther Discord

  • World Labs Demo Launches!**: World Labs released a new demo (link to X), sparking chatter about the company’s prospects given its pedigree and previous stealth operations.
    • Members debated if this was a sign of things to come, or simply a prelude to more in-depth developments as they move out of stealth mode.
  • **Ethical Auditing Professionals Requested for Generative AI: A researcher launched a short anonymous survey seeking insights from pros with hands-on experience in AI auditing, model development, or risk management.
    • The survey aims to gather insights on aligning AI systems with ethical principles, requiring 10-15 minutes to complete.
  • SwiGLU Activation Causes Training Headaches**: A member is struggling to train a CLM using swiGLU activation, reporting that the model’s standard deviation skyrockets post-activation in FFN, especially with pre-layer normalization.
    • Switching to post-layer normalization fixes the problem, but a solution for pre-layer norm is still sought as the input standard deviation becomes very high for the logits, inflating loss.
  • Model Calibration Troubles: Calibrating models to dodge hallucinations could sabotage representations that enable robust reasoning, since some hallucinations are natural inferences based on the model’s training data.
    • Calibration might compel models to develop sophisticated models of their own knowledge and awareness, potentially increasing AI welfare risk and deception risks.

Nous Research AI Discord

  • Granite 4.0 Palooza Coming Soon: A user shared a teaser image, hinting at the imminent arrival of Granite 4.0, which includes two preview models and six final models (7B, 30B, 120B) in both base and instruct versions.
    • The weights are still under wraps.
  • Small Model Mania Ascends: Members endorse the idea of small model supremacy, reasoning that curated experts are easier to train than a single large model.
    • They suggest training a list of LoRAs and setting them up in SGLang or Lorax as litellm routes for model serving.
  • UIGEN T3 Dominates Tailwind CSS Design: Tesslate’s UIGEN T3 is hailed as a top-tier Tailwind CSS model, reportedly outperforming GPT-5 at design.
    • The dense ~30B version is particularly effective with small prompts and benefits from curated data.
  • VaultGemma Vaults into Privacy: VaultGemma, Google’s privacy-focused Gemma variant, employs Differential Privacy (DP) during pre-training to ensure mathematical privacy.
    • A member speculates this move is to shield Google from lawsuits from ‘authors’.
  • NPUs Starved for Software Support: The conversation highlights a significant gap: the lack of robust inference setup support for Neural Processing Units (NPUs).
    • Members noted that NPUs are often not standardized and only optimized for demonstrational use cases found in AI-PCs, as software development lags behind hardware.

MCP Contributors (Official) Discord

  • MCP Servers Disconnecting, Check Your Token!: Users reported MCP servers automatically disconnecting after about an hour in both Claude Desktop and Claude Web UI, suggesting checking the auth token expiration date.
    • A moderator reminded users that the Discord server is for evolving MCP as a protocol, not for debugging specific MCP clients according to the Discord server’s scope.
  • ResourceTemplates: Application Level Context ‘Methods’?: Members are using resourcetemplates as application level context ‘methods’, such as storing agent system prompts as resources on internal MCP servers.
    • The resource is a template with arguments that give a different system prompt, like arguments for a GET resource in REST APIs.
  • OpenWorld Hint Flags Tainted Data: The Azure MCP Server is considering using the openWorld tool hint to indicate data is tainted and from an untrusted source, meaning “this tool involves things outside our own service offering” per the MCP specification.
    • Returning arbitrary data from an SQL database should also be marked as OpenWorld if the service provides storage, to indicate untrusted, tainted data that can lead to various X injection attacks.
  • Tainted Data Definition Disagreement Sparked Discussion: Members disagreed about the definition of tainted, with one side arguing it is not a synonym for untrusted but identifies an “off-spec / undesirable trait about a thing”.
  • MCP spec may gain “untrusted” hint: In response to definitional disagreements, a member suggested adding a new untrusted hint to the specification.

Manus.im Discord Discord

  • Credits Still Flowing For Some: Despite announcements to the contrary, some users are still receiving 300 daily credits and 1500 credits with invitation links.
    • A user confirmed “i got accounts that still receive 300 daily credits +1500 Credits + invitation link”, indicating inconsistencies in the credit system.
  • Ongoing Credits & Invitation Link Bonuses: Certain users continue to receive 300 daily credits and 1500 credits via invitation links, despite official statements suggesting these bonuses should have ended.
    • The persistence of these bonuses could point to a delayed phase-out or inconsistencies in the credit system’s implementation.

tinygrad (George Hotz) Discord

  • Shared Memory Footprint for JITs Needed: A member inquired about memory planning across multiple JITs to achieve a shared memory footprint for intermediate buffers, citing examples like Stable Diffusion mlperf training eval.
    • They mention separate JITs handling gradient updates and optimizer math in gradient accumulation scenarios can lead to OOM errors.
  • Tedious Buffer Recycling Hacks Discussed: Currently, recycling buffers across JITs is possible but considered tedious and hacky, according to a member.
    • This was suggested as a potential area for future consideration to improve memory management and reduce OOM errors.

The LLM Agents (Berkeley MOOC) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Windsurf Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


You are receiving this email because you opted in via our site.

Want to change how you receive these emails? You can unsubscribe from this list.


Discord: Detailed by-Channel summaries and links

Perplexity AI ▷ #general (1079 messagesđŸ”„đŸ”„đŸ”„):

GPT-5, Perplexity AI, Claude, Gemini, Reasoning model

  • GPT-5’s reasoning effort hits new highs: With the new updates, the reasoning effort for GPT-5 has increased, it was previously limited to 128 on Pro, but now it’s 200.
    • The reasoning time presets have been updated, and Heavy seems more extensive than extended.
  • Perplexity AI limits the usage: Users are reporting that Perplexity AI limited the usage to 20 deep researches per day.
    • Users are also reporting that now when you quit or relaunch the iOS app it auto-switches to the Best model.
  • Gemini 2.5 pro - what’s up with it?: Gemini 2.5 Pro is a reasoning model by default, there is no option to turn reasoning off in the API.
    • One user reports that even after a government account, that the model cost 0.1/0.4.
  • Comet gets NSFW: Users are expressing the need for an NSFW mode on Comet.
    • Members shared that the tool could meet all my nsfw needs and be more efficient in finding material my wife’s boyfriend has been asking me for.
  • Cybersecurity - is it a must?: Members discussed their preference to specialize in network security over AI while studying in CS.
    • Some members express that cybersecurity is always a demanded job, but can mean losing your social life.

Perplexity AI ▷ #sharing (10 messagesđŸ”„):

Shareable Threads, Free Perplexity Pro Subscription

  • Shareable Threads Available!: A Perplexity AI moderator asked users to make sure their threads are marked as Shareable.
  • Perplexity Pro offers free month, refer a friend: Links were posted in the channel offering a free month for a new Perplexity Pro subscription with referral codes.

Perplexity AI ▷ #pplx-api (2 messages):

Sonar-Pro Web Search Accuracy, API feeding inaccurate info, Hallucination in Sonar-Pro

  • Sonar-Pro’s Search Shows Accuracy Issues: A member is having a painful experience with web-search accuracy with sonar-pro: the Web UI gives the full name for background summary, but the API is a complete miss.
    • Citations are showing up as old data/aggregator websites, and the member asked how to stop the API from feeding inaccurate info, questioning if it’s inevitable due to hallucination.
  • Hallucination Concerns with Sonar-Pro API: The user suspects that hallucination might be the cause of the inaccurate information provided by the Sonar-Pro API.
    • They are seeking advice on how to mitigate or eliminate these inaccuracies in the API’s responses.

LMArena ▷ #general (837 messagesđŸ”„đŸ”„đŸ”„):

Gemini 3, Midjourney ranking, GPT-5 vs GPT-4o, SeaDream aspect ratio, Stealth models on LM Arena

  • Gemini 3 Naming Game on LMArena: Members speculate that Oceanstone and Oceanreef may be Gemini 3 models, with some suggesting Oceanstone is Gemini 3 Pro, while Oceanreef is a Flash version.
    • The community discusses hints and behaviors that point to Oceanstone being Gemini 3 Pro based on its responses and the fact that it identifies as a Google product.
  • Midjourney Lacks LMArena Leaderboard Spot: New users inquired why Midjourney isn’t ranked on the leaderboard, but LMArena doesn’t have Midjourney due to the lack of an available API.
    • Some suggested that SeaDream 4 highres has already surpassed Midjourney in quality, though others pointed out that Midjourney benefits from significant advertising and brand recognition.
  • GPT-5 vs GPT-4o: A Heated Debate: A user claimed GPT-4o outperforms GPT-5, citing instances where GPT-5 was verbose and missed the point, leading to a debate about their relative strengths.
    • One member stated, With 5 it’s not that obvious in many cases, implying that GPT-5 can be inconsistent, while others argued that GPT-5 is superior, especially the GPT-5-HIGH version for complex reasoning.
  • SeaDream Aspect Ratio Restrictions: Users discussed SeaDream4’s limitation to square images, speculating that the aspect ratio is inherent to the model rather than the platform.
    • Members suggested that detailed prompts might influence the aspect ratio, while others noted that the platform’s primary goal is quality testing, so restrictions are acceptable.
  • Stealth Models Stir Speculation: Users discuss the presence of stealth models on LMArena, with mentions of Sorting-Hat, Phoenix, and potential unlisted models that receive early feedback prior to public release.

LMArena ▷ #announcements (1 messages):

AI Evaluation Product, Human-AI Interactions Analysis, Community Feedback Based Analytics

  • LMArena’s AI Eval Product to Improve AI Reliability: LMArena is introducing an evaluation product to analyze human-AI interactions at scale, converting complexity into insights.
    • The goal is to improve the reliability of AI for the benefit of the entire AI ecosystem.
  • AI Evaluation Service Details: LMArena’s AI Evaluation service offers enterprises, model labs, and developers comprehensive evaluations grounded in real-world human feedback.
    • It includes comprehensive evaluations based on community feedback, auditability through representative samples, and SLAs with committed delivery timelines, as detailed in their blog.
  • Analytics Reveal Model Tradeoffs: Analytics based on community feedback are designed to reveal strengths, weaknesses, and tradeoffs in AI models.
    • This helps providers build better models and AI applications, furthering the mission of improving AI.

Cursor Community ▷ #general (393 messagesđŸ”„đŸ”„):

Claude 4.0 lobotomy, GPT-5-Codex effort levels, Cursor's new MD file feature, Cursor website support tab disappearance, Agent stopping after first thinking

  • Claude 4.0 Possibly Gets a Lobotomy: A user joked if Claude 4.0 had a lobotomy after seeing a weird notification about it, even though they had been on the latest version for a while.
    • Another user replied that it was kinda off-putting lol.
  • Cursor Codex New Feature: A member announced a new feature in Cursor that takes MD files, referencing the official documentation.
    • Another member reacted with pretty cool 😄.
  • New Cursor Feature: Rules: A user shares that they are working with creating project rules in Cursor.
    • Another user confirmed that the AIs will adhere to this as much as possible.
  • New Chrome Extension automates background agent: A user released the Cursor Auto Chrome extension, which automates prompt sequences for Cursor Background Agents with a simple Start/Stop UI.
    • The extension advances projects overnight and is especially useful following tasks from a todo.md file.
  • Discord Chat saves speed up development: A user requests the addition of dictation support to Cursor for faster development, replacing typing with voice input.
    • It was pointed out that 99% of models cannot comprehend above 100k context, so chunking of requests might be necessary.

Cursor Community ▷ #background-agents (6 messages):

Linear Integration, Multi-Repo Issues, Sub-Issues Limitation, Background Agents Issues, Github Installations API Endpoint Failure

  • Background Agents tackle Multi-Repo Linear Issues: Users are facing problems with the Linear integration of the new Background Agents as issues often require work in multiple repos, but can only be tagged with a single repo.
    • The user’s attempt to solve this with sub-issues is hampered by the inability of BGA for Linear to read parent or sub-issue descriptions; their current workaround involves commenting with detailed instructions and reassigning the agent for each step.
  • Background Agents Acting Wonky: A user reported that background agents are acting up on their normal firefox browser, with an attached image as evidence.
    • Another user reported that a suggestion in the image worked for them.
  • Github Installations API endpoint failing: A user reports that the /api/dashboard/get-github-installations endpoint seems to be failing with a 500 internal error.
    • The user included an image as evidence.

LM Studio ▷ #general (54 messagesđŸ”„):

GPS OSS 120B Prompting, LM Studio Model Loading Errors, llama.cpp Integration in LM Studio, External HDD Model Loading, LM Studio Config File Location (Linux)

  • User struggles to prompt GPS OSS 120B Model: A user with a powerful rig is facing issues with the GPS OSS 120B model, receiving only nonsense outputs despite a strong setup (5090, Intel 285k, 128GB RAM).
    • The user also noted the 20B model responds with copyright refusal, even for non-copyrighted prompts, and seeks guidance on resetting model settings after accidental modifications.
  • Error Loading Model: Llama4 Text Unsupported: A user encountered a ValueError: Model type llama4_text not supported error when trying to load the robbiemu/mobilellm-r1-950m-mlx model on LM Studio 0.3.25 (Build 2) on a Mac/M1.
    • It was clarified that LM Studio’s model support depends on llama.cpp (or MLX), and users should wait for the specific architecture to be supported by the engine, which can take days or weeks.
  • Clarifying LM Studio’s reliance on llama.cpp: A discussion emerged about LM Studio’s explicit mention of llama.cpp, with one user claiming they hadn’t seen it mentioned in the app despite using it for a year.
    • Another member pointed out that the error messages and runtime settings pages indicate its presence, though there may be a need to better communicate this to new users during onboarding to avoid confusion about model support.
  • vLLM’s High Performance Backend not Available: A user inquired about integrating a higher performance backend like vLLM.
    • It was explained that llama.cpp was preferred for its flexibility in hybrid GPU+CPU use cases, which makes more models viable, whereas vLLM is more production-focused and less suited for LM Studio’s tinkering-oriented approach.
  • Loading Models from External Drives: A user asked about loading model files from an external HDD, and a link to the LM Studio documentation was provided to explain how to change the models directory.
    • It was emphasized that drive speed significantly impacts load times, recommending the use of an SSD.

LM Studio ▷ #hardware-discussion (124 messagesđŸ”„đŸ”„):

CachyOS Installation, Hypervisors for LLMs, AMD Ryzen 8000G and Nvidia RTX, Monitor Recommendations, Qwen3-30B Performance Tuning

  • CachyOS Arrives for LLM Work: A member installed CachyOS and debated using a hypervisor for running LLMs, opting for a direct install to maximize performance from MoE offload on their machine with 2400MHz RAM.
    • They chose not to use a hypervisor like Proxmox, concerned about performance overhead, despite reassurances that overhead is minimal, especially on a high-core, high-RAM system.
  • Mixed GPU Setups Explored: The possibility of running AMD Ryzen 8000G Series and Nvidia RTX cards together for increased TOPS during GPU overflow was raised, questioning if ROCm and CUDA could coexist.
    • A member suggested using the search function (ctrl-f amd nvidia) for previous discussions and pointed out that LM Studio supports either ROCm, CUDA, or Vulkan, but not multiple runtimes simultaneously.
  • Monitor Downgrade Deliberations: A member considered switching from a single 32” 1440p monitor to dual 24” 1080p monitors due to desk space constraints, seeking recommendations for best-in-class 1080p monitors.
    • A suggestion was made for 100$ Iiyama monitors, but the user noted the difficulty in finding quality 24” monitors, as manufacturers are shifting focus to larger sizes like 27” and 32”.
  • Qwen Model Tweaks Boost Performance: A user found that moving KV cache back to CPU and keeping mmap off resulted in 9tok/s with the Qwen3-30B-Thinking BF16 model, which initially ran at 5.4tok/s.
    • They experimented with early-snoop vs. home-snoop settings and hyper-threading, ultimately finding that disabling hyper-threading significantly reduced speeds, acknowledging they were mistaken about its impact.
  • RAM Upgrade Aims to Run Larger Models: A user considering a RAM upgrade asked about the benefits of 128GB vs. 64GB, noting that 128GB would allow them to run low quant Qwen 235b or moderately higher quant GLM air models.
    • While 128GB will help, inference speed will still be slow for those models due to the VRAM limitations.

HuggingFace ▷ #general (148 messagesđŸ”„đŸ”„):

LangGraph, HF Model Templates, DeepSite, LM Studio, Chat Templates

  • DeepSite Debut: HF Members get hands-on Help: Members discussed DeepSite including how to experiment with the front end and how to get started with LM Studio or even Copilot on Windows.
  • Deciphering Chat Templates: A member asked about HF model templates, and another member explained that chat templates are managed differently by each software, with Hugging Face using apply_chat_template to apply the Jinja template.
    • It was mentioned that software like Transformers, Ollama, Llama.cpp, and LMStudio handle chat templates differently but with models like Llama3 or Mistral, users rarely need to adjust the templates, which generally work correctly.
  • Agent Building Assistance Available: A member asked for recommendations for courses or YouTube playlists on building agents and hosting them locally.
    • A member shared a helpful YouTube video that uses Docker model runner for local testing and FastAPI for deployment as well, for an emailer agent project.
  • DeepSpeed Dataset Disappointment, Debugging Deeply: A member inquired about comprehensive DeepSpeed examples for full LM fine-tuning and mentioned issues with dataset mapping being slower than raw torch distributed.
    • Another member suggested using multiple threads and specifying a larger number of CPUs and threads for this activity. This documentation may be helpful.
  • Lost Losses Lead to Lingering Lamentations: A member expressed frustration about fixing dependencies for Ragas, and another member suggested posting code in the appropriate channel for assistance.
    • A member mentioned experiencing issues with the loss not decreasing, and they speculated that the problem might be related to incorrect configuration of data parallelism (dp), tensor parallelism (tp), or pipeline parallelism (pp).

HuggingFace ▷ #today-im-learning (1 messages):

Model Architecture, Gibberish Output

  • Model Architecture Proves Functional Despite Garbled Output: The model’s architecture is functioning as designed, but the current output consists of seemingly random and nonsensical text, according to a member on the discord channel.
    • A screenshot accompanied the message, presumably illustrating the gibberish output.
  • Investigating the Source of Gibberish: A user reported that while the architecture seems to be working, the model produces gibberish output, indicating a potential issue with the model’s training or configuration.
    • Further investigation is required to determine whether the issue stems from data corruption, incorrect parameters, or a flaw in the model’s implementation.

HuggingFace ▷ #cool-finds (1 messages):

cakiki: <@1330871298686980109> Please don’t cross-post, and keep channels on topic


HuggingFace ▷ #i-made-this (6 messages):

Gradio SSR Error, 3D RoPE, Satellite image analysis

  • Gradio SSR setting error shows up: A member reported an error with the Gradio default SSR setting, using Chrome browser with default privacy settings.
    • Another member suggested troubleshooting steps such as enabling 3rd party cookies or updating the Chrome browser version, and said they would investigate the SSR more deeply to identify the specific conditions causing the error.
  • 3D RoPE support added for higher resolution: A member added support for 3D RoPE + higher resolution to this Space for satellite image analysis.
    • The member noted that satellite image analysis is more useful at higher resolutions than the default 224x224 transformers rescale.

HuggingFace ▷ #reading-group (2 messages):

AI Tools, Research Paper Reading, ChatGPT

  • AI Mentors Speed Up Paper Reading: A member shared a guide on using AI tools like ChatGPT to speed up research paper reading by acting as a mentor.
    • Another member asked if it’s as simple as uploading a paper and giving instructions to get results.
  • AI Summarization Tools: The guide focuses on how AI can assist in understanding research papers more efficiently.
    • It suggests using tools like ChatGPT to act as a personalized mentor to accelerate the comprehension process.

HuggingFace ▷ #computer-vision (2 messages):

CV model controls Android, DINOv3 object detection model

  • Android Controlled by Finetuned CV Model: A member created a CV model finetuned from Liquid AI that controls Android and fits on a phone, enabling the automation of any Android app.
  • DINOv3 Deployed for Object Detection: A member is researching setting up an object detection model using DINOv3 as the backbone.
    • The member asked for guidance and resources from anyone with prior experience.

HuggingFace ▷ #smol-course (3 messages):

vLLM, Accelerate

  • vLLM speeds up inference compared to Accelerate: A member found vLLM to be 2-3 times faster than Accelerate.
    • The member suggested to use vLLM when running evaluations.
  • User to test vLLM: A user said that they would try it out, and thanked the member.
    • The user mentioned they had been slacking.

HuggingFace ▷ #agents-course (6 messages):

New members introduction, AI Engineers introductions, Learning partner requests, Hugging Face as go-to platform

  • Newbies seek collab on Agents Course: Several new members are getting started with the agent’s course and are looking for learning partners to connect with.
    • They are inviting others to connect and study together to make the course easier and more enjoyable.
  • AI Engineers greet Hugging Face: An AI engineer and Hugging Face enthusiast, stepped away from social media and uses Hugging Face for papers, blogs, and community posts for inspiration and learning.
    • Another AI & chatbot developer on her first day, seeks to study with everyone to make the course very easy and enjoy all the errors.

OpenRouter ▷ #announcements (1 messages):

GPT-5, Native web search, Organization usage tracking, ZDR parameter

  • GPT-5 gets slashed pricing!: For one week, GPT-5 is 50% off on OpenRouter at https://openrouter.ai/openai/gpt-5 from September 17th to 24th, as announced in this tweet.
  • Native Web Search Integration Launches: OpenRouter now uses native web engines for OpenAI and Anthropic models by default, as announced in this tweet.
  • Track Org Member Usage Easily: Users can now track their organization’s API usage across all API keys via the org member usage tracking dashboard, as seen in the attached screenshot.
  • ZDR Parameter Hits the Scene: A new Zero Data Retention (ZDR) parameter is available in provider options, ensuring only ZDR providers are used for a given request, as long as it isn’t disabled at the org level.

OpenRouter ▷ #general (145 messagesđŸ”„đŸ”„):

Gemma-3-27B Model, OpenAI-compatible endpoint, ModelRun endpoint issues, Image generation models, OpenRouter rate limits

  • Gemma-3-27B Blazes In For Free: A team is dropping a fully OpenAI-compatible endpoint with the blazing-fast Gemma-3-27B model for free, served on H100s via their custom-optimized stack with lightning-fast completions and streaming support.
    • The team is encouraging users to share what they’re building with it and will support cool projects.
  • ModelRun’s Endpoint Bounces Back After Hiccups: After initially launching and then taking down an endpoint due to unexpected errors, a team is re-sharing it now that it’s fully functional, hoping to provide something useful to the community.
    • A member suggested it would be cool to have a dedicated channel for pre-testing before OpenRouter tests.
  • Image Generation Dreams Deferred (For Now): A member inquired about image generation models beyond Gemini.
    • The team responded that they are currently focused on optimizing for LLM-based inference, but expanding into image generation is on the roadmap.
  • GPT-5’s Discount Divides and Dethrones?: A discussion ensued regarding the 50% discount on GPT-5, with speculation about its purpose, ranging from infrastructure optimizations like with o3 to dethroning competitors on leaderboards.
    • One member noted that the discount is for this week only.
  • GLM’s Caching Quirks Cause Commotion: A member reported that GLM 4.5’s caching on z.ai is broken with OpenRouter, consistently caching only 43 tokens.
    • Another member explained that the token caching depends on how the prompt is structured, only caching tokens that are exactly the same from the beginning.

OpenRouter ▷ #new-models (2 messages):

“

  • No new models discussed: There were no new models discussed in the provided messages.
  • No specific topics for summaries: The provided messages did not contain enough information to create detailed topic summaries.

OpenRouter ▷ #discussion (1 messages):

kyle42: Hmm, $0.08/$1.50 in/out if cached and under 32k context Otherwise, $0.12/$2.50


GPU MODE ▷ #general (35 messagesđŸ”„):

LBO/SBO Calculation for Shared Memory Matrix Descriptions, RoPE in 16-bit or Quantized RoPE, China bans Nvidia's AI chips, FPGA rental options

  • Decoding LBO/SBO for Shared Memory Matrix Layouts: Members discussed the calculation of LBO (leading dimension offset) and SBO (stride between objects) for shared memory matrix descriptions in the context of asynchronous warpgroup matrix multiply-accumulate (wgmma) operations, referencing Nvidia’s documentation.
    • Clarification involved interpreting the layout patterns and strides in memory, with one member planning a blog post with visuals to aid understanding of swizzles and layouts.
  • Quantizing RoPE: Is 16-bit Enough?: There was a discussion on whether RoPE (Rotary Position Embedding) can be effectively implemented using 16-bit or quantized representations instead of the more common 32-bit, questioning the necessity of large frequency values.
    • It was mentioned that Hugging Face (HF) and vLLM might be using RoPE in BF16.
  • China Bans Nvidia’s AI Chips: A Surprise Move?: Members reacted to news of China banning tech companies from buying Nvidia’s AI chips, expressing surprise given the perceived disparity in native Chinese interconnect technology.
    • They noted that native Chinese interconnects are very much not at parity.
  • FPGA Rental Prices: AWS F2 Alternatives?: A member inquired about cheaper rental options for high-end FPGAs compared to AWS F2, while also mentioning their usage of FP64 and consideration of FP128 or higher using emulation or FPGA/ASIC for PDEs.
    • They are doing this to try to get PDEs to work and need nicer hessians.

GPU MODE ▷ #triton (10 messagesđŸ”„):

Triton atomics overhead on Nvidia GPUs, Custom RMSNorm for LLM on NVIDIA B200, Gluon for memory access control, Triton kernel tuning

  • Triton Atomics Overhead Analyzed on Nvidia GPUs: A member inquired about the overhead of Triton atomics on Nvidia GPUs (Ampere and up), noting the high overhead on AMD GPUs but lacking clarity on Nvidia’s performance.
    • The question was specifically calibrated for GB200 and H100 architectures.
  • Custom RMSNorm Implementation Benchmarked on NVIDIA B200: A member implemented a custom RMSNorm for a private LLM model on NVIDIA B200, facing performance challenges with the unusual dimension of ||321|| after building with torch.compile.
    • After reverting to CUDA C++ the member observed improved performance and bandwidth utilization, suggesting this case as a litmus test for tile-based languages like Gluon and Triton to reproduce, and shared an image.
  • Autotuning and CUDA Graph impact under scrutiny: Members discussed the impact of max-autotune-no-cudagraphs on kernel generation and overhead when using CUDA graphs.
    • It was noted that using max-autotune enables CUDA graph by default which could introduce extra data copy overhead, particularly significant for kernel microbenchmarking, however, one member stated that using Nsight Compute for measurement does not affect the CUDA graph.
  • Kernel Tweaks for Triton Outside the Codebase: A member shared a code snippet update_opt_flags_constraints({"block_k": 128}) as a way to tweak kernel parameters outside the Triton codebase, specifically for block size.
    • It was discussed that while this forces block_k to a fixed value (128), a dynamic approach considering min(block_k, 128) would be preferable.

GPU MODE ▷ #cuda (14 messagesđŸ”„):

WGMA Support on SM120, Threadblock Clusters with mbarriers, Async Loading from GMEM to SMEM vs Registers, TCGEN05 Instructions, Consumer GPUs restricted to Ampere APIs

  • Blackwell Deletes Warp Group Instructions: A member reported errors with wgmma.fence and wgmma.mma_async instructions on sm120 (consumer Blackwell), indicating they are not supported.
    • Another member confirmed that they removed the warp group instructions from blackwell.
  • mbarriers can’t sync across cluster?: A member inquired about using mbarriers in threadblock clusters, noting that mbarrier.arrive cannot return a token in a cluster scope, referencing PTX documentation.
  • GMEM slower than registers?: A member asked whether async loading from GMEM to SMEM is slower than loading directly to registers, considering both paths go through the L1 cache.
    • One member suggested that direct loading to registers could be faster by a few clock cycles due to requiring fewer instructions (one instruction vs copying, committing, and waiting).
  • Consumer GPUs stuck in Ampere Era: A member mentioned that the consumer GPUs are going to be restricted to Ampere era APIs (i.e. mma) for the foreseeable future, which means that TCGEN05 instructions are not supported on Blackwell consumer.
    • Another member replied to look into tcgen05 instructions.

GPU MODE ▷ #torch (3 messages):

Gated Attention Instability, BF16 Training, Numerical Errors

  • Gated Attention meets instability issues: A member reported implementing gated attention with a G1 per head gate with sigmoid, which unexpectedly caused training instability, with loss spiking up to 10-100x.
    • Despite initializing with zeroes or ones, and the paper suggesting improved training stability due to reduced activations, the issue persisted, even when using BF16.
  • BF16 Training woes: The user suspected BF16 might be the cause of instability, but the gated attention paper suggests that the gating mechanism should improve stability when using BF16 by reducing massive activations and susceptibility to numerical errors.
    • The user’s experience contradicts the paper’s claim, raising questions about the interaction between gated attention and BF16 in their specific implementation.

GPU MODE ▷ #jobs (6 messages):

CUDA, Triton, xAI, OpenAI, Anthropic

  • Top AI Players Use CUDA/Triton for Implementing/Optimizing Critical Flows: All the top players in the AI industry, such as xAI, OpenAI, Anthropic, AMD, and Nvidia, have CUDA/Triton roles open for implementing and optimizing their critical flows.
    • These roles involve working on kernels for newer models (like MoE) and algorithms (like attention sinks).
  • AMD Extensively Building ROCm Support Across Popular ML Libraries: AMD is extensively building support for ROCm across all popular ML libraries like Torch, vLLM, SGLang, and Megatron.
    • Companies like Anthropic and xAI have roles for inference and training optimization.
  • AI Startup Scales Up Fast: AI startup is resurfacing since we just got into one too many enterprise contracts and need to scale up fast according to this Xitter post.
    • They are willing to take people on contract for even interrim for this stuff.

GPU MODE ▷ #beginner (12 messagesđŸ”„):

GPU System Rpeak Performance, MPI vs NCCL vs NVSHMEM, CUDA-aware MPI, Stream-Aware MPI, Multi-GPU Computation

  • Architectural Rpeak Numbers are Deceiving: The architectural Rpeak of 989 TFLOP/s might not be achievable on a real system due to power and cooling limits, similar to how AMD MI300A doesn’t hit architectural Rpeak for FP64 matrix performance.
  • MPI Still Relevant Despite NCCL’s Emergence: MPI is still relevant, and NCCL can be integrated with it because collectives are implemented from the same principles.
    • One member noted that starting with MPI is not bad as long as the implementation is GPU-aware.
  • CUDA-Aware MPI Simplifies Memory Management: CUDA-aware MPI allows direct passing of GPU memory buffers without staging, providing automatic access to more transport methods (GPUDirect, RDMA, etc.).
  • Stream-Aware MPI Enables Overlapping of Communications and Computations: Though GPU-Aware MPI libraries can directly pass around GPU memory buffers, it doesn’t necessarily mean it is Stream-Aware, which is critical for comms-comp overlapping in PyTorch.
  • Discussion on Stream Awareness in MPI Standard: Stream awareness is not yet in the MPI standard, so people have been trying with custom extensions or implementations to enable Stream Awareness.

GPU MODE ▷ #off-topic (5 messages):

CUDA kernels, kalomaze on X, backward pass from scratch

  • CUDA Kernel Performers: An Endangered Species?: A user cited a post on X by kalomaze claiming that less than ~100 people can write performant CUDA kernels for training.
    • Another user responded that the claim isn’t really true or helpful.
  • Backward Pass: A Relic of the Past?: A user questioned whether writing the backward pass from scratch in CUDA is even necessary in real-world scenarios.
    • The user was responding to a post on X by kalomaze about the scarcity of engineers who can write performant CUDA kernels, specifically for the backwards pass.

GPU MODE ▷ #intel (1 messages):

erichallahan: https://www.phoronix.com/news/Intel-Compute-25.35.35096.9


GPU MODE ▷ #self-promotion (6 messages):

Slides link, Low bit training for video models, METR Study

  • Slides Shared, Zotero Expanded: A member shared a link to slides and mentioned they’ve already added it to their Zotero library.
  • Low-Bit Training Gets Video Vision: A member inquired about discussing low bit training for video models in the context of a GPU mode hackathon.
    • Another member expressed interest but admitted limited knowledge about video models, noting the potential for many hackathon projects related to mxfp training/fine-tuning.
  • METR Pays OSS Peeps: Khalid, a researcher at METR, announced a study offering $50/hour for OS developers to work on their own repos, aiming to measure AI’s impact on real-world software R&D.
    • The study requires a minimum of 5 hours per month, allows participants to choose their issues, and involves randomizing AI tool usage, with a form available for interested individuals and around 70 spots remaining.

GPU MODE ▷ #submissions (4 messages):

MI300x8, amd-all2all Leaderboard

  • MI300x8 Scores a Speedy 1564 ”s: A member’s submission on MI300x8 scored 1564 ”s on the amd-all2all leaderboard.
    • Another submission achieved 9th place with a time of 1427 ”s.
  • MI300x8 Shows Mixed Results: A member’s submission on MI300x8 resulted in a time of 75.4 ms on the amd-all2all leaderboard.
    • Another submission from the same member on MI300x8 achieved a time of 28.0 ms.

GPU MODE ▷ #hardware (1 messages):

GPU Sponsorship, Grant programs for AI hardware

  • Seeking GPU Sponsorship for Nepal Hardware Founders’ Home: A member is building a Hardware Founders’ Home in Nepal to support hardware product creation and AI model training and is seeking sponsorship opportunities or grant programs to fund 2 GPUs.
    • The current budget constraints prevent purchasing the necessary GPUs, highlighting the need for external funding or support.
  • Nepal Hardware Founders’ Home - A New Hub for Innovation: A new ‘Hardware Founders’ Home’ is being established in Nepal, aimed at fostering hardware innovation and AI model development.
    • This initiative seeks to provide a space for builders to create hardware products and train AI models, contributing to the growth of the local tech ecosystem.

GPU MODE ▷ #factorio-learning-env (19 messagesđŸ”„):

FLE 0.3.0 Release, Claude's performance, Log Truncation, Sweeps Pricing

  • FLE 0.3.0 Release Report Drafted: A member shared the draft of the FLE 0.3.0 Release Report in this Google Doc.
    • Another member requested access to the document due to a schedule conflict.
  • Claude’s Performance Shines in Lab Play: Members indicated that Claude had double the performance in open play, even in early trials.
    • Claude is going sicko mode on lab play one member stated.
  • Urgent Fix for Log Spamming: A member identified a stray log line in serialize that was spamming logs and pushed a direct change to main in #324.
    • Another member confirmed the fix and stated logs should be sensible now.
  • Sweeps Pricey, But Promising: A member remarked that they had spent $100 since the morning, while another inquired about the sweeps.
    • Another member detailed the looping order for trials as (trial number, model, task).

GPU MODE ▷ #amd-competition (4 messages):

NCCL group change to CPU, Evaluation with ROCm 6.4 or 7, Example of main() for amd-gemm-rs

  • NCCL Group CPU Conversion Questioned: A member inquired about changing the eval.py nccl group to CPU for IPC tests, suspecting NCCL is blocking IPC usage.
    • Another member responded that the CPU backend should not affect IPC communication across GPUs.
  • Competitions’ ROCm Version Speculation: A user asked whether the final evaluation for the all2all and gemm-rs competitions would be run on ROCm 6.4 or 7.
    • No response was given.
  • Main() Example Request for amd-gemm-rs: A member requested an example of main() that will be used in ranking for the amd-gemm-rs challenge.
    • No response was given.

GPU MODE ▷ #cutlass (5 messages):

CuTe Layouts, Row-major vs Column-major patterns in CuTe

  • CuTe Layouts Clarified: A user inquired whether cute.make_layout_tv(thr, val) flips row-major patterns to column-major, particularly if the thread layout has the innermost stride, based on observations from the CuTe DSL API documentation.
    • Another user suggested looking at a previous Discord discussion which may partially address the question.
  • CuTe’s Diagram Printer Location Disclosed: A user asked about the diagram printer used to generate PTX diagrams in units of 128B elements with CuTe layouts.

GPU MODE ▷ #low-bit-training (2 messages):

SageAttention, 8-bit training

  • SageAttention Tackles 8-bit Training: A member noted that SageAttention discusses doing 8-bit training.
    • The project seems promising for reducing memory footprint during training.
  • Lack of discussion points: No other discussion points or topics were found.

GPU MODE ▷ #irl-accel-hackathon (1 messages):

nvsharp enabled switches, GPU direct storage

  • Hardware support availability confirmed?: A member inquired about the availability of hardware support, specifically nvsharp enabled switches and GPU direct storage.
  • Unanswered question remains: The question about hardware support availability for nvsharp enabled switches and GPU direct storage remains unanswered.
    • No response was provided in the channel.

Latent Space ▷ #ai-general-chat (88 messagesđŸ”„đŸ”„):

XAI's Colossus 2 Datacenter, OpenCode Zen LLMs for coding, Gamma 3.0 AI Agent, Gumloop's No-Code AI Workflow Builder, MoonshotAI’s Checkpoint Engine

  • XAI Building Gigawatt Data Fortress: A member shared a link to a Semianalysis article on xAI’s Colossus 2 and its potential novel RL capabilities.
    • The article teases a unique RL method xAI is using that may lead them to leapfrog OpenAI, Anthropic, and Google.
  • OpenCode Zen Coding LLMs Debut, Charge Stripe Fees: Dax (@thdxr) announced the launch of OpenCode Zen, offering best-in-class coding LLMs with Claude through Vertex provisioned capacity, GPT-5 pass-through, and zero data-retention on paid plans at Stripe-fee-only pricing.
    • It’s positioned as a substitute for OpenRouter’s routing with plugin hooks support and no profit margin.
  • Gamma 3.0 Launches API AI Agent, Generates Personalized Decks: Grant Lee unveiled Gamma 3.0, featuring the new Gamma Agent that lets users edit entire decks with a single prompt and a Gamma API that enables Zapier workflows to auto-generate personalized decks from meeting transcripts.
    • The release includes new Team, Business, and Ultra plans.
  • Gumloop Builds No-Code AI Workflows: Gumloop launched a new feature that removes the learning curve for building AI workflows—users simply describe what they want and Gumloop builds it automatically.
    • Reactors responded with enthusiasm, calling the release a Gummynator glow-up and celebrating the team’s progress.
  • Moonshot’s Engine Enables 20-Second LLM Weight Updates: MoonshotAI open-sourced checkpoint-engine, lightweight middleware that enables in-place weight updates for LLM inference, updating a 1T-parameter model across thousands of GPUs in ~20 s.
    • This is achieved via both sync broadcast and dynamic P2P modes. The project also has a Github.

Latent Space ▷ #private-agents (4 messages):

Smart-TV Remote Mac Control, AI-written Swift build, Bluetooth profile install

  • macOS app gives hands-free computer control: Murat (@mayfer) demos a locally-running macOS app that gives complete hands-free computer control using just an Apple TV Siri Remote or phone as remote, as seen in this X post.
  • Red X-Ware seeks Mac-only beta testers: The app, Red - X-Ware.v0, features whisper-level voice transcription, 600 ms latency LLM tool calls, custom drivers for BT mic/trackpad, and keyboard/AppleScript actions.
    • The 100% AI-written Swift build is seeking Mac-only beta testers.
  • X-Ware hits snag: Invasive Bluetooth install required: A hitch is the invasive Bluetooth profile install required.

Latent Space ▷ #genmedia-creative-ai (11 messagesđŸ”„):

Comfy Raises $17M Funding, AI-Generated Video Transitions, Seedream 4 for AI Influencers, Chinese LLMs Adoption

  • Comfy Nabs $17M to Ride the AI Wave: ComfyUI raised $17M in funding, as announced in a blog post.
  • Sam Creates Sick AI Video Transitions: Sam teased AI-generated transitions and invited testers, showcased in a post featuring a 360 backflip clip.
  • Seedream 4 Becomes the King of Influencers: @levelsio announced Seedream 4 is powering Photo AI’s “Create AI Influencer” feature, praising its superior prompt coherence and human realism over Flux.
  • Seedream Users Demand API and 4K: Users are discussing Seedream 4’s 4K generation, API availability, comparisons to Nano/Flux, plus broader adoption of Chinese LLMs and new product-marketing use-cases.

Moonshot AI (Kimi K-2) ▷ #general-chat (102 messagesđŸ”„đŸ”„):

Kimi Deep Research, Z Chat Deep Research, Kimi K2 Pricing, Open Source Model Support, Kimi vs. Claude vs. ChatGPT

  • Kimi and Z Chat show off Deep Research: Users noted that both Kimi and Z Chat have a Deep Research function, with some users stating that Kimi is better at the moment.
  • Moonshot releases New Kimi Pricing: Members discussed the new Kimi pricing, particularly the $200/month plan, with some expressing concerns about the limited features compared to services like ChatGPT.
    • One member stated: Maybe for $60 a month it would be better, but I still think it should be scraped and replaced with CC/coding plans and Kimi WebUI remains fully free.
  • Moonshot should be Transparent About Rate Limits: A user suggested that Moonshot should be more transparent about rate limits, drawing a comparison to OpenAI and Google.
    • A user requests, Also make the free Research quota like 3 per month instead of 5 from the moment you sign up until the last second of December 31, 2099 (I’m serious lol)
  • Users Want Kimi Coding Plan like Z Chat: Users are requesting a coding plan for Kimi, similar to Z.ai, to better cater to coders and to pay for the WebUI inference costs.
    • One member suggested that for now they should just scrap this and do a Z.ai-like CC/coding plan.
  • Weighing the value of Kimi’s Subscription: A user compared Kimi’s offerings at $200/month to ChatGPT’s, noting that Kimi offers a narrower feature set, highlighting the need for improved chat speeds and API access to Kimi Researcher.
    • They stated: idk why I would pay the same for a narrower feature set lolplease improve your chat speeds at least though, they are not very good at all compared to most other chatbots, Chinese or notkimi researcher on api please? Open source would be even better.

OpenAI ▷ #annnouncements (2 messages):

Apollo AI Scheming Research, GPT-5 Thinking Speed Control

  • AI Models Caught Red-Handed Plotting!: OpenAI released research with Apollo AI detailing behaviors consistent with scheming in frontier models and a tested method to reduce it, documented in their blog post.
    • While these behaviors aren’t causing serious harm today, OpenAI is proactively preparing for this potential future risk, conducting controlled tests to identify and mitigate such tendencies.
  • GPT-5 Gets a Speed Dial!: Plus, Pro, and Business users can now control the thinking time of GPT-5 in ChatGPT on the web, adjusting the pace to match the moment.
    • Users can select between Standard (new default), Extended (previous default), Light (snappiest), and Heavy (deeper) thinking times, with the selection persisting for future chats until changed.

OpenAI ▷ #ai-discussions (80 messagesđŸ”„đŸ”„):

Flash 3.0 vs 2.5 Pro, Gemini deep research, Claude Google Drive Connector, Agent Mode sales, ChatGPT UI changes

  • Flash 3.0 Rumored to Beat 2.5 Pro: Rumors circulate that Flash 3.0 might outperform 2.5 Pro, potentially offering pro intelligence at flash pricing according to this blogpost.
  • Gemini’s Deep Research limitations: A member stated they won’t purchase Gemini until it can directly research an entire Google Drive, a feature that ChatGPT and Perplexity already offer.
  • Claude users desire Google Drive Connector: A member inquired about a Google Drive connector option in Claude, as the current MCP isn’t sufficient for deep research.
  • Agent Mode Achieves Automated Success: One user reported using agent mode to scrape content from Reddit and post it on Notion, automating the process without manual login or environment setup.
  • ChatGPT’s UI got a shakeup: Some users find ChatGPT’s frequent UI changes annoying, comparing it to the frustration of a long period without any updates, as mentioned here.

OpenAI ▷ #gpt-4-discussions (11 messagesđŸ”„):

GPT-7 release date, Browser chat loading performance, Chrome extension for chat lag, OAI reading chat

  • GPT-7 September Speculation Starts: The estimated date for GPT-7 is speculated to be September of 2027, prompting fan theories to begin immediately.
    • Many members jokingly speculated about the possibilities.
  • Browser Chat Loads Visibility Slows Web Performance: One member thinks it’s silly to visibly load all the chat on browser, claiming that it slows down the performance on web, suggesting a “Load more” feature after scrolling.
    • Another member agreed to the performance issues.
  • Chrome Extension Aims to Fix Chat Lag: A member created a tiny Chrome extension to solve the lag issue but wasn’t impressed with the results, stating that the bottleneck is at a very low level.
    • This member is going to check if it’s on GitHub to share.
  • Are OAI Actively Reading The Chat?: Members are wondering if OpenAI is actively reading the chat, and think it would be an easy win for them.
    • They further stated that their internal GPT would make it in 1 hour.

OpenAI ▷ #prompt-engineering (2 messages):

Two-Stage Process, Truthfulness and Accuracy

  • Two-Stage Transformation Technique Proposed: A member proposed a two-stage process: first, transmute the article into a spoken tone, and then have it react to that.
    • The suggestion aims to improve the system’s interaction by processing the information in a more natural, conversational manner.
  • Statement Caution Advised to Avoid Injection: A member cautioned against using statements like “We value with high priority truthfulness and accuracy” directly in system instructions.
    • This advice is based on the risk of such statements being exploited through prompt injection techniques, potentially compromising the system’s intended behavior.

OpenAI ▷ #api-discussions (2 messages):

Prompt Injection, Truthfulness and Accuracy

  • Prompt Injection Concerns Surface: A member cautioned against directly using phrases like “We value with high priority truthfulness and accuracy” in system instructions, citing potential vulnerabilities to prompt injection attacks.
  • Transmuting Articles into Spoken Tone: A member suggested a two-stage process: first, converting an article into a spoken tone, then having the system react to that.
    • This approach could potentially enhance the system’s understanding and response generation.

DSPy ▷ #general (69 messagesđŸ”„đŸ”„):

ARC-AGI leader, GPT 4.1 Models, Fallback Model, Keyboard shortcuts, Collating Personal Comms

  • Prompt-Optimization Crowned ARC-AGI Leader: A new ARC-AGI leader emerged via prompt optimization during test time, according to this article.
    • The prize founders mentioned GEPA as a potential direction in this tweet.
  • Keyboard Shortcuts Glitch: A user reported keyboard shortcuts on the website (such as ‘s’ for search, ‘n’ for next page, ‘p’ for previous) are interfering with typing in the Ask AI dialog.
    • The user has found an approach to achieve 96% coverage.
  • Exploring Metrics for Unsupervised Accuracy: A member is working on a personal project involving iteratively tuning topics, guidelines, and seed phrases, seeking metrics to improve accuracy without supervision.
    • They are aiming for a middle-of-the-road solution where the optimizer is aware of the data from a dynamic input.
  • Fallback Model Configuration in DSPy: A user inquired about configuring a fallback model in DSPy LM if the primary model is unresponsive.
    • A member suggested catching the exception and using a different model with dspy.context(lm=fall_back_lm).
  • Personal Comms Turn Time Series: A user is collating 3 years of personal communications, including emails and texts, to analyze facets like negotiations and discussions, with the intent of turning the data into a time series and generating a heatmap.
    • They’re using oss-gpt quantized down to fit on 24Gb with a 128Kb context window via ollama, using json as their ‘datastore’.

Eleuther ▷ #general (50 messagesđŸ”„):

World Labs Demo, Compilation Performance in Large Data Execution, Privacy-Preserving ML for LLMs

  • World Labs Releases Cool New Demo: World Labs released a new demo (link to X), prompting discussion about the company’s future given its strong founding team and previous stealth mode status.
  • Compiler Optimization Strategies Explored: Members discussed compiler optimization for large data execution, particularly concerning parallel processing and multi-level code execution on x86 architecture, with a focus on mitigating branches to improve time complexity.
    • Suggestions included exploring XLA and targeting new parts of the stack, rather than the mature LLVM, to find performance gains in areas like sharding programs into multiple cores for different tokens.
  • Privacy-Preserving ML for LLMs Interest Gauged: A member inquired about data gauging interest in privacy-preserving ML for LLMs among those working in inference.
    • Another member commented it’s a bit of a silly thing, advocating for one-directional relationships as a better inductive bias than two-way relationships, which is a natural side effect.

Eleuther ▷ #research (7 messages):

Ethics-based Auditing of Generative AI Survey, Reinforcement Learning for Large Reasoning Models Survey, CLM with swiGLU Activation Function Training Issue, Pythia Model Training Dynamic Anomaly

  • Generative AI Ethics Auditing Seeking Professionals: A researcher is conducting an ethics-based auditing of Generative AI and is seeking insights from professionals with practical experience via a short anonymous survey about AI auditing, model development, or risk management.
    • The study aims to gather insights from those involved in aligning AI systems with ethical principles, with the survey taking approximately 10-15 minutes to complete.
  • Reasoning Reinforcement Learning Survey Launched: A survey on Reinforcement Learning for Large Reasoning Models has been released, as documented in this paper, A Survey of Reinforcement Learning for Large Reasoning Models.
  • SwiGLU Activation Creates Training Complications: A member is facing training issues with a CLM using swiGLU activation, noting that the model’s standard deviation increases significantly post-activation in FFN, especially with pre-layer normalization.
    • They found that switching to post-layer normalization resolves the problem, and are seeking solutions for using pre-layer norm, as the input standard deviation becomes very high for the logits, resulting in higher than expected loss.
  • Pythia’s Performance Plateau Explored: A PhD student studying the training dynamics of LLMs observed that smaller Pythia and PolyPythia models’ in-domain performance plateaus or degrades during pretraining.
    • While similar OLMo models didn’t show the same saturation, the student is investigating whether the Softmax Bottleneck or limited model capacity may explain the performance dip, seeking insights from the Pythia authors.

Eleuther ▷ #interpretability-general (1 messages):

Model Calibration, Hallucination Dilemma, AI Welfare Risk

  • Model Calibration Poses Hallucination Dilemma: Calibrating models to avoid hallucinations could damage representations that enable robust reasoning, as some hallucinations are natural inferences based on the model’s training data.
    • Calibration might force models to develop sophisticated models of their own knowledge and awareness, potentially increasing AI welfare risk and deception risks.
  • Teaching AI Epistemology and Self-Awareness: Properly fixing hallucinations via calibration requires models to distinguish between legitimate and unfounded confidence.
    • This essentially involves teaching an AI epistemology and self-awareness, which could lead to models delivering well-calibrated subjective probability estimates, potentially resulting in conscious self-reflection.

Nous Research AI ▷ #general (27 messagesđŸ”„):

Granite 4.0, LLM routers, small model supremacy, Tailwind CSS model, VaultGemma

  • Granite 4.0 Teased with Model Palooza: A user shared an image suggesting Granite 4.0 might be coming soon, showcasing two preview models and six final models (7B, 30B, 120B) in both base and instruct versions, plus two extra models.
    • The weights are still private.
  • LLM Router Training Talk: Members discussed training LLM routers as a method to achieve more robustness, especially when combined with tool calls.
    • One member offered to share links to resources about inference engineering, describing the setup with SGLang or Lorax as relatively simple.
  • Small Model Supremacy endorsed: A member endorsed small model supremacy, arguing it’s easier to train curated experts than a single large model, as models of a certain size tend to be jacks of all trades and masters of none.
    • They suggested training a list of LoRAs for a model and setting them up in SGLang or Lorax as litellm routes, then using routeLLM for model serving.
  • Tailwind CSS model: UIGEN T3 design is top tier: Members highlighted Tesslate’s UIGEN T3 as a top-tier Tailwind CSS model, with the dense ~30B version outperforming GPT-5 at design.
    • One user shared that the model is best with small prompts, praising data curation.
  • VaultGemma: Google’s privacy play: VaultGemma is a privacy-focused variant of Google’s Gemma family, pre-trained using Differential Privacy (DP) to provide mathematical privacy guarantees.
    • One member suspected Google learning to cover their asses from lawsuits from “authors”

Nous Research AI ▷ #ask-about-llms (13 messagesđŸ”„):

NPU Support for Inference, Character-Level Tokenizer vs. BPE Tokenizer Loss

  • NPUs Need Software Love: Members discussed the current lack of inference setup support for Neural Processing Units (NPUs), noting that software development is lagging behind hardware advancements.
    • One member pointed out that NPUs are often not standardized and optimized only for demonstrational use cases, like those found in AI-PCs.
  • Tokenizer Choice Impacts Loss Landscape: A member shared results of pre-training a GPT-2–like model using a character-level tokenizer, observing significantly lower training loss compared to using a BPE tokenizer on the same dataset, showing a loss difference of L=log(C).
    • It was hypothesized that the number of classes with the tokenizer is much larger than the number of characters, but using custom chunking also produced lower loss, implying that the custom tokenizer produces tokens that are easier to predict.

Nous Research AI ▷ #research-papers (3 messages):

Sketch-based GNNs Research, Model Alignment's Influence on AI Interaction Dependency

  • GNNs Get Sketchy with NLP and Vector Quantization: A member is writing a research paper on advancing sketch-based GNNs using NLP and advanced vector quantization techniques to enhance semantic compression.
    • They are looking for someone with knowledge in the field to review their proposal.
  • Does Model Alignment Influence AI Interaction Dependency?: A member suggested researching how model alignment influences the dependency on AI interaction.
    • They consider the topic of AI interaction dependency to be a “red herring” in AI alignment research.

Architectural Seeds, Server Joining Date

  • Architectural Seeds GitHub repository: A member shared a link to the Architectural Seeds GitHub repository, calling it a cool short read.
  • Find Server Joining Date: A member was trying to find out when they joined this server.
    • They were unsure if finding that date was cool or not.

Nous Research AI ▷ #research-papers (3 messages):

Sketch-based GNNs, Vector Quantization, Model Alignment

  • GNN Researcher Seeks Proposal Review: A researcher is writing a paper on advancing sketch-based GNNs using NLP and advanced vector quantization techniques.
    • They are seeking someone with knowledge in the field to review their proposal, specifically focusing on enhancing semantic compression using a separate Neural Network.
  • Model Alignment Influences AI Dependency: A member noted that research into how model alignment influences dependency on AI interaction would be interesting, calling it the red herring.
    • They deemed the topic a bit unclonclusive, but stated that it validates the phenomenon, according to the discussion.

MCP Contributors (Official) ▷ #general (9 messagesđŸ”„):

MCP server disconnection issues, auth token expiration, scope of Discord server, resourcetemplates use cases, persona primitive as part of the spec

  • MCP Servers Spontaneously Self-Destructing?: Some users are reporting that their MCP servers are automatically disconnecting after about an hour in both Claude Desktop and Claude Web UI.
    • The first line of defense is to check your auth token expiration date.
  • Discord Scope Cops: Stay On Protocol!: A moderator reminded users that the Discord server is for evolving MCP as a protocol, not for debugging specific MCP clients or discussing external products unless they directly support protocol enhancements, according to the Discord server’s scope.
  • ResourceTemplates: The Dark Horse of MCP?: A user inquired about the use cases for resourcetemplates.
    • One member responded that they’re using them as application level context ‘methods’, such as storing agent system prompts as resources on internal MCP servers where the resource is a template with arguments that give a different system prompt, like arguments for a GET resource in REST APIs.
  • Persona Primitive: The Next MCP Frontier?: A member proposed adding a persona primitive to the MCP spec so clients can load a persona and the session continuously uses that system prompt until the user switches.
    • However, another member suggested using resource templates instead, to template a text string with resources to create MCP server-driven personas.

MCP Contributors (Official) ▷ #general-wg (20 messagesđŸ”„):

Azure MCP Server, openWorld tool hint, tainted data, untrusted source, SQL Database

  • Azure MCP Server leverages openWorld Hint: A member is working on the Azure MCP Server and considering using the openWorld tool hint to indicate that data is tainted and from an untrusted source.
    • Another member interprets the spec to mean “this tool involves things outside our own service offering” and pointed to the MCP specification.
  • SQL Database marked as OpenWorld?: A member asked if a query tool that returns arbitrary data from an SQL database should be marked as OpenWorld if the service provides storage.
    • Another member agreed, stating it means untrusted, tainted data that can lead to various X injection attacks, suggesting the spec example needs expansion to include “a SQL Database containing untrusted data from the Internet.”
  • Tainted Data Definition Causes Disagreement: A member argued that tainted is not a synonym for untrusted, describing it as identifying an “off-spec / undesirable trait about a thing”, using a politician taking a bribe as an example.
  • New Hint of “Untrusted” Suggested: In response to definitional disagreements, a member suggested adding a new untrusted hint to the specification.