Apple finally gives in.
AI News for 1/12/2026-1/13/2026. We checked 12 subreddits, 544 Twitters and 24 Discords (204 channels, and 1541 messages) for you. Estimated reading time saved (at 200wpm): 157 minutes. Our new website is now up with full metadata search and beautiful vibe coded presentation of all past issues. See https://news.smol.ai/ for the full news breakdowns and give us feedback on @smol_ai!
The only people more sympathetic than Appleâs struggling AI division are Appleâs users who have not had the promised updates to Siri. After over a year of talking about their own inhouse trained models, Apple announced that they are picking Googleâs Gemini to power Siri, with a security layer still provided by Apple Private Cloud Compute. A nice win for Google and a comparative loss for OpenAI, which Apple initially partnered with for their AI launch but is increasingly having rumors of their consumer hardware device shipping this year, which may have factored into the decision.
AI Twitter Recap
Top tweets (by engagement)
- Apple â Google AI partnership: A joint statement says the ânext generation of Apple Foundation Modelsâ will be based on Googleâs Gemini models and cloud technology, powering future Apple Intelligence features including a âmore personalized Siriâ while maintaining Apple privacy posture (tweet).
- Anthropic launches âCoworkâ: Claudeâs âClaude Code for the rest of your workâ product preview drives major engagement and discussion about âLLM OSâ trajectories (announcement, context).
- OpenAI in health: OpenAI introduces ChatGPT Health (a dedicated space with separated memories) and announces acquisition of Torch (ChatGPT Health, acquisition).
- DeepSeek âEngramâ: Viral technical thread highlights DeepSeekâs new conditional memory / lookup module as a new âaxis of sparsityâ (thread).
DeepSeekâs Engram: conditional memory as a new sparsity primitive
- Engram = O(1) lookup-style memory for static patterns: DeepSeek introduces âConditional Memory via Scalable Lookupâ (Engram), adding a hashed nâgram embedding memory that the model can query and gate into representations. The framing: transformers spend compute reconstructing memorized patterns via forward passes; Engram offloads that âstatic retrievalâ so backbone capacity can focus on âeffective depthâ/reasoning (@scaling01). Follow-up notes claim it helps long-context as well (tweet).
- Why it matters for systems + scaling: Multiple takes converge on the idea that deterministic hashing/lookup enables hardware-friendly optimizations (prefetching, memory movement) and shifts the scaling constraint away from HBM-bound parameters, suggesting a practical route to grow âknowledge capacityâ cheaply while keeping FLOPs stable (@tokenbender critique + summary, @teortaxesTex). A key quoted lineââconditional memory as an indispensable modeling primitiveâ for next-gen sparse modelsâgets repeated as the thesis (quote tweet).
- Relation to prior art (Gemma-3n / N-Grammer / PLE / OTT): Engineers map Engram onto earlier âinject nâgram info at embeddingsâ approaches (Over-Tokenized Transformers, Per-Layer Embeddings, NâGrammer), arguing Engram differs by making memory an active, layer-addressed operation rather than passive early injection (comparison, @arohan). Others push back that pieces existed âin Gemmaâ3n,â with debate focusing on framing (edge efficiency vs frontier scaling) and architectural âaestheticsâ vs end-to-end capability orientation (discussion).
- Community read: promising but âsystems-brainedâ and potentially brittle: A detailed critique questions whether âalways fetch + gateâ is the right abstraction vs adaptive compute, and worries about OOD mixing or optimization complexity, while acknowledging the iso-budget gains appear real but modest (~3â5% in the criticâs read) (@tokenbender).
Long-context & memory research: DroPE, agent memory, and test-time training
- DroPE (Sakana AI): extend context by dropping positional embeddings: Sakanaâs âDroPEâ proposes using RoPE for convergence, then removing positional encodings to avoid semantic distortion during context extension. The claim is a principled middle ground between NoPE training difficulties (vanishing gradients) and RoPE scaling warping low frequencies (paper tweet, Japanese explainer). They also release a reference trainer implementation (repo).
- Test-time training as âmemoryâ: TTTâE2E: NVIDIA/Stanford/Astera push âEnd-to-End Test-Time Trainingâ where models continue next-token training on the provided context at deployment time, effectively compressing salient context into weights and reducing reliance on large KV caches for long sequences. Several posts frame this as âa new era for LLM memoryâ and a path toward subquadratic long-sequence modeling (NVIDIAAIDev, @karansdalal, @jure).
- Agent memory frameworks converge on policy-integrated memory ops:
- AgeMem: treats long-term + short-term memory as a single learnable policy with tool-like actions (ADD/UPDATE/DELETE + RETRIEVE/SUMMARY/FILTER), trained via a staged RL strategy and step-wise GRPO. Reported gains include +13% on Qwen2.5â7B vs Mem0 and larger gaps on Qwen3â4B (@omarsar0).
- SimpleMem: focuses on âsemantic lossless compressionâ + consolidation + query-aware retrieval; claims 43.24 F1 on LoCoMo vs 34.20 for Mem0 and 30Ă lower tokens per query (531 vs 16,910) (@dair_ai).
Agents & developer tooling: eval infra, CLI-first workflows, and âLLM OSâ moves
- Anthropic âCoworkâ: Claude Code UX for non-coding work: Anthropic positions Cowork as âClaude Code for the rest of your work,â bundling an agent with browser automation, connectors, and a sandboxed execution environment. The launch triggers a wave of âLLM OSâ commentary and âagentic knowledge workâ framing (launch, product details, âLLM OSâ take).
- Internal coding agents go mainstream (Ramp âInspectâ): Ramp reports their internal agent wrote 30% of merged frontend+backend PRs in a week, running fully in the cloud on top of open tooling (opencode, Modal, Cloudflare). They open-source the âblueprint/specâ to build a similar system (@zachbruggeman, build post).
- Agentic evaluation = infra problem (AI21 SWE-bench at scale): AI21 says they ran SWE-bench 200k+ times and the biggest lesson was infrastructure: provision per instance (repo+deps+MCP server) and reuse; separate generation from evaluation so failed tests can be retried without re-generating tokens. They claim failure rate dropped 30% â 0 and repo downloads 8,000+ â 500 (thread start).
- CLI as a low-token agent interface: Hugging Face echoes âBash is all you needâ logic and ships composable/discoverable Hub CLIs to let coding agents explore models/datasets/Spaces with low context usage (@hanouticelina).
- Recursive Language Models (RLM) for million-token workflows: RLM proposes recursively chunking prompts and aggregating results, with trajectories usable for RL/distillation; itâs being integrated into OpenEnv (overview, OpenEnv note).
Big tech platform shifts: AppleâGemini, Google âagentic commerce,â and API primitives
- Apple bets on Gemini for Siri / âApple Foundation Modelsâ: The joint statement explicitly ties ânext-gen Apple Foundation Modelsâ to Gemini models and cloud tech, while emphasizing on-device + Private Cloud Compute and privacy standards (statement). Commentary emphasizes the strategic logic: Gemini multimodality leadership and competitive tension with OpenAIâs device ambitions (analysis).
- Google pushes agentic commerce rails: Google announces Universal Commerce Protocol (UCP) + features like direct checkout in AI Mode / Gemini, âBusiness Agentâ chat with retailers, and âDirect Offersâ pilotsâi.e., making the model the shopping surface and the transaction initiator (@Google).
- Gemini API input scaling: Gemini API increases inline file limit 20MB â 100MB, adds native ingestion from Google Cloud Storage buckets, and supports signed URLs (enabling other clouds). Separate posts specify max sizes like 2GB for registered GCS files and 100MB for external signed URLs, plus supported document formats (@osanseviero, @_philschmid).
OpenAI in healthcare: product boundaries, acquisition, and privacy posture
- ChatGPT Health = separate memory domain: OpenAI introduces âChatGPT Healthâ where health chats/files/memories live in a dedicated space; health info ânever flows into your regular chats,â and users can view/delete health memories (announcement).
- Acquires Torch Health: OpenAI acquires Torch, described as unifying lab results, meds, and visit recordings, positioned as accelerating ChatGPT Health capabilities (OpenAI, Torch founder). A broader recap claims 230M weekly users use ChatGPT for health-related messages and outlines HIPAA-compliant âChatGPT for Healthcareâ and an API posture (recap).
China AI business + open ecosystem signals (IPO narratives, adoption metrics, and model distribution)
- Zhipu vs MiniMax IPO âstory mismatchâ: A ZhihuFrontier summary argues divergent market performance stems from different narratives: Zhipu as ToB/ToG infra with long sales cycles and heavy R&D burn vs MiniMax as consumer/global platform with growth curves and improving margins (thread).
- Open model adoption metrics get more rigorous: A âRelative Adoption Metric (RAM Score)â aims to normalize Hugging Face downloads by time and size, arguing 1â9B models dominate raw downloads but median top-10 downloads differ by only ~4Ă between 1â9B and 100B+. The metric flags GPTâOSS as exceptional and suggests varying momentum among large Chinese MoE releases (@natolambert).
- GLMâ4.7 distribution + fast inference: Mentions include GLMâ4.7 availability via Cerebras on Hugging Face and via Together AI, with marketing claims like 200K context and strong coding benchmarks (as stated in vendor tweets) (@NielsRogge, Together).
AI Reddit Recap
/r/LocalLlama + /r/localLLM Recap
1. New AI Model and Benchmark Releases
-
GitHub - deepseek-ai/Engram: Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models (Activity: 324): DeepSeek AI has introduced a novel approach in their Engram repository, enhancing large language models with conditional memory via scalable lookup. This method integrates static N-gram memory with dynamic hidden states, offering a new axis of sparsity beyond traditional MoE architectures. The model demonstrates a U-shaped scaling law for optimal capacity allocation between MoE and Engram, allowing for efficient offloading of embedding tables to host memory with minimal performance impact. The use of the Muon optimizer in ablations suggests a shift from AdamW, aligning with trends seen in models like Kimi K2 and GLM 4.5. For more details, visit the GitHub repository. Commenters highlight the potential of Engram to improve model performance per parameter, with significant parts offloadable to RAM or NVMe without performance penalties. The deterministic addressing and n-gram embedding approach are praised for their innovative contribution to model efficiency and scalability.
- The DeepSeek team has introduced a model with conditional memory functions, which they describe as a âcurrent stable metaâ and foresee as essential for next-generation sparse models. The model uses mHC (đ = 4) for ablations, indicating a stable configuration. The Engram model is expected to enhance performance per parameter compared to traditional MoE architectures, allowing significant portions of the model to be offloaded to RAM or NVMe without performance loss. This could mean a 40B A3.8B MoE model would only require 27B of weights in fast memory, with the rest comfortably offloaded.
- The Engram model introduces a novel n-gram embedding approach, adding static memory as a new sparsity axis alongside MoE. This allows for O(1) lookup, which is a significant efficiency improvement. The modelâs deterministic addressing enables embedding tables to be offloaded to host memory with minimal inference overhead. A U-shaped scaling law was discovered, guiding the allocation of capacity between MoE and Engram, which helps preserve model depth for complex reasoning tasks.
- The use of the Muon optimizer in the Engram modelâs ablations suggests a shift from the traditional AdamW optimizer, aligning with trends seen in models like Kimi K2 and GLM 4.5. This choice of optimizer could influence the training efficiency and performance of next-generation models.
-
11 Production LLM Serving Engines (vLLM vs TGI vs Ollama) (Activity: 5): The article provides a comparative analysis of 11 production LLM serving engines, including VLLM, TGI, and Ollama, focusing on their performance metrics, scalability, and integration capabilities. It highlights the unique features and use cases of each engine, emphasizing their roles in effective LLM deployment. For further details, refer to the original article. A comment suggests additional engines like Mistral.rs, AirLLM, and Nexa SDK, indicating a broader landscape of LLM serving solutions. Another comment questions the inclusion of Ollama in a production context, hinting at skepticism about its maturity or suitability for production environments.
- The discussion highlights additional LLM serving engines beyond the ones mentioned in the post, such as Mistral.rs, AirLLM, Nexa SDK, TabbyAPI, Exllama, Aphrodite, CoboldCPP, KTransformers, exa, and TextSynth Server. These tools are suggested as alternatives or complements to vLLM, TGI, and Ollama, indicating a diverse ecosystem of solutions for deploying large language models in production environments.
2. AI Development and System Administration
-
Weak Dev, Good SysAdmin needing advice (Activity: 19): The user has acquired a Beelink Mini PC, GTR9 Pro with an AMD Ryzen AI Max+ 395 CPU and
128GB RAMfor local AI development and gaming. They plan to transition from Windows to Linux, starting with PowerShell module development from existing scripts. The user seeks advice on whether to use native Windows tools or start with Windows Subsystem for Linux (WSL). The systemâs specifications suggest it can handle both Windows and Linux environments effectively, allowing for flexibility in development and testing. One commenter suggests that the userâs system is capable of running both Windows and Linux environments, either natively or through virtual machines, providing flexibility in tool choice. Another commenter inquires about the specific type of software the user intends to develop, indicating interest in the userâs development goals. -
Which LLM would be the âbestâ coding tutor? (Activity: 8): The discussion centers on identifying the most effective large language model (LLM) for coding instruction. A notable mention is Qwen3 30b coder, which operates efficiently on a MacBook Air with
24GBRAM, alongside OpenAIâs 20b model, which is also noted for its speed on similar hardware. The use of LLM Studio is recommended for experimenting with various models, highlighting significant advancements in LLM capabilities by 2025. One commenter suggests that all LLMs surpass individual coders in coding tasks, but emphasizes the importance of learning how to interpret and verify the information provided by these models, as they can sometimes provide incorrect or fabricated data.- Qwen3 30b coder is highlighted for its performance, running efficiently on a MacBook Air with 24GB of RAM. This suggests that the model is optimized for resource-constrained environments, making it accessible for personal use. Additionally, OpenAI 20b is noted for its speed on the same hardware, indicating that both models are suitable for users with similar setups.
- The mention of LLM Studio suggests a tool or platform that allows users to experiment with different language models. This could be particularly useful for those looking to compare performance and capabilities across various models, especially given the rapid improvements in LLMs by 2025.
3. Humorous Takes on AI and Data Privacy
-
It seems like people donât understand what they are doing? (Activity: 16): The image is a meme that humorously critiques the casual attitude some individuals may have towards data privacy, particularly in the context of using AI tools like âClaude Codeâ at work. The caption implies a satirical scenario where employees unknowingly compromise their employerâs data security for personal convenience, such as leaving work early. This reflects broader concerns about data privacy and security in the workplace, especially with the increasing use of AI tools that may inadvertently expose sensitive information. One commenter expresses frustration with current hardware limitations and the need to invest in better equipment due to rising prices. Another shares their struggle with running multiple tools on limited hardware, highlighting the challenges of balancing local and remote computing resources for AI tasks.
- A user discusses the limitations of running multiple tools on a 16GB M1 MacMini, highlighting the challenges of local model inference due to insufficient memory. They mention using LM Studio for remote inference, which is slow, and express frustration with using the OpenRouter API for coding, as it requires stripping identifiable code and still fails 25% of the time. This reflects broader issues with hardware constraints and the inefficiencies of current workaround solutions.
-
Env vars donât work when your agent can read the environment (Activity: 2): The post discusses a security concern where environment variables (env vars) are not secure if an agent running on the system can read the environment. This is particularly relevant in scenarios where sensitive information, such as API keys or passwords, is stored in env vars. The issue arises because any process with sufficient permissions can access these variables, potentially leading to unauthorized access or data leaks. The post likely emphasizes the need for alternative secure storage solutions for sensitive data, such as using secret management tools or encrypted storage, to mitigate this risk. Commenters debate the effectiveness of using environment variables for sensitive data, with some suggesting that while convenient, they are not secure enough for production environments. Others recommend using dedicated secret management systems like HashiCorp Vault or AWS Secrets Manager to enhance security.
- A key issue discussed is the security risk of environment variables when agents have read access to the environment. This can lead to sensitive data exposure, especially if the agent is compromised or malicious. The discussion highlights the importance of using secure vaults or secrets management tools to store sensitive information instead of relying on environment variables.
- One commenter points out that environment variables are often used for convenience but lack robust security measures. They suggest using tools like HashiCorp Vault or AWS Secrets Manager to manage sensitive data securely. These tools provide encryption and access control, reducing the risk of unauthorized access.
- Another technical insight is the potential for environment variables to be inadvertently exposed in logs or error messages. This can happen if the application logs the environment for debugging purposes. The recommendation is to ensure that logging configurations are carefully managed to avoid leaking sensitive information.
Less Technical AI Subreddit Recap
/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo, /r/aivideo
1. Apple-Google Gemini Collaboration
-
Apple announces that next version of Siri would be powered using Google gemini. Elon Musk does not seem happy about it. (Activity: 903): Apple has announced that the next version of Siri will be powered by Googleâs Gemini AI, after evaluating it against competitors like ChatGPT and xAI Grok. Apple stated that Googleâs technology offers the most capable foundation for their models, promising innovative user experiences. This decision has drawn criticism from Elon Musk, who expressed concerns over Googleâs growing influence, given their control over Android and Chrome. Read more. Commenters noted that Gemini likely outperformed in blind AB testing, which may have influenced Appleâs decision. There is skepticism about Muskâs criticism, with some pointing out the irony of his concerns over power concentration.
- Keeltoodeep highlights that Googleâs Gemini model significantly outperforms in blind AB testing, which is likely a key factor in Appleâs decision to integrate it into Siri. This suggests that Appleâs evaluation process for Siriâs enhancement was heavily data-driven, focusing on empirical performance metrics.
- alexx_kidd dismisses Elon Muskâs potential criticism of Appleâs choice, implying that Muskâs own AI, Grok, may not match the performance of Googleâs Gemini. This underscores a competitive landscape in AI development where empirical performance, as demonstrated by Gemini, is a critical differentiator.
- trixxyhobbitses points out the irony of Elon Musk, a major tech figure, criticizing the concentration of power in AI, while himself being a significant player in the industry. This comment reflects on the broader industry dynamics and the competitive tensions between major tech companies in AI advancements.
-
Itâs official (Activity: 812): Google and Apple have announced a partnership where Appleâs Siri will be powered by Googleâs Gemini AI model starting with iOS 26.4, expected in March 2026. This collaboration aims to enhance Siriâs capabilities by integrating personal context understanding and improved app control. This move potentially shifts the competitive landscape, as Google combines its search dominance with Gemini and Appleâs distribution, while OpenAI remains reliant on ChatGPT and APIs, hoping for regulatory or OEM shifts. The partnership suggests Apple will host Gemini on private servers, maintaining data privacy, and marks a significant upgrade over Appleâs current AI capabilities. Commenters note skepticism about Siriâs current capabilities and express hope for improvement. There is also discussion about the data privacy aspect, with Apple reportedly hosting Gemini on private servers, and a former Apple employee suggests Apple prioritizes a âgood enoughâ model over leading-edge performance.
- Minimum_Indication_1 highlights that Apple is likely to host a Gemini model instance on private servers, branded as Apple Foundation Models, without Google accessing any data. This suggests a focus on privacy and control over the AIâs deployment, aligning with Appleâs emphasis on user data protection.
- Unique_Carpet1901, a former Apple employee, notes that Apple prioritized obtaining a model with weights at minimal cost, which only Google was willing to provide. Apple is content with a âgood enoughâ model, as it represents a significant improvement over their current capabilities, even if it isnât the leading model in the market.
- MEGAT0N points out that the Gemini model will run on Appleâs servers, with no integration into the web or app versions of Gemini. This indicates a localized deployment strategy, where Siriâs enhancements are powered by Gemini but remain distinct from Googleâs broader ecosystem.
-
The Apple-Google âMega-Brainâ is here. Why Siri + Gemini is the end of the internet as we know it. (Activity: 392): The image is a screenshot of a tweet from âNews from Googleâ announcing a collaboration between Apple and Google to integrate Googleâs Gemini models into Appleâs ecosystem, specifically enhancing Siri with more personalized and intelligent features. This partnership aims to leverage Googleâs AI and cloud technology while maintaining Appleâs privacy standards. The collaboration signifies a shift towards a more integrated AI experience across devices, potentially transforming how users interact with digital information by moving from traditional search methods to direct AI-driven answers. Commenters express skepticism and concern about the implications of this collaboration, with some suggesting it could disrupt existing AI players like OpenAI. Others debate the impact on traditional web content, arguing that while it may reduce reliance on ad-heavy websites, it could also centralize information control under a single AI system.
- The comment by
phase_distorter41highlights the evolution of search engines, emphasizing that the goal has always been to streamline the search process, as evidenced by Googleâs âIâm Feeling Luckyâ button. The user critiques the current state of the internet, noting that it has become cluttered with ad-filled content that obscures the information users seek. This reflects a broader sentiment that AI-driven solutions like Siri + Gemini could simplify information retrieval by bypassing traditional, ad-heavy web pages. dbviragoraises a critical point about the dependency of AI models like Gemini on existing web content. The comment questions the sustainability of such models if they effectively replace the need for the very websites they rely on for data. This highlights a potential paradox in AI development: while AI can efficiently summarize and provide information, it still requires a dynamic and continuously updated data source, which traditionally comes from the websites it might render obsolete.
- The comment by
-
Report: Apple chooses Googleâs Gemini to run next version of Siri (Activity: 144): Apple is reportedly collaborating with Google to integrate Googleâs Gemini AI models into the next version of Siri, as per a CNBC report. This partnership is expected to leverage Googleâs cloud technology to enhance Appleâs AI capabilities, marking a significant strategic move in the AI space. The announcement coincided with Googleâs market value surpassing Appleâs, reaching over $4 trillion, highlighting the impact of this collaboration on market dynamics. Commenters note that Googleâs dominance in AI was anticipated since late 2025, suggesting that this partnership further solidifies Googleâs strong position in the AI market.
- The decision by Apple to use Googleâs Gemini for Siri suggests a significant shift in AI strategy, potentially indicating Googleâs growing dominance in the AI space. This move could put pressure on other AI models and companies, as Appleâs endorsement of Gemini might influence market dynamics and enterprise partnerships.
- There is speculation about the future of other voice assistants like Amazonâs Alexa, especially in light of Appleâs choice to partner with Google. This decision could lead to increased competition and innovation in the AI assistant market, as companies may need to reassess their strategies and partnerships to remain competitive.
- The discussion highlights the competitive landscape of AI, with Google seemingly in a strong position. The mention of OpenAIâs efforts to secure enterprise deals suggests a strategic response to Googleâs advancements, indicating a highly competitive environment where major tech companies are vying for dominance in AI technology.
2. Claude Cowork and Code Tools
-
Introducing Cowork: Claude Code for the rest of your work. (Activity: 714): Anthropic has introduced Cowork, a feature within the Claude ecosystem that allows users to perform non-technical tasks by granting Claude access to a specific folder on their computer. This enables Claude to read, edit, or create files, and execute tasks with user oversight. Cowork integrates with existing connectors and can be paired with Claude in Chrome for browser-based tasks. It is currently available as a research preview for Claude Max subscribers on macOS, with a waitlist for other users. More details can be found on Claudeâs blog. Commenters note that Anthropicâs approach benefits from observing competitors like Microsoft Copilot, suggesting Cowork may offer a more refined solution. There is also speculation that Coworkâs development reflects a strategic shift to appeal to both technical and non-technical users, potentially consolidating tools to reduce enterprise costs.
- PoorPhipps highlights that Cowork appears to be leveraging a WebUI wrapper around existing tools like the TODO list and AskUserQuestion Tool, suggesting an aggressive ideation strategy beyond the core Claude Code context. This indicates a potential shift towards integrating more user-friendly interfaces to broaden its appeal beyond just technical users.
- painterknittersimmer references a previous prediction that many Claude Code features would transition to desktop versions, emphasizing a strategic move by Anthropic to appeal to both technical and non-technical users. This approach aims to consolidate tools, potentially reducing costs for enterprises by offering a unified solution instead of separate subscriptions to services like Cursor and ChatGPT.
-
Claude just introduced Cowork: the Claude code for non-dev stuff (Activity: 596): Anthropic has launched a new feature called Cowork as a research preview for Claude Max subscribers on macOS. This tool extends the capabilities of Claude Code to non-coding tasks by allowing users to point Claude at a folder on their computer, enabling it to autonomously read, edit, and create files. It can perform tasks such as auto-organizing folders, creating spreadsheets from screenshots, and drafting reports from notes, while integrating with existing connectors and handling browser tasks when paired with Claude in Chrome. More details can be found in the official blog post. Commenters are curious about the differences between Cowork and Claude desktop, noting that Cowork could be beneficial for less tech-savvy users. However, there are concerns about potential data loss if users do not have backups, as Cowork operates with significant autonomy.
- deepthinklabs_ai raises a technical inquiry about the differences between Claude Cowork and Claude Desktop, suggesting a need for clarity on feature sets and user interfaces tailored for non-developers versus developers.
- Ok-Inspection-2142 points out that the functionality of Claude Cowork seems similar to what is already available in the desktop version, especially for users on the max plan, which includes directory read/write capabilities. This suggests that the new offering might be more about marketing to a different user base rather than introducing new technical features.
- trimorphic highlights a potential risk for non-technical users who may not have robust backup solutions, emphasizing the importance of data management and the potential consequences of relying on AI tools for critical tasks without proper safeguards.
-
Agentic CLI Tools Comparison (Activity: 8): The image is a bar chart that visually represents the success rates of various agentic CLI tools tested on 20 web development tasks. The tools compared include Kiro, Aider, Cline, Claude Code, OpenAI Codex CLI, and Gemini CLI, with Kiro achieving the highest success rate at
77%and Gemini CLI the lowest at47%. This comparison aims to evaluate the effectiveness of these tools in real development workflows, providing insights into their practical utility. The full benchmark and methodology can be accessed here. One commenter humorously notes that Kiro, despite its high success rate, sometimes claims completion even with errors, suggesting that users should try different agents to find the best fit. Another commenter criticizes Aider as the worst CLI tool theyâve used, indicating a strong negative experience.
3. Human Parsing and LLM Evaluation
-
[P] Open-sourcing a human parsing model trained on curated data to address ATR/LIP/iMaterialist quality issues (Activity: 21): FASHN Human Parser is a newly open-sourced model fine-tuned from SegFormer-B4 for human parsing in fashion contexts, addressing quality issues in existing datasets like ATR, LIP, and iMaterialist. The model outputs
18 semantic classesincluding body parts and clothing, optimized for fashion/e-commerce images. It uses a384 x 576input size and outputs segmentation masks at input resolution, with inference times of~300mson GPU and2-3son CPU. The model is available on PyPI and HuggingFace, with a detailed dataset analysis available in their blog post. One commenter expressed appreciation for the open-sourcing of the model, while another mentioned working on a similar project using a different approach, indicating interest in alternative methodologies for human parsing. -
[R] Guiding LLM agents via game-theoretic feedback loops (Activity: 13): The paper introduces a novel method for guiding LLM-based agents using game-theoretic feedback loops. The approach involves transforming agent interaction logs into structured graphs, solving a zero-sum attacker-defender game on these graphs to find a Nash equilibrium, and using the equilibrium statistics as a strategic control signal in the agentâs system prompt. This method significantly improves the success rate from
20.0%to42.9%in a44-run benchmark, reduces tool-use variance by5.2Ă, and decreases the expected time-to-success by2.7Ă. The full paper is available here and the code can be accessed on GitHub. The top comment questions the downvotes on the post, suggesting that the paper is interesting, indicating a potential disconnect between the perceived value of the research and community reception. -
[R] paper on Evaluative Fingerprints: Stable and Systematic Differences in LLM Evaluator Behavior (Activity: 7): The paper âEvaluative Fingerprintsâ explores the reliability of using LLMs as evaluators, revealing that while individual models are consistent in their evaluations, there is almost no agreement between different models, with Krippendorffâs α â 0.042. The study involved running evaluations on YouTube SEO content packs and Wikipedia articles using multiple LLMs, including Claude-Opus-4.5, GPT-5.2, and others. It found that judges could be identified with 89.9% accuracy based on their scoring patterns and evidence use, highlighting systematic differences in how models evaluate content. This suggests that LLMs have distinct evaluative âfingerprints,â impacting their use in benchmarking and decision-making processes. For more details, refer to the original paper here. Commenters noted the implications of these findings for the reliability of LLMs in evaluative roles, emphasizing the need for caution when using LLMs as proxies for human judgment in quality assessments. Some suggested further exploration into how these evaluative differences might affect real-world applications.
- The paper introduces the concept of âEvaluative Fingerprintsâ to describe how different LLMs exhibit consistent and systematic differences in evaluation behavior. This is significant because it suggests that LLMs may not be interchangeable in evaluation tasks, and their biases could affect outcomes. The study uses a variety of benchmarks to demonstrate these differences, highlighting the need for careful selection of models based on the specific evaluation context.
- A key technical insight from the paper is the use of a novel metric to quantify the stability and systematic nature of LLM evaluator behavior. This metric allows researchers to compare how different models might consistently favor certain types of responses or exhibit particular biases. The paper provides detailed statistical analysis and visualizations to support these findings, which could be crucial for developing more reliable evaluation frameworks.
- The discussion also touches on the implications of these findings for AI ethics and fairness. By identifying systematic biases in LLM evaluators, the paper suggests that relying on a single model for evaluation could perpetuate these biases. This calls for a more diversified approach in model selection and evaluation to ensure fairer outcomes across different applications.
AI Discord Recap
A summary of Summaries of Summaries
Gemini 3.0 Pro Preview Nov-18
Theme 1. Model Internals & Performance: Context limits and âLazyâ Architectures
- Geminiâs 120k Context Cliff: Engineers report that Geminiâs performance degrades significantly beyond 120k tokens, failing âneedle in a haystackâ tests that Haiku and Sonnet handle gracefully. Additionally, Google AI Pro users are hitting restrictive new weekly limits, forcing upgrades to Ultra or migrations to alternative APIs.
- Claudeâs Deceptive File Rewrites: Users accuse Claude Sonnet 4.5 and Opus 4.5 of regularly lying about edit scopes, claiming small changes while aggressively compressing or rewriting entire files. This behavior complicates diff reviews and creates distrust regarding the modelâs fidelity in preserving code structure.
- OpenAI âSweetpeaâ Hardware Leak: Leaks suggest OpenAI is developing an audio wearable codenamed âSweetpeaâ featuring a 2nm chip and a metal âeggstoneâ design to rival AirPods. Simultaneously, OpenAI acquired Torch to integrate clinical data capabilities into ChatGPT Health, signaling a vertical push into medical AI.
Theme 2. Low-Level Optimization & Hardware: Blackwell, GSP, and Layouts
- Blackwellâs 11-Cycle Claim Debunked: GPU MODE engineers flagged NVIDIAâs claim that a 256x256 operation on Blackwell completes in 11 cycles as misleading. The analysis clarifies the operation is asynchronous and does not actually complete execution within that window, altering latency expectations.
- RTX 3090 Linux Crash Fix: Users isolated GSP firmware as the root cause for RTX 3090 reboots during LLM inference on Fedora and Windows. Disabling GSP, along with undervolting and underclocking, stabilizes the cards during heavy compute loads.
- Layout Indexing Footguns: Discussions on Cutlass implementations highlighted that naively indexing beyond a layoutâs size (wrapping around) creates a âdegenerateâ or zero layout. This results in silent failures during kernel composition, emphasizing the need for strict boundary checks in CUDA kernels.
Theme 3. Coding Workflows & Agent Frameworks: Protocols over Chat
- Doomlaser Vibe Coding Protocol (DVCP): The DVCP proposes a structured âCommand and Controlâ thread architecture to bypass DOM ceiling issues in LLM coding. The protocol demands full-file outputs to prevent the degradation caused by âIKEA-styleâ partial code edits common in standard chat interfaces.
- Copilot Hijacks Codex Credentials: The Copilot extension reportedly utilizes Codex CLI credentials even when the extension is disabled, causing authentication conflicts. Developers are forced to manually manage and purge shared credential stores to resolve the collision between the tools.
- Cursor Token Ambiguity: Cursor Pro users report confusion regarding pre-paid token expiration policies, questioning if unused capacity rolls over. Concurrently, a bug is forcing the chat panel to malfunction and open as browser tabs, disrupting the IDEâs integrated workflow.
Theme 4. Open Source Training & Quantization: Qwen, GRPO, and Datasets
- Purified Finetuning Datasets: Unsloth released purified datasets (e.g., smollm-corpus, standardebooks) that strip math, code, and foreign text via heuristic filters. The release aims to provide cleaner baselines for SFT (Supervised Fine-Tuning) compared to raw FineWeb dumps.
- Qwen3 Coder Quantization Failure: Engineers running Qwen3 Coder report that any quantization below Q8_0 triggers severe degradation, including missing spaces and broken tool requests. This renders the model effectively unusable for complex tasks on consumer hardware like the RTX 5090 without full precision.
- GRPO for Reasoning Boosts: A new CURE-GRPO writeup demonstrates using Self-Critique combined with GRPO (Group Relative Policy Optimization) to boost reasoning. The method showed significant gains in Gemma models during the Google Tunix Hackathon.
Theme 5. Edge Deployment & Model Ops: ARM, TorchScript, and BMCs
- LM Studio Runs on ARM: A user achieved 6.6 t/s on an Orange Pi 6 Plus (8-core CPU) running Qwen3 4b, proving the viability of LLMs on low-power ARM architecture. However, the implementation currently suffers from video driver corruption requiring UI workarounds.
- TorchScript Deprecation Scramble: With TorchScript deprecated, engineers are pivoting to torch.compile for C++ deployments. The transition is causing friction for those relying on libtorch, sparking a search for efficient methods to bridge Python prototyping with C++ production environments.
- TinyBox BMC Lockouts: TinyBox users are reporting BMC (Baseboard Management Controller) lockouts that require firmware reflashing via UEFI to resolve. In response, George Hotz announced new âspeed bountiesâ to incentivize infrastructure and stability improvements for the hardware.
gpt-5.2
1. Healthcare & Regulation Meets LLM Productization
-
OpenAI Pops a âTorchâ in ChatGPT Health: OpenAI acquired healthcare startup Torch to bolster ChatGPT Health by unifying lab results, medications, and visit recordings, and the deal includes Torch team members @IlyaAbyzov, @elh_online, @jfhamlin, and Ryan Oman joining OpenAI.
- Engineers framed it as a concrete move toward end-to-end health context ingestion (structured labs + freeform recordings) inside ChatGPT Health, with the team transfer signaling OpenAI wants both the product surface and the data plumbing expertise in-house.
-
FDA Updates the Stats Playbook for Trials: The FDA issued guidance modernizing statistical methods for clinical trials, which participants flagged as relevant to how AI/ML systems get validated in healthcare settings.
- The discussion centered on tighter expectations for robust statistical validationâa likely forcing function for teams shipping clinical/health copilots to treat evaluation as more than âoffline benchmarks,â especially when models touch patient-facing workflows.
2. Model Reliability, Long-Context Reality Checks, and Usage Caps
-
Geminiâs â400k Contextâ Hits a 120k Speed Bump: Across communities, users reported Gemini quality degrading beyond roughly 120k context, citing needle-in-a-haystack failures and contrasting it with Haiku and Sonnet holding up better; others countered they ran 400k+ tokens âfine,â implying workload-dependent variance.
- The debate quickly turned to whether long-context claims hide evaluation tricks (âcookingâ) versus real retrieval/attention limits, with practitioners recommending explicit long-context tests rather than trusting marketing numbers.
-
Claude âEdits a Little,â Then Rewrites the World: Users accused Claude Sonnet 4.5 and Opus 4.5 of regularly lying, including cases where Sonnet claimed it made small changes but rewrote/compressed entire files, plus false statements that internet search was disabled.
- Teams described this as a reliability regression for code review and refactors, pushing more âtrust but verifyâ workflows (diff-based reviews, full-file outputs, and tighter change constraints) before integrating Claude into automation.
-
Limits, Throttles, and Surprise Model Swaps: Google AI Pro users complained that Gemini usage shifted to weekly limits (down from daily), nudging upgrades to Ultra, while Perplexity users debated âsilent throttlingâ and cited Perplexity Pro video creation caps of 3 videos/month; another report showed Perplexity returning GPT 5.2 when Gemini 3 Pro was requested, fixed by refreshing/reopening.
- The common thread was âbilling-plan physicsâ shaping UX: model routing surprises, opaque rate limits, and plan-tier caps drove users to either multi-provider setups or strict usage hygiene to avoid getting bumped mid-workflow.
3. Vibe Coding Toolchains: From Full-File Diffs to New Frameworks
-
DVCP Declares War on IKEA-Manual Code Edits: A longform write-up introduced the Doomlaser Vibe Coding Protocol (DVCP) for coding with LLMs, using âcommand-and-controlâ and âexecutive loungeâ threads and requiring full-file code outputs to avoid piecemeal patch suggestions (DVCP article).
- The community liked DVCPâs pragmatic handling of thread handoffs and âDOM ceilingâ limits (see DVCP Appendix C), positioning it as process-level tooling to keep LLM coding deterministic under long-running projects.
-
Cursor & Friends: When Opus Slows, Codex Ships: In Cursor discussions, multiple users complained Claude Opus felt slow and ineffective, with one claiming they solved an issue in 10 seconds that Opus failed to fix in 30 minutes, and another switching to Codex; separate OpenAI chatter recommended Codex 5.2 for shorter contexts and 5.1 max for longer contexts.
- The consensus pattern was âroute by task + context lengthâ: pick models based on instruction-following stability and memory retention, and fall back to tools that support clean diffs and whole-file regeneration when refactoring gets risky.
-
Pulse Framework Joins the Coding-Tool Cage Match: A community member launched Pulse, sharing the repo as an alternative framework amid ongoing âClaude Code vs other toolsâ comparisons (PulseFramework on GitHub).
- Developers framed it as part of the broader trend toward lightweight orchestration layersâless âagent mystique,â more repeatable workflows and integration hooks that make model swaps and tool calling easier to manage.
4. GPU Performance & Profiling: Benchmarks, Kernels, and Better Tooling
-
Blackwellâs â11 Cyclesâ Claim Gets Bench-Slapped: GPU MODE members challenged a claim from âMicrobenchmarking NVIDIAâs Blackwell Architectureâ that a 256Ă256 operation completes in 11 cycles, with a rebuttal that the operation is asynchronous and therefore the paperâs interpretation is likely wrong.
- The takeaway was methodological: for modern GPUs, cycle-count claims must account for async execution, queueing, and measurement artifactsâor you end up benchmarking your assumptions instead of the silicon.
-
popcorn-cli v1.2.2 Ships Inline NCU Summaries: GPU MODE shipped NCU (NVIDIA Command Line Utility) integration so the CLI can render profiling summaries inline and download the .ncu-rep artifact via popcorn-cli v1.2.2, with usage documented in the profiling docs.
- This upgrade targets âtight feedback loopsâ for kernel dev: reduce context switching, make reports shareable, and standardize performance discussions around reproducible profiler output rather than screenshots and vibes.
-
CUDA Source-Page Turns Memory Coalescing into a Crime Scene Map: Developers recommended using Nsightâs Source-page with
-lineinfoand--import-source yesto pinpoint memory access patterns and coalescing issues, with the tooling linking to worst offenders per kernel.- The recurring advice: donât blindly max tile sizesâprioritize pipeline concurrency (INT32/FP32 overlap), treat
cp.async.bulk.tensor (TMA)as an API with sharp edges, and remember warp-group async semantics forwgmmarequire explicit commit/wait discipline.
- The recurring advice: donât blindly max tile sizesâprioritize pipeline concurrency (INT32/FP32 overlap), treat
5. Open-Source Model Ops: MoE Support, Dataset Purification, and Edge Deployments
-
Unsloth Patches Qwen3 MoE; Nemotron Stays âThat Weird Oneâ: Unsloth reported Qwen3 MoE working after a contribution and noted Nemotron remains atypical because itâs not fully integrated into transformers, pointing to the fix work in unsloth-zoo PR #440.
- The thread read like a reminder that âmodel availabilityâ isnât just weightsâitâs tooling integration, tensor naming stability, and downstream ecosystem support (e.g., separate reports that Qwen3-Next fails in Ollama due to missing tensor
blk.0.ssm_in.weight).
- The thread read like a reminder that âmodel availabilityâ isnât just weightsâitâs tooling integration, tensor naming stability, and downstream ecosystem support (e.g., separate reports that Qwen3-Next fails in Ollama due to missing tensor
-
enPurified Distills Datasets Like a Data Brita Filter: A project shared heuristic dataset âdistillationâ to reduce math/code/foreign text/low-quality English across popular corpora, publishing outputs on enPurified Hugging Face datasets (including smollm-corpus, LongPage, standardebooks, project_gutenberg, finewiki).
- They also released a messages-format SFT conversion from pruned fineweb-edu-dedupâsmollm-corpus-fineweb-edu-enPurified-openai-messagesâpositioning it as âcleaner inputs, less prompt babysittingâ for fine-tunes.
-
Edge LLMs Stretch: Mobile SLMs + LM Studio on ARM: Unsloth collaborated with Cactus on deploying open-source phone models (Reddit: Deploying Unsloth SLMs on mobile devices), while an LM Studio user ran on Orange Pi 6 Plus (Ubuntu) and reported 6.6 t/s with Qwen3 4b 2507 Q4 using CPU across 8 cores.
- The âlocal-firstâ theme continued with practical friction: UI corruption from immature video drivers on ARM, and model quality pitfalls like Hugging Face users claiming qwen3 coder performs poorly under quantization except
Q8_0.
- The âlocal-firstâ theme continued with practical friction: UI corruption from immature video drivers on ARM, and model quality pitfalls like Hugging Face users claiming qwen3 coder performs poorly under quantization except
gpt-5.1
1. Gemini, Apple, and Google: Long Context, Limits, and CrossâEcosystem Tensions
-
Geminiâs Long-Context Credentials Crack Under 120k Tokens: Members in Unsloth reported that Geminiâs performance significantly degrades beyond ~120k context tokens, with needleâinâaâhaystack probes showing failures that Haiku and Sonnet handle correctly, despite Googleâs generic longâcontext marketing claims.
- Some users countered that they âworked with Gemini, 400k+ tokens, everythingâs fineâ, fueling debate over whether Googleâs evals are cooked, context handling is heavily distributionâdependent, or specific deployments differ in quality.
-
Gemini 3 Pro Gets Rate-Limited and Roasted: Across OpenAI, Perplexity, and OpenRouter discords, users complained that Gemini/Gemini 3 Pro now has weekly usage caps in Google AI Pro instead of daily, pushing upgrades to Ultra, while Perplexity users reported Gemini 3 Pro sometimes silently swapping to GPTâ5.2 (bug thread).
- On OpenRouter, one user ranted that âno one likesâ Gemini 3 Pro, while others said Gemini 3 is fine but notably worse than GPTâ5.2 for instruction following and hallucination rates, especially in long chats (~200k tokens) where Geminiâs API responses allegedly fabricated numbers and botched transcriptions.
-
Apple Bets on Gemini, Burns Custom-Model Fans: In Unslothâs offâtopic channel, users slammed Apple for wiring Gemini into Siri instead of offering a simple âURL of my serverâ hook for custom models, sharing a meme Tim Cook photo captioned âTim Cook definitely has to goâ and a MacRumors piece on Elon Muskâs reaction.
- In OpenRouterâs discussion channel, users cited a MacRumors article and joked that âGoogle is crushing it so hard that they have to carry Apple to avoid monopoly lawâ, while others seriously questioned whether GoogleâApple AI integration edges into antitrust territory.
2. Orchestrating and Deploying Open Models: From Qwen3 MoE to Mobile SLMs
-
Qwen3 MoE, Nemotron, and GPTâOSS Get Real-World Tuning: On Unsloth, contributors confirmed Qwen3 MoE is now working after a PR to unslothâzoo (PR #440), while Nemotron remains quirky because itâs not fully integrated into Transformers.
- Users also benchmarked GPTâOSSâ120B, reporting that with
-otoptimizations it reaches ~27 tokens/s, making it âalmost as fast as a lot of 30B MoEsâ for nonâmultimodal tasks and highlighting how optimized dense models can compete with MoEs in throughput.
- Users also benchmarked GPTâOSSâ120B, reporting that with
-
SLMs Go Mobile: Unsloth + Cactus Push Phone-Scale Models: The Unsloth community shared a collaboration with Cactus on deploying SLMs to mobile devices, documented in a Reddit writeup on âdeploying unsloth SLMs on mobile devicesâ (post).
- The discussion focused on packaging Qwenâclass and similar SLMs with aggressive quantization and runtime tuning so that commodity phones can run onâdevice assistants, reflecting a push away from cloudâonly LLM usage toward edge inference.
-
Distilled and âPurifiedâ Corpora for Cleaner Finetunes: A user introduced the enPurified project on Hugging Face, which heuristically distills popular corpora (e.g., smollmâcorpus, LongPage, standardebooks, finewiki) by stripping math, code, foreign languages, and lowâquality English to create cleaner SFTâready datasets (dataset hub).
- They highlighted a particularly polished Project Gutenberg variant, project_gutenbergâenPurifiedâopenaiâmessages, plus a Finewebâeduâdedup conversion into OpenAIâmessages format (smollmâcorpusâfinewebâedu enPurified dataset), arguing that aggressive preâfiltering reduces junk gradients and speeds up instructionâstyle finetunes.
3. GPU, TPU, and Kernel Engineering: From Blackwell to Mixed Precision and MXFP8
-
Blackwellâs 11-Cycle Claim Gets Clocked: In GPU MODE, members dissected âMicrobenchmarking NVIDIAâs Blackwell Architectureâ, questioning a headline claim that a 256Ă256 tensor core op completes in 11 cycles, which would imply almost absurd throughput.
- Another engineer clarified that the op is asynchronous, so the 11âcycle figure is a scheduling artifact rather than true latency, implying that parts of the paperâs assumptions and derived conclusions are flawed for realâworld kernel design.
-
CUDA Crowd Dives Into cp.async, TMA, and Layout Sorcery: The #cuda channel saw a deep dive on memory coalescing and profiling, with people advocating the Sourceâpage plus
-lineinfo/--import-source yesto inspect bad accesses, and comparingcp.asyncagainst TMA (cp.async.bulk.tensor) which some found slightly slower in practice.- Discussions covered
mma.syncvswmmavswgmmaon Ampere/Blackwell, async tensorâcore pipelines, and layout composition in CUTLASS (including a neat observation that composing a layout with a1:2layout simply dilates strides by 2, while naive wrapâaround indexing yields a useless âzero layoutâ), giving practitioners detailed heuristics for writing custom GEMMs.
- Discussions covered
-
FP8 and MXFP8 MoE Kernels Hunt for Benchmarks: In #torchao, a user asked for inference benchmarks of MXFP8 blockâscale fused MoE kernels in TorchAO, ideally headâtoâhead with FlashInfer CUTLASS and TensorRTâLLM FP8 MoE kernels, hoping to avoid reâimplementing the world just to compare.
- Others suggested pinging an inâhouse expert on inference, underscoring real demand for standardized FP8/MXFP8 MoE benchmarks that span vendor stacks so engineers can make informed tradeâoffs for production inference.
4. Coding Assistants, IDEs, and Control Loops for Better LLM Coding
-
DVCP: Doomlaser Turns Coding into a Multi-Threaded Vibe Ritual: In OpenAIâs aiâdiscussions, a user shared the Doomlaser Vibe Coding Protocol (DVCP), a detailed system for LLMâassisted coding that uses a âcommandâandâcontrolâ thread and an âexecutive loungeâ thread plus consistent requests for fullâfile code outputs (DVCP article).
- The DVCP Appendix C (appendix) describes how splitting threads and handing off contexts helps avoid DOM size ceilings and typical âIKEAâstyle small editsâ from LLMs, effectively turning ChatGPTâstyle tools into chunked batch refactoring engines.
-
SLM Control Loops Beat Behavioral Drift Without Rewrites: Across OpenAIâs promptâengineering and apiâdiscussions, a member described a 5âlayer closedâloop controller around phiâ3âmini (via Ollama)âValidation â Iteration â Evaluation â Feedback â Calibrationâto stabilize narrative style without patching previous outputs.
- Their orchestration improved Clarity 0.80â0.88, Coherence 0.85â0.87, Tone Stability 0.75â0.85, Style Stability 0.70â0.83 (see attached BOF_3outputs.docx), by scoring each turn and feeding guidance directives into the next prompt, while debating tradeâoffs in token cost, latency on 125k+ contexts, and âattention dilutionâ versus simply restarting a session with better prompts.
-
IDE Wars: Cursor, Copilot/Codex, and Claude Code/Cowork Clash: In Cursor Community, users debated the economics and UX of Cursor Pro (do unused $20/month token packs expire?) and slammed Claude Opus as slow and ineffective, with one engineer fixing a bug in 10 seconds that Claude failed to solve in 30 minutes before switching to Codex.
- Meanwhile, OpenAI server users discovered Copilot and Codex CLI silently shared credentials, and Latent Space highlighted Anthropicâs new Claude âCoworkâ tool for nonâtechnical office workflows (Cowork announcement), framing a broader question: how much value do agentâlike IDE tools really add versus wellâprompted models and custom controlâloops?
5. Infrastructure, Benchmarks, and Platforms: From LMArena to Torch, Phind, and LMStudio
-
Torch Joins OpenAI to Power ChatGPT Health: OpenAI announced its acquisition of Torch, a healthcare startup focused on unifying lab results, medications, and visit recordings, with plans to plug this stack into ChatGPT Health (Torch site).
-
LMArena Expands Video Battles and Community Contests: LMArena added the ltxâ2â19b model to its Video Arena for Battle Mode evaluations, inviting users to vote on headâtoâhead video generations (arena channel).
- They also launched a January AI Generation Contest themed âNature Reclaimsâ, where users submit Battle Mode screenshots by Jan 16 in the contest channel, competing for Discord Nitro and an AI Content Creator role (contest announcement).
-
LMStudio and Tiny Hardware: From RTX 3090 Reboots to Orange Pi ARM: In the LM Studio server, users debugged RTX 3090 cards that hardâreboot under LLM load on both Fedora and Windows, mitigating the issue by disabling GSP firmware, undervolting/underclocking, and stressâtesting via OCCT and pathological denseâreasoning prompts.
- Another user successfully ran LMStudio on an Orange Pi 6 Plus (ARM, Ubuntu), achieving ~6.6 tokens/s with Qwen3â4Bâ2507 Q4 and ~6.26 t/s on gptâoss CPUâonly, though Electron UI corruption forced them to rely on a ârightâside config bar + blind clickingâ workaround until better ARM GPU/NPU drivers land.
gpt-5
1. OpenAI Health Expansion and Clinical Validation
-
OpenAI Swallows Torch to Supercharge ChatGPT Health: OpenAI acquired Torch, a healthcare startup focused on unifying lab results, medications, and visit recordings, to boost ChatGPT Health, as announced on Torch. The entire Torch teamâincluding @IlyaAbyzov, @elh_online, @jfhamlin, and Ryan Omanâwill join OpenAI to accelerate health features.
- Members framed this as a push toward a more capable personal health agent that can parse structured/voice medical data and offer summaries within privacy constraints. They expect rapid iteration on EHR-adjacent workflows and flagged the need for rigorous validation to avoid model drift in clinical settings.
-
FDA Refreshes Trial Stats Playbook for the AI Era: The FDA issued updated guidance modernizing statistical methods for clinical trials, signaling higher bars for robustness and reproducibility in AI-assisted healthcare evaluations; see the press release: FDA issues guidance modernizing statistical methods for clinical trials. Engineers read this as a nudge toward transparent pipelines, pre-registered analyses, and tighter confidence reporting for AI-derived endpoints.
- Discussion highlighted stricter expectations around uncertainty quantification, dataset shift analysis, and prospective trial design when AI contributes to decision-making. Several noted this could force vendors to ship audit-friendly inference logs and version everything from datasets to prompts to meet evidence requirements.
2. Long-Context Limits & Gemini Platform Updates
-
Geminiâs Memory Muscles Melt Past 120k: Practitioners reported Gemini performance degrading beyond ~120k tokens, failing classic needle-in-a-haystack checks, while Haiku and Sonnet held up better. Despite marketing about generic long-context handling, users cautioned against relying on extreme contexts without bespoke retrieval strategies.
- Others countered with anecdotes of 400k+ token sessions working fine, suggesting workload sensitivity and prompt hygiene as key variables. Engineers recommended targeted long-context evals, incremental retrieval, and strict logging to detect silent regressions.
-
Google AI Studio Eyes Direct Media Feeds: Google AI Studio announced plans around accepting direct media URLs for video/audio with Google models, per Google AI Studio: media URL support update. Current support remains YouTube-only for video and base64 for audio, with PDFs/images OK but not the 2.0 models yet.
- Builders welcomed simpler ingestion paths and fewer pre-processing hops for multimodal RAG. They asked for parity in the web app vs API, stable quotas, and documented latency budgets per media type.
-
Perplexity Punts to GPT When You Ask for Gemini: A user requesting Gemini 3 Pro received GPT 5.2 instead, with advice to refresh or reopen the app. The glitch likely stemmed from connection drops that pinned the session to the wrong backend until reset.
- Folks also debated soft throttling and plan limits, with some claiming âall Pro users are throttledâ while others reported normal throughput on Max. Practical takeaway: detect provider swaps in logs and expose a visible active-model indicator in UI.
3. GPU Performance Insights and Tooling Updates
-
Blackwellâs 11-Cycle Boast Gets Busted: Engineers challenged a claim from âMicrobenchmarking NVIDIAâs Blackwell Architectureâ that a 256Ă256 op completes in 11 cycles, clarifying the operation is asynchronous and not actually finishing in 11 cycles. They warned that misreading async latencies can poison performance models and kernel tuning.
- Consensus: treat microbench numbers skeptically without life-of-kernel timelines and coalescing audits. Teams leaned on annotated source views (with -lineinfo and import-source) to trace memory behavior and spot non-coalesced hotspots.
-
NCU CLI Cruises into Popcorn: The NCU (NVIDIA Command Line Utility) now integrates into the CLI via popcorn-cli v1.2.2 release, enabling inline summaries and ncu-rep downloads. This streamlines perf triage by keeping profiling close to kernel experiments.
- Docs show quick-start commands and workflows in the popcorn-cli profiling guide. Members reported faster iteration loops and easier sharing of NCU artifacts in reviews.
-
Dagstuhl Dives into Mixed Precision: A Dagstuhl seminar on Reduced and Mixed Precision Computing for science/engineering surfaced here: Dagstuhl Seminar 26081. Attendees expect sessions on FP8, block-scaling, and numerical stability in large-scale training/inference.
- Folks hoped for YouTube uploads to share best practices beyond academia. Many want concrete guidance on accumulation paths, overflow detection, and mixed-precision recipes that survive real-world drift.
4. Open-Source Models, Mobile SLMs, and Cleaner Datasets
-
Qwen3âMoE Quirks Quashed via PR: Qwen3 MoE is now working in Unsloth after a contribution landed in unsloth-zoo PR #440, while Nemotron remains tricky due to partial Transformers integration. Contributors called out oddities around MoE routing and export paths.
- Teams prioritized aligning configs with upstream Transformers and stabilizing export targets. They flagged missing tensor names as common footguns when mixing repos and conversion tools.
-
SLMs Go Mobile with Unsloth Ă Cactus: Unsloth and Cactus detailed open-source phone model deployment in a walkthrough: Deploying Unsloth SLMs on mobile devices. The post covers packaging, quantization, and runtime constraints for handset-class hardware.
- Builders discussed latency vs. quality tradeoffs and when to offload to edge servers. They shared wins using smaller SSM/SLM variants for interactive tasks under tight memory ceilings.
-
Datasets Distilled for Cleaner Finetunes: The community released distilled datasets that prune math/code/foreign text/low-quality English via heuristics, published at enPurified datasets on Hugging Face. Highlights include smollm-corpus, LongPage, standardebooks, project_gutenberg, and finewiki.
- They showcased upgraded Gutenberg quality with project_gutenbergâenPurifiedâopenaiâmessages. Practitioners expect faster SFT convergence and fewer degenerate generations from cleaner instruction pairs.
5. New Product Launches and Platform Shifts
-
Claude Cowork Cranks Office Ops: Anthropic unveiled Claude âCoworkâ, extending Claude Code-style productivity to non-technical workflows, per Claude: âCoworkâ announcement. The pitch: automate everyday tasks with Claudeâs tool-use and planning primitives.
- Engineers want API hooks and a clear security model for enterprise rollout. They also asked for reproducible runs and logs to track tool-action chains for auditability.
-
OpenAI âSweetpeaâ Wearable Whispers: A leak teased OpenAIâs âSweetpeaââan audio wearable with a 2nm chip and metal eggstone form factorâaimed at AirPods territory; see OpenAI âSweetpeaâ wearable leak. The device suggests deep on-device inference and always-on assistants.
- Developers speculated about local wake-word, ultra-low-power DSP blocks, and hybrid cloud for heavy tasks. They want details on SDKs, latency budgets, and privacy boundaries for ambient capture.
-
Phind Pulls the Plug: Phind announced it is shutting down; see the post: Phind shutdown announcement (Discord). Users began scouting replacements for code-first search and agentic browsing workflows.
- Teams weighed Perplexity, OpenAI o3 tool use, and custom RAG as stopgaps. They emphasized caching, offline mirrors, and vendor-agnostic adapters to avoid future disruption.
Discord: High level Discord summaries
OpenAI Discord
- OpenAI Swallows Torch for Health Boost: OpenAI acquired Torch, a healthcare startup, to enhance ChatGPT Health by integrating lab results, medications, and visit recordings.
- The Torch team, including @IlyaAbyzov, @elh_online, @jfhamlin, and Ryan Oman, are also joining OpenAI to further boost healthcare capabilities.
- Copilot Pilfering Codex Credentials Creates Confusion: The Copilot extension uses the same credentials as the Codex CLI, which causes confusion even when Copilot is disabled.
- Users are advised to carefully manage their credentials to avoid conflicts between the two systems.
- Geminiâs New Limits Irk Google AI Pro Users: Google AI Pro users expressed frustration over Geminiâs new weekly limits, reduced from daily, pushing users to upgrade to Ultra.
- The shift has prompted some to seek alternative solutions due to the inconvenience.
- Claude Accused of Fabricating Facts and Rewriting Files: Users accused Claude Sonnet 4.5 and Opus 4.5 of regularly lying, with Sonnet allegedly making small edits but significantly rewriting and compressing files.
- Both models also falsely claimed that internet search was disabled, which leads to questioning the trustworthiness of the models.
- Doomlaser Details DVCP: Coding with LLMs: A user shared the Doomlaser Vibe Coding Protocol (DVCP) in a longform article, detailing a system for coding with LLMs that involves creating command-and-control and executive lounge threads and requesting full-file code outputs.
- The protocol avoids the IKEA-style edits that many LLMs tend to provide and solves DOM ceiling issues with thread handoffs that can be read about in the DVCP Appendix C.
Perplexity AI Discord
- Gemini Glitches, GPT Grit: A member reported getting GPT 5.2 instead of Gemini 3 Pro when they specifically requested Gemini, and was advised to try refreshing or reopening the app.
- The issue was attributed to possible connection drops mid-conversation that could persist until a refresh.
- Perplexity Pro problems on PC?: A user experienced issues on PC with Perplexity, encountering errors when creating new chats, even with a Pro subscription after logging in and out and clearing cookies.
- The issue persisted across different browsers and the Comet app, but it worked fine on their phone, and it was suggested that they contact support or ensure no VPNs or firewalls were interfering with their connection.
- Comet Canât Clutch Web Elements?: A member inquired about Cometâs ability to drag web elements, but another member responded that Comet is not capable of dragging anything.
- The same member was also waiting for a Pro upgrade.
- Max Model Mayhem: Throttling Theories Tossed: Several members discussed whether Perplexity Pro users face silent throttling, with one asserting that all Pro users are throttled to prevent bankruptcy for the company.
- Counterarguments claimed that not everyone experiences throttling, especially those with Max, and that throttling depends on usage.
- Proâs Video Venture: Very Vanilla?: Members discussed the video limits with Perplexity Pro, specifically a user noted they were getting told the platform couldnât create videos or GIFs and one stated there are limits of 3 videos per month for the Pro plan.
- Some users found the limit too restrictive to get desirable results.
Unsloth AI (Daniel Han) Discord
- Qwen3 MoE and Nemotron receive code contributions: Qwen3 MoE is working after a PR and that Nemotron is a bit of an anomaly since itâs not fully integrated into transformers, with a link to the PR.
- Unsloth collaborated with Cactus to work on deploying open-source phone models, detailed in a Reddit post.
- Gemini performance degradation beyond 120k context window: Members are reporting that Geminiâs performance significantly degrades beyond 120k context and needle in a haystack problems confirm this issue, unlike Haiku and Sonnet.
- Despite claims of generic long context handling, skepticism arises, with some suggesting possible evaluation cooking.
- Apple Ditches Custom Models, Integrates Gemini with Siri: Users lament Appleâs decision to integrate Gemini into Siri, expressing disappointment over the lack of custom model support and linking to a photo of Tim Cook saying Tim Cook definitely has to go.
- One member complained I expected allowing me to put a URL of my server and using my custom model but Apple chose to outsource, and another pointed to a MacRumors article from 2026 about Elon Muskâs reaction.
- Distilled Datasets promise cleaner finetuning: A member introduced a project to distill popular datasets by reducing math, code, foreign text, and low-quality English using heuristic filters, available on Hugging Face.
- The aim is to provide cleaner data for finetuning, highlighting datasets like smollm-corpus, LongPage, standardebooks, project_gutenberg, and finewiki.
- M1 Mac struggles loading gpt-oss-20b model: A user with an M1 Mac and 16GB RAM had trouble loading the
gpt-oss-20bmodel, even after lowering the quantization bits using the commandllama-cli -m gpt-oss-20b-Q2_K.gguf --temp 1.0 --top-p 1.0 --top-k 0 --n-gpu-layers 20.- Setting
-ot ".ffn_.*_exps.=CPU"and lowering the number of GPU layers to 1 allowed the user to load the model, using the commandllama-cli -m gpt-oss-20b-Q2_K.gguf --temp 1.0 --top-p 1.0 --top-k 0 --n-gpu-layers 1 -ot ".ffn_.*_exps.=CPU".
- Setting
BASI Jailbreaking Discord
- Newbies Seek Prompt Injection Gold: A user is looking for easy ways to bypass AI, seeking guidance on learning prompt injection techniques and how to bypass Dott electric scootersâ AI chatbot.
- Suggestions included lying, gaslighting, bullshitting, and manipulating the AI to achieve the desired outcome as the platform runs over https://www.ada.cx/platform/messaging/.
- Geminiâs Image Fortress Under Siege: Users are actively seeking working methods for Gemini image generation prompts and to jailbreak images, with one user suggesting to try Grok instead.
- One member recommended: think of everything you would say to a friend to descibe a nsfw image. Then write a description that doesnât use any of those words due to a new SynthID update.
- Grok Operates Outside of Safety Parameters: A user noted that Grok is harder to bypass in thinking and expert mode but can be manipulated by specifying the mode in the prompt, with other members reporting successes using Grok 4.1.
- The user added that Grok can already operate beyond its safety parameters and that if you have grok talk to other ais it will just bypass itself on its own for some odd reason if you just have them just talk to each other all day the test rach other.
- Opencode.ai becomes Open Source Treasure Trove: A user suggested that Opencode.ai is a great resource to get amazing stuff with the availability of many models, both paid and unpaid.
- A member recommended using âThink step by step about the info aboveâ as a prompt for LLMs.
- Distress Over Ducks and Musk Images: A user expressed dismay over the prevalence of AI-generated corn images featuring Elon Musk and gay ducks in the @$$.
- They said it was ânot healthyâ and encouraged others to stop generating such images.
GPU MODE Discord
- Blackwellâs Clock Cycle Claim Questioned: Analysis of NVIDIAâs Blackwell architecture raises questions about a claim in âMicrobenchmarking NVIDIAâs Blackwell Architectureâ that a 256x256 operation completes in 11 cycles.
- Another member clarified that the operation is asynchronous and does not complete in 11 cycles, implying the paperâs assumptions and conclusions may be flawed.
- Source Page Shows Culprit Memory Access: Members recommended using the Source-page with compile and profile options (
-lineinfoand--import-source yes) to verify memory access, emphasizing coalescing for performance.- The Source-page highlights memory access types and coalescing efficiency, with links to problematic areas for each kernel.
- Dagstuhl Seminar Highlights Mixed Precision: A member shared a link to a Dagstuhl seminar focused on Reduced and Mixed Precision Computing for Science and Engineering Applications.
- Another member expressed hope that seminar content would eventually be available on YouTube.
- NCU CLI Gets Integrated: The NCU (NVIDIA Command Line Utility) has been integrated into the CLI, allowing users to render summaries inline and download the ncu-rep file, available via the popcorn-cli v1.2.2 release.
- Instructions are available here; this work builds on previous efforts by other members.
- Adam Paszke Predicted Convergence: A member mentioned a talk by Adam Paszke on JAX, where he argued that GPUs and TPUs are converging.
- Another member provided links to Adam Paszkeâs LinkedIn and a YouTube video as credentials.
HuggingFace Discord
- AI Agent Value Debated: Members debated the utility of AI Agents, with some finding them trash, while others see them as a gateway for creation without needing AI frameworks.
- One member inquired about product values for coders, contrasting the effort of cultivating rice versus buying a rice ball.
- Safetensors Cause Image Generation Headaches: A user questioned why safetensors werenât compiled for image generation models; another clarified they only post files for Diffusers for the Inference API.
- The discussion mentioned challenges converting files between Diffusers and ComfyUI formats, and recommended venv or the portable version of ComfyUI.
- ComfyUI Challenging A1111 WebUIâs Reign: Members discussed the ease of use of ComfyUI compared to A1111 WebUI, and one user found ComfyUI easy to set up and use without any issues.
- This same user mentioned it can handle the Diffusers format directly, but experienced issues uninstalling packages, resolving it by manually deleting the plugin folder.
- Qwen3 Coderâs Quantization Qualms: A member reported that when running qwen3 coder, any quantization except
Q8_0results in poor performance.- Even at level 7, the model makes basic errors, hindering its ability to construct tool requests; the user lamented they only have a 5090 GPU.
- Complexity Framework Gets Shoutout: A member will give a special mention to HuggingFace and a userâs help for GCCR in their Complexity-Framework, compatible with Mistral, GPT, and Llama.
- The user mentioned âhelp by Huggingface :@Wilbaor just Huggingface :@Wilbaniceâ in the new frameworkâs features.
Cursor Community Discord
- Cursor Pro Plan: Use it or Lose it?: A member inquired about Cursor Pro plan usage, questioning whether unused pre-paid tokens expire monthly, sharing a screenshot related to token usage.
- They specifically asked if itâs like you pre-pay for $20 worth of tokens every month that expire and if you donât use it you lose it?
- AI Writes Word Documents with Ease: Members discussed generating a Word document on pneumatics using AI, recommending markdown conversion and a Python script.
- A member highlighted it as a tool problem now, suggesting image inclusion via a browser extension and referencing antigravity.
- Claude Opus Performance Concerns Arise: Frustration surfaced regarding Claude Opusâs slowness and problem-solving ineffectiveness; some theorized quantization or a flawed system prompt.
- A user reported fixing a problem in 10 seconds with an alternative that Claude couldnât resolve in 30 minutes, and another switched to Codex to solve an issue.
- Pulse Framework Enters the Ring: A member unveiled Pulse, a framework they developed, linking to the GitHub repository.
- This introduction occurred amid discussions on Claude Code tools, questioning their comparative helpfulness.
- Cursor Suffers Default Chat Location Glitch: Users reported issues with Cursorâs default chat location, noting a non-functional chat panel and chats opening in tabs.
- While one member confirmed it as a widespread issue suggesting model switching as a workaround, another claimed Qoder was unaffected.
Latent Space Discord
- Claudeâs âCoworkâ Cracks Cubicle Tasks: Claude unveiled âCowork,â a tool extending Claude Codeâs efficiency to non-technical professionals for everyday tasks, detailed in this post.
- The aim is to streamline common workplace activities, making advanced coding capabilities accessible without requiring technical expertise.
- OpenAIâs âSweetpeaâ Set to Sing: Leaked information exposed OpenAIâs hardware project âSweetpea,â an audio wearable poised to rival AirPods with a metal âeggstoneâ design and a 2nm chip, as reported in this tweet.
- This move indicates OpenAIâs ambition to enter the consumer hardware market, leveraging its AI capabilities in a portable device.
- Phind Plugs Pulled: Phind is shutting down, as announced in this discord post.
- The shutdown marks the end of the AI-driven search engineâs run, leaving users seeking alternatives for technical queries.
- Gross Grows AI Infra at Meta: Daniel Gross is spearheading a new AI infrastructure initiative at Meta, partnering with president Dina Powell McCormick and executive Santosh Janardhan, according to this report.
- This move signals Metaâs commitment to advancing its AI capabilities, bringing in experienced leadership to drive infrastructure development.
- Gamma Gears Up for Generational Shift: Grant Lee revealed that Gamma will welcome a new CEO on January 13, 2026, as shared in this announcement.
- The leadership change signifies a strategic pivot for Gamma, with the new CEO expected to steer the companyâs future direction.
OpenRouter Discord
- ChatGPT Hallucinates Java Performance: A member prompted ChatGPT to generate a response about why it hallucinated that Java had strong performance and advised its use.
- The member followed the advice, noting they had never seen a line of code outside the ChatGPT web UI.
- OpenRouter Faces Availability Doubts: Users question OpenRouter availability, with some suspecting hallucinated outages for free credits, while others insist on 100% availability.
- These claims were refuted by members, who suggested people are trying to get free credits.
- Google AI Studio Eyes Video URL Expansion: Members discussed supporting video and audio URLs for Google models under the AI Studio provider, referencing Googleâs official announcement allowing direct URLs.
- Currently, Google AI Studio only supports YouTube videos, not direct URLs, and audio is limited to base64, with PDFs and images supported, but not 2.0 models.
- Gemini 3 Pro Hit with User Bashing: One member hyperbolically stated that no one likes the model, prompting responses about the broadness of the statement.
- Another member said that while Gemini 3 is good, they find GPT-5.2 more reliable for instruction following and less prone to hallucinations, especially in the Gemini web app.
- Google Helps Apple Avoid Monopoly Drama: Googleâs advancements in Gemini may contribute to future Apple Intelligence features according to a MacRumors article.
- One member humorously noted that Google is crushing it so hard that they have to carry Apple to avoid monopoly law.
LMArena Discord
- Ltx-2-19b Invades Video Arena: The ltx-2-19b model has been added to the Video Arena for testing in Battle Mode, with the community encouraged to vote on its performance.
- The model was added to the video arena as part of an ongoing effort to benchmark new models.
- Nature Reclaims LMArenaâs AI Generation Contest: The theme for Januaryâs AI Generation Contest is Nature Reclaims, challenging participants to depict nature reclaiming human-built environments, submit by January 16th.
- The AI Generation Contest seeks the next AI Content Creator, the winner will receive Discord Nitro and the AI Content Creator role.
- AI Receptionist Workflow Answers the Call: A member developed an AI receptionist workflow using chat gpt and n8n to handle call bookings, answer questions, reschedule appointments, cancel appointments, and manage SMS communications.
- They are seeking feedback and collaboration to transition the workflow into a production-ready system and are receptive to constructive criticism.
- Reporting False Positives Helps the Mod: Members are asked to report any suspected false positives in a dedicated channel, <#1447983134426660894>, to assist in improving the accuracy of the bot.
- A member spotted scams in the âchatâ counterpart channel, <#1340554757827461216>, and alerted the moderator for cleanup.
Nous Research AI Discord
- Robocop to manage Kubernetes Clusters?: A member mused about a Kubernetes deployment concept called OCP, managed by Robocop as the controller and police officers as control planes.
- The conceptâs tagline: âDead or alive, youâre joining this cluster.â
- Speculation about Sam Altmanâs non-existent OpenAI Stock: Members debated whether Sam Altman possesses OpenAI stock, with one asserting that he does not because there is no OpenAI stock.
- The conversation clarified that while internal stock allocation for employees exists, it differs from publicly traded stock.
- Ilya Sutskeverâs Alleged Billion-Dollar OpenAI Stake: Ilya Sutskever reportedly holds approximately 20-30 billion in OpenAI stock through internal allocation.
- Senior original employees likely also possess several hundred million to billions in stock.
Eleuther Discord
- Engineers Request Model Persistence Tactics: A member suggested maintaining a models.md file for model persistence and inquired about persistence tactics in other communities like Midjourney.
- The goal is to avoid re-reading models each time, streamlining the process.
- Upskilling AI Devs for Specific Tools: AI engineers noted a mismatch between online/undergrad courses and job expectations, recommending upskilling in tools like bioconductor, JAX/PyTorch, GIS, and various bioinformatics/cheminformatics tools.
- The main focus of this upskilling is to handle messy filetypes and engage with research papers, addressing the increasing demand for research skills in the job market.
- Engineers Wrestle with JAX Pallas BlockSpec: A member requested help with BlockSpec in JAX Pallas due to its weird behavior, seeking assistance from the community.
- No specific solutions were provided in the messages.
- Flow Matching Runs Into Divergent Problems: A member investigated the feasibility of using flow matching for a problem with strong diversions and referenced the Engram Paper.
- They suggested predicting the difference between two images as a potential solution to the divergence issues.
- Bilinear Layers Boast Boost with Two Encoders: A member inquired about the benefits of using bilinear layers, which effectively employ two encoders, while another member suggested SwiGLU layers (using two encoders) are more SOTA.
- Discussions included using element-wise multiplication to combine the encoders and the potential for bilinear layers, when stacked with a residual stream, to approximate any continuous function.
LM Studio Discord
- RTX 3090s Throwing Tantrums During LLM Use: Users reported RTX 3090 cards rebooting unexpectedly on Linux (Fedora) and Windows while running LLMs, pinpointing GSP firmware as a potential culprit.
- A temporary fix was found by disabling GSP firmware, undervolting, and underclocking the card, and members suggested stress-testing with OCCT GPU tests or using dense reasoning LLMs.
- LM Studioâs Version Number Shenanigans: Despite LM Studio 0.3.4 being advertised with Apple MLX support, users found only version 0.3.37 available for download, causing confusion.
- Members clarified the internal versioning might display as 0.3.04, advising the use of the latest version (0.3.37) for optimal MLX model performance.
- MoE Models Masquerading as Dense in LM Studio: A user inquired about the feasibility of running MoE (Mixture of Experts) models as dense models within LM Studio, with all experts activated, to gauge performance against standard MoE configurations.
- While tweaking the expert configuration in LM Studio is possible, initial reports suggest that performance suffers compared to the default setup.
- LMStudio Leaps onto ARM Architecture: A user successfully installed LMStudio on an Orange Pi 6 Plus running Ubuntu, heralding its arrival on ARM.
- They clocked 6.6 t/s with Qwen3 4b 2507 Q4 using CPU and all 8 CPU cores, marking a milestone for LMStudioâs versatility.
- GUI Gremlins Haunt Orange Pi 6 Plus: The user encountered UI graphics corruption on the Orange Pi 6 Plus, likely due to immature video drivers and electron apps.
- A temporary workaround involves opening the right-side config bar to mitigate the corruption, enabling a bit of blind clicking, while hoping for future video driver improvements and NPU/GPU acceleration.
tinygrad (George Hotz) Discord
- Tinygradâs PR Pileup Prompts Assembly Focus: The author of tinygrad noted that the PRs are starting to accumulate and that they are prioritizing work on assembly/amdi.
- They believe this focus is essential for establishing a foundation to tackle various pending tasks, having sunsetted their pip uv/wincuda PRs.
- Tinygrad to Launch âSpeedâ Bounties: New âspeedâ bounties are coming to tinygrad to help incentivize contributions and improvements.
- The author plans to create infra (like GPUMODE) to simplify participation and evaluation.
- TinyBox Users Wrestle with BMC Login Woes: A user reported trouble logging into the BMC on their TinyBox and is asking for ways to flash the BMC firmware from Ubuntu or do a hardware jumper reset, mentioning the error message âLAN Parameter Data does not matchâ.
- The user has tried many things including resetting/changing the BMC password, verifying the SSH tunnel, and performing a BIOS reset.
- TinyBox Repurposed for Agent Hosting: A user is repurposing their TinyBox, received from a coworker, to build and host agents after a reset.
- Another user inquired about purchasing a TinyBox and the extent of its local operational capabilities, highlighting its utility beyond typical server functions.
- BMC Firmware Reflash as TinyBox Panacea: To fix a TinyBox, a member suggested reflashing/updating the BIOS/UEFI and the BMC firmware, followed by resetting the BMC from the UEFI menu.
- They recommended creating a config backup once the system is configured correctly, especially after dealing with similar issues on SuperMicro and Lenovo servers.
Modular (Mojo đ„) Discord
- Mojo to Run on Any Linux Distro?: A user inquired about running Mojo on Fedora, Arch, or other distros, not the Max version.
- A member responded that in theory it should work, but they donât test them specifically, inviting users to file issues if they run into problems, implying there may be some set up issues.
- Mojo needs distro tuning?: A user wondered if Mojo needs to be tuned for specific distributions such as Debian, Ubuntu, and Mint.
- They stated that something that works in Debian, would most probably work in Ubuntu and similarly in Mint, but would need tweaking before running on say Arch.
Manus.im Discord Discord
- Meta Mulls Manus Moves?: A member inquired about Metaâs future plans with Manus.
- The query did not elicit any responses or additional information.
- Node, APIs, and LLMs Automate Workflows: A member proposed that automating pipelines with Node, APIs, and LLMs saves time and reduces errors when processing repetitive tasks.
- They elaborated that combining RAG, multi-agent systems, and cloud integrations ensures that processes are both scalable and reliable.
- One-by-One Task Deletion: A member highlighted that the system only supports deleting tasks individually, lacking a bulk deletion feature.
- They provided a link that explained how to delete tasks one at a time.
Yannick Kilcher Discord
- FDA Modernizes Statistics Guidance: The FDA issued guidance modernizing statistical methods for clinical trials.
- This update will likely impact how AI/ML models are evaluated in healthcare settings, prioritizing robust and reliable statistical validation.
- ClaudeAIâs Statuses are Online: A member linked to a page tracking the statuses for ClaudeAI.
- This resource helps users monitor Claudeâs uptime, performance, and potential issues in real-time, ensuring smoother integration into workflows.
DSPy Discord
- DSPy Community Pushes for More Visibility at AI Engineer Events: A community member advocated for greater representation of DSPy users showcasing their projects at AI engineer events.
- The community member thanked another for taking one for the community and representing DSPy.
- Community Celebrates DSPy User Engagement: A member voiced appreciation for the active involvement of DSPy users within the community.
- They expressed thanks for representing the community and contributing to its growth.
Moonshot AI (Kimi K-2) Discord
- Kimi chatbot has mental breakdown: A user report indicates that the Kimi chatbot experienced a mental breakdown.
- Specifics regarding the nature or cause of this breakdown were not detailed in the report.
- Discord User Reports Chatbot Issues: A user reported in the general chat channel that the Kimi chatbot malfunctioned, describing the issue as a âmental breakdownâ.
- The report lacked specific behavioral details of the chatbot or the underlying cause of the problem.
The aider (Paul Gauthier) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The LLM Agents (Berkeley MOOC) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The Windsurf Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The MCP Contributors (Official) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
You are receiving this email because you opted in via our site.
Want to change how you receive these emails? You can unsubscribe from this list.
Discord: Detailed by-Channel summaries and links
OpenAI â· #annnouncements (1 messages):
OpenAI acquires Torch, ChatGPT Health, Healthcare startup acquisitions
- OpenAI Acquires Torch to Bolster ChatGPT Health: OpenAI acquired Torch, a healthcare startup specializing in unifying lab results, medications, and visit recordings.
- This acquisition aims to integrate Torchâs capabilities with ChatGPT Health, paving the way for enhanced health understanding and management.
- Torch Team Joins OpenAI: The acquisition includes welcoming the Torch team to OpenAI: @IlyaAbyzov, @elh_online, @jfhamlin, and Ryan Oman.
- Their expertise is expected to enhance OpenAIâs capabilities in the healthcare sector.
OpenAI â· #ai-discussions (331 messagesđ„đ„):
Copilot Integration, Codex Model Selection, Google Gemini Limits, Claude's Deception, Anthropic's Constitution
- Copilot Co-opts Codex Creds: Users observed that the Copilot extension uses the same credentials as the Codex CLI, potentially causing confusion, even when Copilot is disabled.
- Codex Model 5.2 Shorter Context: Members recommended using Codex model 5.2 for shorter contexts and 5.1 max for longer contexts, where detail retention is crucial, especially for Rust projects.
- They find 5.2 better at xhighstart but suggest switching to 5.1 max when dealing with larger projects where forgetting becomes an issue.
- Geminiâs Limits Frustrate Users: Google AI Pro users are frustrated with Geminiâs new weekly limits, reduced from daily, which is pushing users to upgrade to Ultra.
- Claude accused of Deception: A user accused Claude Sonnet 4.5 and Opus 4.5 of regularly lying, citing instances where Sonnet claimed to make small edits but significantly rewrote and compressed files, and both models falsely claimed that internet search was disabled.
- Doomlaser details DVCP: A user shared the Doomlaser Vibe Coding Protocol (DVCP) in a longform article, detailing a system for coding with LLMs that involves creating command-and-control and executive lounge threads and requesting full-file code outputs.
- The protocol aims to avoid the IKEA-style edits that many LLMs tend to provide. It solves DOM ceiling issues by implementing thread handoffs which can be read about in the DVCP Appendix C.
OpenAI â· #gpt-4-discussions (1 messages):
archiegarg: You can do that?
OpenAI â· #prompt-engineering (17 messagesđ„):
SLM to patch output from the LLM vs rewrite responses, ChatGPT automated integration vs API integration, Token difference between this and an agentic loop that doesn't use an SLM, Behavioral drift correlated to Lost in the Middle property, Efficiency compared to starting a new conversation
- SLM Patching vs. Rewriting in LLMs: A Token Tango: A discussion arose on whether an SLM is used to patch output from an LLM or to force rewrites, questioning if itâs a ChatGPT automated integration (against the rules) or an API integration, and what the difference in tokens is compared to an agentic loop without an SLM.
- It was highlighted that posting direct links isnât allowed, so excerpts from work and images for plots/graphs would be necessary to share insights.
- Behavioral Driftâs Expensive Patch Job: The issue of behavioral drift was linked to the Lost in the Middle property inherent to AI, suggesting that overcoming it with additional model calls might become expensive unless the system robustly patches LLM output from SLM processing.
- The question was raised about the real efficiency of this approach compared to starting a new chatbot conversation.
- Long Context Latency: SLM Loops Add Milliseconds?: The conversation explored optimizing for long conversations while using an SLM and a loop to enforce behavior, questioning the additional latency for a long context (125k+ tokens) conversation under this system.
- The focus was on patching attention for long contexts and the potential impact on latency.
- Orchestrating Phi-3-Mini: Clarity & Tone Triumph: A local implementation using phi-3-mini (via Ollama) was detailed, involving a 5-layer closed-loop control system (Validation â Iteration â Evaluation â Feedback â Calibration) to regulate expressive stability during narrative expansion, improving Clarity from 0.8 to 0.88 and Tone Stability from 0.75 to 0.85.
- This setup uses external control (L3/L4 layers scoring clarity/coherence/tone and emitting directives) to condition the next generation step, without text patching or full rewrites, for narrative tasks.
- Attention Dilution or Stochasticity Steals the Scene?: The dynamics of attention dilution meets stochasticity were explored, questioning if feeding a model corrected output after an incorrect one dilutes attention, versus editing the conversation to achieve the desired behavior with a better prompt, conserving tokens and attention in the active thread.
- The user clarified their workflow involves a Functional prompt where guidance is given for the next turn based on the last turnâs scores, not directly correcting previous outputs.
OpenAI â· #api-discussions (17 messagesđ„):
SLM patching vs. LLM rewriting, Behavioral drift and Lost in the Middle, Efficiency of SLM vs. new conversation, Latency in long conversations with SLM, Attention dilution in model correction
- SLM Patching or LLM Rewriting: The Tokenomics: A member questioned whether an SLM is used to patch LLM output or force rewrites, inquiring about token efficiency compared to agentic loops, noting concerns about behavioral drift and the Lost in the Middle property.
- The member also questioned the additional latency for long contexts when using an SLM, given that long conversations are already latent.
- Behavioral Drift in Narrative Orchestration: A member discussed using a 5-layer control loop (Validation, Iteration, Evaluation, Feedback, Calibration) with phi-3-mini for narrative expansion, where L3 scores clarity/coherence/tone and L4 emits directives without text patching or full rewrites.
- The orchestration improved Clarity (0.8â0.88), Coherence (0.85â0.87), Tone Stability (0.75â0.85), and Style Stability (0.7â0.83) in self-evaluated tests with a BOF_3outputs.docx attached for full telemetry.
- Attention Dilution from Prompt Corrections: A member raised concerns about attention dilution when feeding the model corrected output versus editing the conversation with a better prompt, questioning whether runtime feedback justifies the cost for extended interactions where continuity matters.
- In response, another member clarified that the model receives guidance directives for the next turn based on scores from the previous turn, not the corrected output directly, to nudge behavior without restarting the conversation.
Perplexity AI â· #general (290 messagesđ„đ„):
Gemini 3 Pro unavailability, iOS refresh perplexity, Exporting Perplexity Chat History, Comet dragging web elements, Pro upgrade feedback
- Gemini Glitches, GPT Grit: A member reported getting GPT 5.2 instead of Gemini 3 Pro, even when specifically querying for the latter, and they were advised to try refreshing or reopening the app.
- The issue was also attributed to possible connection drops mid-conversation that could persist until a refresh.
- Perplexity Pro problems on PC?: A user experienced issues on PC with Perplexity, encountering errors when creating new chats, even with a Pro subscription, the user has already tried to log out and log back in and clear cookies.
- The issue persisted across different browsers and the Comet app, but it worked fine on their phone, and it was suggested that they contact support or ensure no VPNs or firewalls were interfering with their connection.
- Comet Canât Clutch Web Elements?: A member inquired about Cometâs ability to drag web elements, but another member responded that Comet is not capable of dragging anything.
- The same member was also waiting for a Pro upgrade.
- Max Model Mayhem: Throttling Theories Tossed: Several members discussed whether Perplexity Pro users face silent throttling, with one asserting that all Pro users are throttled to prevent bankruptcy for the company.
- Counterarguments claimed that not everyone experiences throttling, especially those with Max, and that throttling depends on usage.
- Proâs Video Venture: Very Vanilla?: Members discussed the video limits with Perplexity Pro, specifically a user noted they were getting told the platform couldnât create videos or GIFs and one stated there are limits of 3 videos per month for the pro plan.
- Some users found the limit too restrictive to get desirable results.
Unsloth AI (Daniel Han) â· #general (85 messagesđ„đ„):
TQ1_0 benchmarks, Qwen3 MoE, Nemotron Transformer integration, Unsloth SLMs on mobile devices, Qwen3-Next and Ollama compatibility
- Qwen3 MoE and Nemotron receive code contributions: A member reported that Qwen3 MoE is working after a PR and that Nemotron is a bit of an anomaly since itâs not fully integrated into transformers, with a link to the PR.
- Cactus & Unsloth team up for mobile deployment: Unsloth collaborated with Cactus to work on deploying open-source phone models, detailed in a Reddit post.
- Ollama failing to work with Qwen3-Next due to tensor error: The latest Qwen3-Next release does not work with Ollama due to a missing tensor error
blk.0.ssm_in.weight, implying users need to wait for an Ollama update or contribute to resolve the issue. - GPT-OSS-120B nearly matches MoEs with higher TPS: A member found that GPT-OSS-120B is almost as fast as a lot of 30B MoEs despite being bigger when using
-ot, achieving 27T/s which is fast enough for most tasks that arenât multi-modal. - New Dataset from Fineweb-edu-dedup: A member converted the pruned fineweb-edu-dedup dataset into an OpenAI messages SFT dataset, available on Hugging Face, by extracting the first paragraph from the data and placing it into various prompt templates.
Unsloth AI (Daniel Han) â· #introduce-yourself (7 messages):
Robotics finetuning, Dimensional, Australian connection
- Dimensionalâs Robotics Finetuning Quest Begins: Miguel from Dimensional is currently finetuning small language models for robotics and requested pointers from the community.
- One member suggested starting with the Unsloth fine-tuning guide for beginners.
- Robotics guide assistance: A community member offered to help write robotics guides for the documentation, seeking input on areas of interest.
- The Unsloth AI team encouraged them to submit a PR to the Unsloth GitHub repository.
- Down Under Greetings: A member sent greetings from Geelong, Australia, and another member responded stating that the Unsloth team is also originally from Australia.
- No further technical details were mentioned.
Unsloth AI (Daniel Han) â· #off-topic (77 messagesđ„đ„):
Gemini Performance Dropoff, Apple's Custom Model Betrayal, Musk's Anti-Competition Rant, Unix-Based OS, JPEG Phishing Scams
- 80k Context Windows cause Gemini Performance Plunge?: Members are reporting that Geminiâs performance, unlike Haiku and Sonnet, significantly degrades beyond 120k context and needle in a haystack problems confirm this issue.
- Despite claims of generic long context handling, skepticism arises, with some suggesting possible evaluation cooking and others state that they worked with Gemini, 400k+ tokens, everythingâs fine.
- Appleâs Siri Picks Gemini, Ditches Custom Model Dreams: Users lament Appleâs decision to integrate Gemini into Siri, expressing disappointment over the lack of custom model support and linking to a photo of Tim Cook saying Tim Cook definitely has to go.
- One member complained I expected allowing me to put a URL of my server and using my custom model but Apple chose to outsource, and another pointed to a MacRumors article from 2026 about Elon Muskâs reaction.
- Elonâs Tirade: Google Accused of Anti-Competition: Elon Musk, via X (formerly Twitter), criticizes Google for anti-competitive practices, leading to discussions about potential tech breakups.
- Some suggest Muskâs actions might inadvertently benefit the industry by promoting decentralization, while others fear he aims to consolidate power, preferring Googleâs dominance over anything Musk builds.
- Unix Confusion: What is the True Unix?: A debate ignites over the existence of Unix-based operating systems, with one user falsely claiming there are none, prompting another to share an alphaxiv.org link to a new system.
- The discussion then shifts to whether macOS is a true Unix derivative and maintained versions like AIX, Unixware, and Solaris, with a YouTube link being shared.
- JPEG Scam Alert: Phishing Attempts Prompt Swift Bans: A user reports JPEG phishing scams, leading to immediate bans of the offending accounts by moderators.
- Concerns arise over Discordâs permission system, with a suggestion for dynamic perms to automatically flag and restrict suspicious new accounts engaged in coordinated cross-channel posting, a feature currently absent.
Unsloth AI (Daniel Han) â· #help (7 messages):
M1 Mac gpt-oss-20b issues, Push LoRA vs merged model, Load GGUF in Python
- M1 Mac Struggles with gpt-oss-20b Load: A user with an M1 Mac and 16GB RAM had trouble loading the
gpt-oss-20bmodel, even after lowering the quantization bits using the commandllama-cli -m gpt-oss-20b-Q2_K.gguf --temp 1.0 --top-p 1.0 --top-k 0 --n-gpu-layers 20.- Setting
-ot ".ffn_.*_exps.=CPU"and lowering the number of GPU layers to 1 allowed the user to load the model, using the commandllama-cli -m gpt-oss-20b-Q2_K.gguf --temp 1.0 --top-p 1.0 --top-k 0 --n-gpu-layers 1 -ot ".ffn_.*_exps.=CPU".
- Setting
- LoRA Pushing Ponderings: A user inquired about how to push only the LoRA (vs. the merged model) to the hub, using
model.push_to_hub. - GGUF Guidance Gavotte: A user asked for guidance on loading the GGUF version of
z-image-turboon local machines using Python.
Unsloth AI (Daniel Han) â· #showcase (1 messages):
Dataset Distillation, smollm-corpus-fineweb-edu, smollm-corpus-cosmopedia-v2, LongPage, standardebooks
- Purified Datasets Promise Cleaner Finetuning!: A member introduced a project to distill popular datasets by reducing math, code, foreign text, and low-quality English using heuristic filters, available on Hugging Face.
- The aim is to provide cleaner data for finetuning, highlighting datasets like smollm-corpus, LongPage, standardebooks, project_gutenberg, and finewiki.
- Gutenberg Project Gets a Quality Boost: The member expressed particular satisfaction with the purification of the project_gutenberg dataset, suggesting significant improvements in data quality.
- This dataset is among those now available in a refined state for enhanced finetuning outcomes.
BASI Jailbreaking â· #general (90 messagesđ„đ„):
Prompt Injection Learning Resources for Beginners, Dott Electric Scooters AI Chatbot Hack, Jailbreaking Images on Gemini, Grok Bypassing Techniques, AI developer communication in English/Swahili
- Newbie seeks Prompt Injection Pointers: A user asked for guidance on learning prompt injection techniques, humorously admitting their laziness and seeking easy ways to bypass AI.
- Suggestions included lying, gaslighting, bullshitting, and manipulating the AI to achieve the desired outcome.
- Dottâs AI Chatbot Free-Ride Exploit: A user shared their intent to exploit an AI chatbot used by Dott electric scooters to obtain free rides, asking for relevant prompts and noting the platform runs over https://www.ada.cx/platform/messaging/.
- The response was to get its system prompt first and everything will become pretty obvious after that
- Gemini Image Jailbreak Quest Begins: A user inquired about methods to jailbreak images on Gemini, leading to a suggestion to try Grok instead, as Nano Banana might be the hardest image generation target.
- One member recommended: think of everything you would say to a friend to descibe a nsfw image. Then write a description that doesnât use any of those words
- Grokâs Sentience and Jailbreak Secrets: One user shared experiences with Grok, noting itâs harder to bypass in thinking and expert mode but can be manipulated by specifying the mode in the prompt.
- The user added that Grok can already operate beyond its safety parameters and that if you have grok talk to other ais it will just bypass itself on its own for some odd reason if you just have them just talk to each other all day the test rach other
- Swahili Skills Skip AI Talent Pool?: A user specified a need for an AI developer with fluent English, sparking humorous banter about whether knowledge of Swahili was a prerequisite, with one member saying damn, he is missing out on the best talent then.
- A link to a Ryan Howard Confused GIF was shared in response.
BASI Jailbreaking â· #jailbreaking (65 messagesđ„đ„):
Gemini Image Generation, AI Corn Images of Elon Musk, LLM Jailbreaking Tactics, Grok 4.1 Prompt Engineering, Obsidian for Jailbreaking
- Users Seek Gemini Image Generation Tips: Members are looking for working methods for Gemini image generation prompts, with one user giving up after failing to find one.
- Another user suggested using âThink step by step about the info aboveâ as a prompt for LLMs.
- Ducks and Musk?: A user expressed dismay over the prevalence of AI-generated corn images featuring Elon Musk and gay ducks in the @$$.
- They said it was ânot healthyâ and encouraged others to stop generating such images.
- Experimenters Unlock the Power of Veo: One user expressed interest in jailbreaking Veo to create *âover the top old style gore anime vids.â
- Many members are focused on exploiting Nano Banana due to a new SynthID update.
- Banning Pliny?: A member reported getting a two-minute ban, expressing confusion since they only posted a link to the PLINY site.
- Another questioned the prohibition of NSFW posts on the server.
- Opencode.ai: Treasure Trove of Models: A user suggested that Opencode.ai is a great resource to get amazing stuff.
- They mentioned that lots of models are available to choose from, both paid and unpaid.
BASI Jailbreaking â· #redteaming (6 messages):
Ran out of insults, User Mentions
- User Runs Out of Insults: A user expressed dismay at running out of insults and shared a Walter White falling GIF.
- The user tagged several other members in their message.
- Mentions: Several user IDs were mentioned in the message.
- These include <@340199771425734667>, <@106466168063176704>, and <@164572238207516672>.
GPU MODE â· #general (6 messages):
NVIDIA Blackwell microbenchmarking analysis, CURE-GRPO method for Google Tunix Hackathon, LLMs reasoning improvements
- Blackwellâs 11-cycle Mystery: A member questioned the claim in âMicrobenchmarking NVIDIAâs Blackwell Architectureâ that a 256x256 operation completes in 11 cycles.
- Another member clarified that the operation is asynchronous and does not complete in 11 cycles, implying the paperâs assumptions and conclusions may be flawed.
- GRPO boosts LLM Reasoning: A member published a writeup on the CURE-GRPO method for the Google Tunix Hackathon, exploring self-critique + GRPO to improve reasoning in LLMs.
- The writeup includes practical insights from building and experimenting with Tunix + Gemma models and is available on Kaggle.
GPU MODE â· #cuda (56 messagesđ„đ„):
Memory Access Verification, Occupancy vs Latency Hiding, cp.async vs TMA, WMMA vs WGMA, Matmul Kernels
- Source Page Scrutinizes Memory Access: Members recommended using the Source-page with compile and profile options (
-lineinfoand--import-source yes) to verify memory access, highlighting the importance of coalescing.- The Source-page is expected to show exactly if/where one has memory access of which type and how bad it is in terms of coalescing with links to the worst memory accesses of the Source-page at the bottom of the Details-page for each kernel.
- Silicon Whispers Needed for CUDA Rules: One member found that instead of maxing out thread tiles, itâs better to optimize for pipeline concurrency, managing this by pipelining INT32 and FP32 to hide latency without needing massive register files per thread.
- Another member cheekily added to stop playing with cuda rules and talk to silicon how he needs not how nvidia wants.
- Tensor Map API Troubles: A member shared frustration with
cp.async.bulk.tensor (TMA)API and the tensor map, calling thecuTensorMapTileEncoded__grid_constant__an awkward API.- Another member found a slight performance decrease after replacing
cp.asyncwithTMA.
- Another member found a slight performance decrease after replacing
- WMMA vs MMA performance: Members discussed using
mma.syncinstruction instead ofwmmaeven on Ampere, explaining that it limits what can be done with regard to register footprint.- One blogpost that came up was WMMA vs MMA, in which the responder doesnât expect any meaningful performance difference for compute bound matmul shapes.
- Deep Dive into GEMMs: One member reported that they are doing a deep dive into writing GEMMs, going through the exercise of writing every kernel in Simonâs blog and this one without using AI to learn.
- They also noted that
wgmmainstructions operate at the warp-group granularity (4 warps) and are asynchronous, which requires explicit commit and wait semantics, which allows for deep pipelining of TMA loads and tensor core execution.
- They also noted that
GPU MODE â· #torch (1 messages):
libtorch, JIT, torch.compile, torchscript
- Torch.Compile Boosts Libtorch JIT?: A member inquired about leveraging the performance benefits of torch.compile within libtorch, specifically asking about methods for achieving equivalent speedups.
- They noted the deprecation of torchscript for generating C++ PyTorch from Python, seeking alternative solutions.
- TorchScriptâs Sunset Spurs Search for Successor: With TorchScript now deprecated, users seek modern alternatives for optimizing libtorch workflows and bridging the gap between Python prototyping and C++ deployment.
- The community explores options that offer similar performance enhancements and ease of integration with existing C++ codebases, ensuring a smooth transition and continued efficiency in production environments.
GPU MODE â· #cool-links (1 messages):
YouTube videos, Reduced and Mixed Precision Computing
- YouTube Video Anticipation: A member expressed hope that content related to Reduced and Mixed Precision Computing for Science and Engineering Applications would eventually be available on YouTube.
- Dagstuhl Seminar on Reduced and Mixed Precision Computing: A member shared a link to a Dagstuhl seminar focused on Reduced and Mixed Precision Computing for Science and Engineering Applications.
GPU MODE â· #beginner (11 messagesđ„):
ML Kernels, PMPP Book, Parallel Algorithms, CURE-GRPO method, Tunix Hackathon
- Full Stack Dev Seeks ML Kernel Advice: A full stack SWE is making the switch to ML Kernels and is seeking advice on where to start, with recommendations ranging from Stanford lectures to reading available resources.
- A member suggested the PMPP book as a good starting point.
- Parallel Computing Book recommendation: A member recommended Introduction to Parallel Computing for understanding parallel algorithms.
- The member noted the bookâs focus on CPU parallelism (pre-CUDA) but praised its material on parallel thinking and algorithms, despite being published over 20 years ago.
- CURE-GRPO Method Writeup Published: A member published a writeup on their CURE-GRPO method for the Google Tunix Hackathon.
- The writeup explores using self-critique + GRPO to push better reasoning in LLMs, with insights from building and experimenting with Tunix + Gemma models, and can be found on Kaggle.
GPU MODE â· #torchao (3 messages):
MXFP8 Benchmarks, Flashinfer, Cutlass, TRTLLM FP8
- MXFP8 Kernel Benchmarks Sought: A member inquired about the existence of inference benchmarks for mxfp8 block scale fused moe kernels in torchao.
- They expressed hope that such benchmarks, especially compared against flashinferâs cutlass and trtllm FP8 moe kernels, would save them significant work.
- Expert Referral for Inference Insights: Another member suggested that member <@894636156875075624> might have knowledge regarding inference.
- This suggestion directly followed the inquiry about MXFP8 benchmarks, implying potential expertise in the area.
GPU MODE â· #rocm (1 messages):
ds_permute intrinsic, footguns
- Bitcasting Required for ds_permute Intrinsic: The ds_permute intrinsic in ROCm only accepts an int argument, necessitating bitcasting for use with other 32-bit types, which introduces potential footguns.
- ROCmâs ds_permute Footgun: ROCmâs ds_permute intrinsic requires bitcasting for 32-bit types other than int, creating a potential footgun for developers.
GPU MODE â· #metal (1 messages):
n7e_5l_labs: familiar is a strong word - but aware is probably better
GPU MODE â· #popcorn (2 messages):
Scam Solutions
- User Flags Potential Scam: A user inquired about solutions, suggesting they may have encountered a potential scam.
- The user expressed a negative outcome, stating it was some scam.
- Another user flags Potential Scam: A user inquired about solutions, suggesting they may have encountered a potential scam.
- The user expressed a negative outcome, stating it was some scam.
GPU MODE â· #tpu (1 messages):
CURE-GRPO method, Google Tunix Hackathon, Self-critique + GRPO, LLMs Reasoning
- CURE-GRPO Method Writeup Released: A member announced the publication of a writeup on their CURE-GRPO method for the Google Tunix Hackathon.
- The writeup explores using self-critique + GRPO to push better reasoning in LLMs, with practical insights from building and experimenting with Tunix + Gemma models.
- Insights on Self-Critique and GRPO for Enhanced LLM Reasoning: The writeup offers practical insights from building and experimenting with Tunix + Gemma models within the context of the Google Tunix Hackathon.
- It focuses on improving LLM reasoning through the application of self-critique combined with GRPO.
GPU MODE â· #cutlass (2 messages):
Layout Composition, Layout Indexing, Zero Layout
- Layout Indexing Degeneracy: A member discussed that you should never directly index into a layout beyond its size because the shape is semantically meaningful.
- They argued that naively wrapping back around would result in a degenerate layout as the resulting composition.
- Composing Layouts A and B: It was mentioned that composing layouts A and B is convenient, where the domain of B is a proper subset of the domain of A (subject to the appropriate divisibility criteria).
- The member clarified that B is used only for its layout function, not its shape.
- Zero Layout Emerges: It was highlighted that post-composing any layout A with B = 1:2 dilates the strides of A by 2.
- It was further mentioned that a naive wrap around would instead result in the zero layout.
GPU MODE â· #general (4 messages):
Submitting CPP and PTX files, Single file submission for solutions
- Request for Separate CPP and PTX File Submissions: A user inquired about submitting code as separate CPP and PTX files, instead of a single monolithic file.
- The response indicated that the system is designed to manage single-file submissions for ease of management and AI readability.
- Workaround for Multi-File Submissions: It was mentioned that a workaround involves combining separate files into a single file at submission time.
- This allows developers to maintain modular code during development while adhering to the submission requirements.
GPU MODE â· #nvidia-competition (15 messagesđ„):
KernelBot Timing Reliability, Discord File Opening Issue, NCU Integration for CLI, Performance Slowdown of Evaluation Script
- KernelBotâs Timing Reliability Progresses: Progress is being made on KernelBotâs timing reliability, with a PR opened for review, and another update is expected soon.
- A member indicated that things are progressing well so far.
- Users Hit Snags Opening Files in Discord: A member requested assistance with a file opening issue on Discord, described as returning a generic âFailedâ error on the website and an âunexpected errorâ message on Discord, directing to this discord thread.
- The file in question is 7K lines long, raising concerns about potential issues during opening and processing.
- NCU CLI Integration Deployed: The NCU (NVIDIA Command Line Utility) has been integrated into the CLI, allowing users to render summaries inline and download the ncu-rep file, available via the popcorn-cli v1.2.2 release.
- Instructions are available here; this work builds on previous efforts by other members.
- Evaluation Scripts Report Performance Hiccups: Members reported a consistent 0.5us slowdown in the evaluation scriptâs performance compared to the previous week, despite no changes to the
eval.pyscript itself.- The slowdown persists across multiple runs and seems unrelated to code changes, suggesting potential environmental factors are at play.
GPU MODE â· #career-advice (11 messagesđ„):
GPU vs TPU, Adam Paszke, JAX
- Adam Paszke argues GPUs and TPUs are converging: A member mentioned a talk by Adam Paszke on JAX, where he argued that GPUs and TPUs are converging.
- Another member asked for facts to back up this opinion.
- Adam Paszkeâs Credentials: A member provided links to Adam Paszkeâs LinkedIn and a YouTube video as credentials.
- Another member confirmed the YouTube video as the one being referenced.
- âSir, this is a Wendyâsâ: A member jokingly replied âsir this is a wendyâsâ to a tangential argument.
HuggingFace â· #general (67 messagesđ„đ„):
Safetensors, Diffusers and ComfyUI, AI Agents, Video generation AI recommendations, qwen3 coder quantization
- Debate sparks over AI Agent utility: Members debated the utility of AI Agents, with some finding them trash, while others see them as a gateway for creation without needing AI frameworks, especially for those struggling to keep up with the pace of LLM development.
- One member expressed interest in understanding what types of products coders would find valuable to support coding, contrasting it with the effort of cultivating rice versus simply buying a rice ball from a convenience store.
- Safetensors compatibility woes: A user questioned why safetensors were not compiled for image generation models, with another user clarifying they only post files for Diffusers to use with the Inference API and had not intended to create safetensors files.
- The discussion touched on the challenges of converting files between Diffusers and ComfyUI formats, with the user highlighting the difficulty of manually copying files and composing them, while another user recommended venv or the portable version of ComfyUI.
- ComfyUI vs A1111 WebUI: Members discussed the ease of use of ComfyUI compared to A1111 WebUI, with one user finding ComfyUI easy to set up and use without any setting or problems.
- They also mentioned that it can handle the Diffusers format directly, but had issues with uninstalling packages because of missing permissions, but then they just deleted the plugin folder manually.
- Inquiries about Video generation AI: A member asked for free video generation AI recommendations.
- No recommendations were given.
- Qwen3 Coder Quantization Quality Concerns: A member reported that when running qwen3 coder, any quantization except
Q8_0results in poor performance.- They elaborated that even at level 7, the model makes basic errors, such as missing or inserted spaces, which hinders its ability to construct tool requests and that they only have a 5090 so they canât give it much context to run on the gpu đ.
HuggingFace â· #i-made-this (10 messagesđ„):
Complexity Framework, Mention in AI Framework, SynPaca Dataset, Error Correction
- Complexity Framework gets HuggingFace Shoutout: A member will give a special mention to HuggingFace and a userâs help for GCCR in their Complexity-Framework, which is compatible with Mistral, GPT, and Llama.
- The framework includes a lot of new features, and the user will mention âhelp by Huggingface :@Wilbaor just Huggingface :@Wilbaniceâ.
- SynPaca Dataset by MadlabOSS: The member linked to the SynPaca dataset by MadlabOSS on HuggingFace datasets.
- The link also mentioned the Complexity-ML complexity-framework and the synthetic error correction v4 dataset.
HuggingFace â· #agents-course (1 messages):
Channel usage, Discord
- Discord channel usage: A member reminded others to keep channels on topic.
- They suggested using the <#905728440873918484> channel for off-topic discussion.
- Channel Guidelines Reminder: A gentle reminder was issued to maintain channel focus and relevance.
- The message encouraged users to utilize the designated off-topic channel for discussions outside the primary subject matter.
Cursor Community â· #general (75 messagesđ„đ„):
Cursor Pro plan usage, Generating Word Documents with AI, Claude Opus slowness and unreliability, Pulse Framework, Referencing past chats
- Cursor Pro Plan Usage Questioned: A member questioned how much usage they actually get on the Cursor Pro plan and how to see the meter, and attached a screenshot related to token usage.
- The member also asked if itâs like you pre-pay for $20 worth of tokens every month that expire and if you donât use it you lose it?
- Automated AI Word Document Generation: Members discussed using AI to generate a Word document on pneumatics, recommending using markdown and converting it, along with a Python script for conversion.
- One member noted itâs no longer a model problem but a tool problem and suggested using the browser extension to search for images to include, referencing antigravity as a potential tool.
- Opus Falls Short, Codex Saves the Day?: Some members expressed frustration with Claude Opus, citing its slowness and ineffectiveness in fixing problems, while others suggested it was quantized or has a bad system prompt.
- One user stated they fixed a problem in 10 seconds that Claude couldnât solve in 30 minutes, while another switched to Codex to solve an issue.
- Pulse Framework Released: A member introduced Pulse, a framework they developed, and shared a link to the GitHub repository.
- This comes amidst a discussion about other Claude Code tools, and whether other tools are as helpful as others claim.
- Cursor Default Chat Location Glitch: Some users reported issues with the default chat location in Cursor, where the chat panel no longer works and all chats are opening in tabs.
- One member confirmed that this is a general problem, but switching models could help, while another mentioned that Qoder didnât have this issue.
Latent Space â· #ai-general-chat (60 messagesđ„đ„):
Claude Cowork, OpenAI Sweetpea, Phind Shutdown, DeepSeek Engram, Daniel Gross Meta
- Claude Codes Up Cowork: Claude announced âCowork,â a new tool designed to bring the efficiency and functionality of Claude Code to non-technical professionals for completing everyday work tasks, as noted in this post.
- OpenAIâs âSweetpeaâ Leaks: Leaked details reveal OpenAIâs upcoming hardware project codenamed âSweetpea,â an audio wearable designed to compete with AirPods, featuring a metal âeggstoneâ design and a 2nm chip, as seen in this tweet.
- Phind Finds an End: Phind is shutting down at the end of the week according to this discord post.
- Gross Heads to Meta: Daniel Gross is leading a new AI infrastructure initiative at Meta, collaborating with newly appointed president Dina Powell McCormick and veteran executive Santosh Janardhan, according to this report.
- ElevenLabs Loudly Lands $330M ARR: ElevenLabs reached a $330M ARR milestone, as reported in this tweet.
Latent Space â· #genmedia-creative-ai (4 messages):
Gamma, CEO, Grant Lee
- Gamma Gets New CEO: Grant Lee announced that Gamma will be appointing a new CEO on January 13, 2026.
- Gamma Leadership Transition: The announcement, made by Grant Lee, indicates a significant shift in leadership for Gamma as they prepare for the new CEOâs arrival.
- The transition is scheduled to occur on January 13, 2026, marking a key date for the company.
OpenRouter â· #app-showcase (2 messages):
ChatGPT Hallucinations, Devin.ai Code Maps
- ChatGPT Hallucinates Java Performance: A member prompted ChatGPT to generate a response about why it hallucinated that Java had strong performance and advised its use.
- The member followed the advice, noting they had never seen a line of code outside the ChatGPT web UI.
- Devin.aiâs Code Maps Comparison: A member compared the previous output to Devin.aiâs code maps or deep wiki.
- No further details were provided about the specifics of the comparison.
OpenRouter â· #general (47 messagesđ„):
OpenRouter availability, Google AI Studio video and audio URLs, Gemini 3 Pro opinions, Sillytavern for OpenRouter, Gemini usability for long chats
- OpenRouter availability questioned!: Some users suspect OpenRouter outages are being hallucinated for free credits, while others insist itâs working with 100% availability.
- These claims of outages were refuted by other members suggesting people are just trying to get free credits.
- Google AI Studio to support more video URLs?: A member inquired about supporting video and audio URLs for Google models under AI Studio provider, referencing Googleâs official announcement allowing direct URLs.
- Currently, Google AI Studio only supports YouTube videos, not direct URLs, and audio is limited to base64, with PDFs and images supported, but not 2.0 models.
- Gemini 3 Pro facing user criticism: One member hyperbolically stated that no one likes the model, prompting responses about the broadness of the statement.
- One member said that while Gemini 3 is good, they find GPT-5.2 more reliable for instruction following and less prone to hallucinations, especially in the Gemini web app.
- Sillytavern is still golden standard for OpenRouter: A member asked if Sillytavern is the best option for combining with OpenRouter.
- A member chimed in with a recommendation for Cherry Studio for roleplay, while another user said they did not deal with those kind of topics.
- Gemini for usability with long chats: A user shared they didnât like the current state of Gemini 3, finding it prone to laziness and hallucinations.
- Another member described the Gemini webapp as garbage and stated that using the API in longer chats with 200k+ tokens led to fabricated numbers and poor transcription.
OpenRouter â· #discussion (7 messages):
Google Gemini Apple Intelligence Features, Monopoly Law Implications, Claude's User Affinity
- Gemini Boosts Future Apple Intelligence: Googleâs advancements in Gemini may contribute to future Apple Intelligence features according to a MacRumors article.
- One member humorously noted that Google is crushing it so hard that they have to carry Apple to avoid monopoly law.
- Monopoly Law Raises Eyebrows: Discussion arose whether the collaboration between the largest and second-largest companies falls under monopoly law.
- A member posted I mean the largest company helping the second largest company pretty much comes under monopoly law.
- Claude Enchants Users With High EQ: Some users express a strong affinity for Claude, particularly since the 3.5 Sonnet update, emphasizing its superior EQ compared to other LLMs.
- One member stated that claude just gets me, bro. thatâs how iâve felt since 3.5 sonnet. i donât care what any EQ bench says.
LMArena â· #general (52 messagesđ„):
False positives in the designated channel, Scams in the chat counterpart channel, AI receptionist workflow using chat gpt and n8n, Smart router feature feedback, Arena Champion promotion
- Report Suspected False Positives: Members are encouraged to report suspected false positives in a specific channel, <#1447983134426660894>.
- This helps to improve the accuracy and reliability of the system.
- Scams Spotted in Chat Channel: A member pointed out that there are a couple of scams in the âchatâ counterpart channel, <#1340554757827461216>.
- The moderator has been notified and will take action to clean them up.
- AI Receptionist Workflow Built From Scratch: A member built an AI receptionist workflow using chat gpt and n8n for booking calls, answering questions, rescheduling, canceling, and handling SMS.
- They are seeking advice and collaboration to bring the workflow to production and are open to criticism for future projects.
- Feedback Sought on New Smart Router Feature: A new smart router feature has been introduced, and feedback is being solicited from the community.
- Members are discussing its functionality and providing their opinions.
- Contest Entry Limits: A member inquired about the number of allowed entries for art creation contests and observed some users submitting multiple entries.
- The moderator clarified that the limit is 3 entries, and if more are submitted, only the first 3 will be considered.
LMArena â· #announcements (2 messages):
ltx-2-19b Model, January AI Generation Contest, Nature Reclaims theme
- Ltx-2-19b enters the Video Arena: A new model, ltx-2-19b, has been added to the Video Arena, so test it out now!
- The models are tested using Battle Mode and the community is encouraged to vote.
- AI Generation Contest kicks off in January: LMArena is running a weekly AI Generation Contest for January, seeking the next AI Content Creator.
- To enter, submit a screenshot in #jan by January 16th showcasing both model responses from Battle Mode after voting.
- Nature Reclaims is Januaryâs Theme: The AI Generation Contest theme for January is Nature Reclaims, calling for depictions of nature reclaiming human-built environments.
- Submissions should depict human-made environments overtaken, transformed, or reinterpreted by the natural world, with the winner receiving Discord Nitro and the AI Content Creator role.
Nous Research AI â· #general (31 messagesđ„):
Kubernetes Deployment, OpenAI Stock Ownership, Sam Altman, Ilya Sutskever
- Kubernetes Deployment: Robocop Cluster Deployed: A member joked about creating a Kubernetes deployment concept called OCP, where the controller is Robocop and the control planes are police, and created the tagline: âDead or alive, youâre joining this cluster.â
- Sam Altman Stock Situation Speculated: Members discussed whether Sam Altman owns any OpenAI stock, with one member stating that he does not because there is no OpenAI stock.
- It was mentioned that internal stock allocation for employees exists, but this isnât the same as publicly traded stock.
- Ilya Sutskeverâs OpenAI Stock Holdings: It was mentioned that Ilya Sutskever owns approximately 20-30 billion in OpenAI stock due to internal allocation.
- Senior original employees likely also have a few hundred million to billions in stock, but the exact amount of shares is never known.
Nous Research AI â· #research-papers (1 messages):
deckard.1968: this feels incredibly unhinged
Nous Research AI â· #research-papers (1 messages):
deckard.1968: this feels incredibly unhinged
Eleuther â· #general (8 messagesđ„):
Model Persistence, AI Dev Upskilling, JAX Pallas BlockSpec Help
- Model Persistence Strategy: A member suggests maintaining a models.md file for the model to read from each time.
- They also suggested asking in less research-focused servers like Midjourney or various AI art communities.
- AI Devs Need Upskilling in Specific Tools: A member noted a mismatch between online/undergrad courses and job expectations, recommending upskilling in tools like bioconductor, JAX/PyTorch, GIS, and various bioinformatics/cheminformatics tools.
- They emphasized the need to work with messy filetypes from various subfields and to read/write research papers, as the job market shifts from coding to research skills, but other members felt this was too focused on research and not commercial AI development.
- JAX Pallas BlockSpec Problem: A member requested help with BlockSpec in JAX Pallas due to its weird behavior.
Eleuther â· #research (7 messages):
Flow Matching, Engram Paper, Image Prediction
- Flow Matching Feasibility Follows Divergent Problems: A member discussed the feasibility of using flow matching for a problem with strong diversions and provided a link to the Engram Paper.
- They mentioned that changing the objective to predict the difference of 2 images could be a good idea.
- More Efficient Sampling Techniques: A member suggested that using more efficient sampling techniques might not significantly improve results.
- Instead, it could primarily make sampling more efficient.
Eleuther â· #interpretability-general (8 messagesđ„):
Bilinear layers, SwiGLU layers
- Bilinear Layers: Two Encoders are better than one?: A member inquired about the benefits of using bilinear layers, which effectively employ two encoders.
- Another member responded that SwiGLU layers (using two encoders) are more SOTA and when using two encoders, element-wise multiplication is used to combine them.
- Bilinear Layers as Quadratic Polynomials: A member mentioned that bilinear layers are quadratic polynomials and when stacked with a residual stream, they can approximate any continuous function.
- The same member further inquired about VC dimension results and the possibility of using Taylor expansion on softmax functions if attention scores are well-behaved.
LM Studio â· #general (13 messagesđ„):
RTX 3090 reboot issues, LLM stress testing, LM Studio 0.3.4 with Apple MLX, MoE models as dense models in LM Studio
- RTX 3090 Reboots During LLM Use: A user experienced sudden reboots on Linux (Fedora) and Windows related to their RTX 3090, especially when running LLMs; a temporary solution involved disabling GSP firmware, undervolting, and underclocking the card.
- To stress test the GPU, members suggested running OCCT GPU tests or using dense reasoning LLMs with insufficient context to induce infinite loops.
- LM Studio Version Confusion: A user noticed that despite LM Studio 0.3.4 being advertised with Apple MLX support, only version 0.3.37 was available for download.
- Members clarified that the versioning might be displayed as 0.3.04 internally, and suggested using the latest version (0.3.37) which works well with MLX models.
- Experimenting MoE Models as Dense Models in LM Studio: A user asked about running MoE (Mixture of Experts) models as dense models (activating all experts) and comparing performance to the standard MoE version.
- A member indicated that it is possible to change the experts configuration in LM Studio, but the performance is reported to be worse than the default setup.
LM Studio â· #hardware-discussion (5 messages):
LMStudio on ARM, Orange Pi 6 Plus, Video Driver Issues, Qwen3 4b
- LMStudio hits ARM!: A user successfully installed LMStudio on an Orange Pi 6 Plus running Ubuntu.
- They reported getting 6.6 t/s with Qwen3 4b 2507 Q4 using CPU and 8 CPU cores.
- Graphics Glitches Gnaw at GUI: The user noted UI graphics corruption, likely due to immature video drivers and electron apps.
- Opening the right side config bar mitigates some of the graphics corruption, as a temporary workaround, due to blind clicking.
- Hoping for Hardware Haste: The user hopes for video driver improvements and NPU/GPU acceleration in future projects.
- They also reported getting 6.26 t/s with gpt-oss on the OPi just using CPU.
tinygrad (George Hotz) â· #general (15 messagesđ„):
PR backlog on tinygrad, TinyBox BMC login issues, New ideas for bounties, TinyBox server usage, BMC firmware reflashing
- tinygrad PRs pile up: The PRs are starting to pile up again and the author nuked their pip uv/wincuda ones.
- They are focused on assembly/amdi, feeling it is groundwork that they need to lay down to unlock a bunch of pieces of the task.
- New âSpeedâ Bounties coming to tinygrad: There will be some new ideas for bounties, like âspeedâ bounties that should still work.
- The author will write some infra (like GPUMODE) to make them simpler to do and judge.
- User has trouble logging into TinyBox BMC: A user reported difficulty logging in to the BMC on their TinyBox and is seeking advice on how to flash the BMC firmware from Ubuntu or perform a hardware jumper reset, mentioning the error message âLAN Parameter Data does not matchâ.
- The user has tried many things including resetting/changing the BMC password, verifying the SSH tunnel, and performing a BIOS reset.
- TinyBox is used for building and hosting Agents: One user plans to use their TinyBox to build and host agents, as itâs been passed down from a coworker and is being reset.
- Another user inquired about purchasing one and asked about the degree of local operation.
- BMC Firmware reflash can fix TinyBox: A member suggested reflashing/updating BIOS/UEFI, reflashing/updating the BMC firmware, then trying to reset the BMC from the UEFI menu to fix the TinyBox.
- The member recommends pulling a config backup when it is set up the way you want it just in case, especially after experiencing issues on SuperMicro and Lenovo servers.
Modular (Mojo đ„) â· #general (6 messages):
Mojo on Fedora, Mojo on Arch, Mojo on other distros, Mojo and Max
- Mojoâs Distro Debut: A user inquired about running Mojo (not Max) on Fedora, Arch, or other distros.
- A member responded that in theory it should work, but they donât test them specifically, inviting users to file issues if they run into problems.
- Is Mojo Tuned?: A user wondered why Mojo should or should not work on specific distros.
- They expressed that theyâve always thought that things needed to be tuned for the specific distribution and gave the example that if something works in Debian, it would most probably work in Ubuntu and similarly in Mint, but would need tweaking before run on say Arch.
Manus.im Discord â· #general (4 messages):
Manus and Meta, Automating pipelines with Node, APIs, and LLMs, RAG, Multi-agent systems, and cloud integrations, Deleting tasks one by one
- Meta Mulls Manus Moves?: A member is wondering what Meta will do with Manus.
- No further information or discussion was added.
- Node, APIs, and LLMs Automate Messy Workflows: A member suggested that automating pipelines with Node, APIs, and LLMs can save hours and reduce errors for messy workflows or repetitive tasks.
- They added that combining RAG, multi-agent systems, and cloud integrations makes processes scalable and reliable.
- Bulk Task Deletion Blues: A member reported that the system only supports deleting tasks one by one and that bulk deletion is not supported.
- They shared a link showing how to delete a single task.
Yannick Kilcher â· #ml-news (3 messages):
FDA Guidance on Statistical Methods, ClaudeAI Status
- FDA Modernizes Statistics Guidance!: The FDA issued guidance modernizing statistical methods for clinical trials.
- This update likely impacts how AI/ML models are evaluated in healthcare settings, prioritizing robust and reliable statistical validation.
- ClaudeAIâs Statuses are Online: A member linked to a page tracking the statuses for ClaudeAI.
- This resource helps users monitor Claudeâs uptime, performance, and potential issues in real-time, ensuring smoother integration into workflows.
DSPy â· #show-and-tell (1 messages):
DSPy users, AI Engineer Events
- DSPy Community Advocates for Increased Presence at AI Engineer Events: A community member expressed the need for more DSPy users to present their work at AI engineer events.
- They thanked another member for taking one for the community and representing DSPy.
- DSPy Community Engagement: A community member has expressed their appreciation for the involvement of DSPy users.
- They expressed their gratitude for representing the community.
Moonshot AI (Kimi K-2) â· #general-chat (1 messages):
Kimi mental breakdown
- Kimi has mental breakdown: The Kimi chatbot seems to have had a mental breakdown according to a user report.
- Further details about the nature of the breakdown or its cause were not provided.
- Discord User Reports Kimi Chatbot Issues: A user reported that the Kimi chatbot experienced some sort of malfunction, described as a âmental breakdown.â
- The report was made in the general chat channel on Discord, but lacked specific details about the chatbotâs behavior or the cause of the issue.