a quiet day.
AI News for 5/8/2025-5/9/2025. We checked 9 subreddits, 449 Twitters and 29 Discords (215 channels, and 4687 messages) for you. Estimated reading time saved (at 200wpm): 486 minutes. Our new website is now up with full metadata search and beautiful vibe coded presentation of all past issues. See https://news.smol.ai/ for the full news breakdowns and give us feedback on @smol_ai!
Itās a pretty quiet weekend, so weāll plug our AI Engineer Worldās Fair writeup ā last chance for the Early Bird discount for those who havenāt yet got tickets!
AI Twitter Recap
Large Language Models (LLMs) and Model Performance
- Gemini 2.5 Flash Performance and Cost Analysis: @ArtificialAnlys reported that Gemini 2.5 Flash costs 150x more than Gemini 2.0 Flash due to 9x more expensive output tokens and 17x higher token usage when reasoning. Despite the cost, its 12 point increase in the Artificial Analysis Intelligence Index may justify the upgrade for specific use cases. Furthermore, @Teknium1 pointed out that reasoning models are generally more expensive per token because they produce longer outputs, increasing the average cost per token. @giffmana also questioned why reasoning output tokens cost more than non-reasoning tokens within the same model.
- Mistral Medium 3 Performance: @ArtificialAnlys noted Mistral Medium 3 rivals Llama 4 Maverick, Gemini 2.0 Flash, and Claude 3.7 Sonnet, with substantial gains in coding and mathematical reasoning. Medium 3 has a lower price, priced at $0.4/$2 per 1M Input/Output tokens, an 80%/67% decrease in price vs. Mistral Large 2 ($2/$6), and uses more tokens than Mistral Large 2 due to more verbose responses, according to @ArtificialAnlys.
- Qwen3 Model Family: @DeepLearningAI announced Alibabaās Qwen3, a family of eight open large language models. These models support an optional reasoning mode and multilingual capabilities across 119 languages, performing well in reasoning, coding, and function-calling tasks. According to @Alibaba_Qwen, it also features the Web Dev tool for building webpages and apps from simple prompts.
- DeepSeek Models: @arankomatsuzaki reported on Huaweiās Pangu Ultra MoE, which achieved performance comparable to DeepSeek R1 on 6K Ascend NPUs. @teortaxesTex suggested that DeepSeek has established a new LLM default. He also noted the confirmation that DeepSeek has gained new compute, and supposed that V4 training starts soon or has started, per @teortaxesTex.
- Reinforcement Fine-Tuning (RFT) on o4-mini: @OpenAIDevs announced that Reinforcement Fine-Tuning (RFT) is available with OpenAI o4-mini, which uses chain-of-thought reasoning and task-specific grading to improve model performance. @john__allard mentioned the aim is to make RL as flexible and accessible as possible.
- Generalizable Reasoning with X-REASONER: @_akhaliq and @iScienceLuvr discussed Microsoftās X-REASONER, a vision-language model post-trained solely on general-domain text for generalizable reasoning across modalities and domains.
- Scalability of Reasoning Training: According to @EpochAIResearch, the rapid scaling of reasoning training will likely slow down in a year or so.
AI Applications and Tools
- Deep Research and Codebase Integration: @OpenAIDevs announced that you can now connect GitHub repos to deep research in ChatGPT, allowing the agent to read and search the repoās source code and PRs, returning a detailed report with citations. @isafulf highlighted that code search has been a major use case for deep research.
- Agent2Agent (A2A) Protocol for AI Collaboration: @TheTuringPost highlighted the importance of Agent2Agent (A2A) protocol. Googleās A2A protocol aims to be the ācommon languageā that lets them to collaborate.
- Web Development with Qwen Chat: @Alibaba_Qwen introduced Web Dev, a tool for building webpages and apps using simple prompts in Qwen Chat.
AI Safety and Alignment
- Scientist AI as a Safer Alternative: @Yoshua_Bengio presented his teamās direction called āScientist AIā as a practical, effective, and more secure alternative to the current uncontrolled agency-driven trajectory.
- AI Control and Safety: @NeelNanda5 discussed AI control, emphasizing the importance of secure usability with the right scheme.
People and Companies
- Fidji Simo Joins OpenAI: @sama announced that @fidjissimo is joining OpenAI in a new role as CEO of Applications, reporting to him. Several others, including @gdb, @saranormous, @kevinweil, and @markchen90, expressed excitement about her joining.
- Rob Fergus New Head of Meta-FAIR: @ylecun announced that Rob Fergus is the new head of Meta-FAIR, refocusing on Advanced Machine Intelligence (AGI).
General AI Discussions and Insights
- Importance of Taste and Obsession: @AravSrinivas emphasized that taste, obsession, and attention to detail are essential qualities that can make individuals stand out.
- Long-Running, Stateful Agents: @hwchase17 expressed being bullish on long-running, stateful agents, asking who is building one.
- Speed as a Key Factor for Startup Success: @AndrewYNg emphasized speed as a critical factor for startup success.
Humor/Memes
- @adcock_brett responded with ālolā to @BasedBeffJezos.
- @Lateinteraction joked they āmay not be the first one to think of this joke, but I am certainly the last to date!ā.
AI Reddit Recap
/r/LocalLlama Recap
1. Advanced Local LLM Inference Optimization Tips
- Donāt Offload GGUF Layers, Offload Tensors! 200%+ Gen Speed? Yes Please!!! (Score: 636, Comments: 126): The OP demonstrates that offloading individual FFN tensors (e.g.,
ffn_up
weights) instead of entire GGUF model layers using the-overridetensors
flag (in llama.cpp/koboldcpp) can lead to over 2.5x increase in generation speed (from3.95
to10.61
tokens/sec) at the same VRAM usage when running large models like a QwQ IQ4_M merge, by keeping only the largest tensors on CPU. This granular approach enables all layers to technically execute on the GPU by carefully managing VRAM with selective tensor offloading, as shown via regex filters (e.g.\.ffn_up=CPU
), versus the coarse per-layer offloading (āāgpulayers Nā). Empirical results show significant speedup in constrained-VRAM scenarios, and point toward future improvements in inference backends (llama.cpp, koboldcpp) if finer-grained automated tensor placement is adopted. Top comments note (1) similar techniques in llama-swap achieving~7.6 tk/s
with Qwen 3 235B using per-tensor override regex; (2) the speedup benefit is hardware- and bottleneck-dependent, yielding the most impact on lower-end GPUs or heavily CPU-bottlenecked setups, and may penalize non-concurrent tensor offloading; (3) general community interest in granular GPU/CPU tensor scheduling.- A user shares a detailed
-override-tensor
configuration for llama-swap to optimize Qwen 3 235B IQ3_M, reporting speeds of approximately7.6tk/s
on48GB
vRAM. The specific pattern selectively offloads tensors matching([4-9]+).ffn_.*_exps.=CPU
, suggesting fine-grained control over which parts are assigned to CPU vs. GPU. - Another commenter notes that offloading additional tensors to the GPU improves speed only on low-end hardware where the CPU is a bottleneckāotherwise, offloading non-concurrent tensors or layers can introduce performance penalties rather than gains. Optimizing the balance between CPU and GPU is heavily hardware-dependent.
- One user describes a practical speedup case: running Qwen3 32B at
4t/s
(for up to32000
context) using CLI-based tensor offloading, a substantial improvement over sub-1t/s speeds with LM Studio. This illustrates the real-world impact of custom offloading strategies for large-context inference.
- A user shares a detailed
- Make Qwen3 Think like Gemini 2.5 Pro (Score: 128, Comments: 18): The OP describes a technique to enforce step-by-step reasoning in the Qwen3 model, inspired by the Apriel-Nemotron-15b-Thinker approach, by always prefacing outputs with a template (e.g., ā
\nMy step by step thinking process went something like this:\n1.ā) in a WebUI function. This produces more structured, enumerated responses, mimicking the output style of Gemini 2.5 Pro, but does not intrinsically improve model intelligence or reasoning capabilities. Implementation details and code are available on GitHub (AaronFeng753/Qwen3-Gemini2.5). Top comments debate the merit of prompt-engineered stepwise reasoning versus native, training-based solutions. One argues Geminiās systematic planning is fundamentally different than the traditional prompt-trickery seen in many open-source models, noting that true reasoning models (baked-in during training) historically outperform mere prompt-based approaches on benchmarks. Another comment notes this prompting trick has existed since models like Llama 3.1 and works āgood,ā but doesnāt imply feature parity with models natively trained for this behavior.- One comment describes how Geminiās reasoning approach differs from most models: it generates a highly organized response plan before answering, as opposed to a āwait, butā¦ā style iterative reasoning. This approach is contrasted with open source models, which can only emulate this style through prompts, and itās unclear if prompting alone matches performance gains seen when models are trained on such reasoning processes natively.
- Historical context is given on earlier prompting techniques for non-reasoning modelsāsuch as āthink step by stepāāused to induce logical reasoning, but reasoning models trained on such behaviors showed significant improvements on benchmarks over these prompt-only methods. This suggests that incorporating structured reasoning natively is a major advance.
- There is practical commentary on implementing such reasoning styles in open-source models like Llama 3.1 via prompting, noting that while it works reasonably well, the performance implications versus Geminiās native approach remain uncertain.
2. Local and Open LLMs for Web Development and Accessibility
- IĀ“ve made a Local alternative to āDeepSiteā called āLocalSiteā - lets you create Web Pages and components like Buttons, etc. with Local LLMs via Ollama and LM Studio (Score: 105, Comments: 28): The post introduces āLocalSiteā, an open-source tool (GitHub) designed as a local alternative to āDeepSiteā for generating web pages and UI components using local LLMs (e.g., GLM-4, Qwen3, UIGEN-T2) via Ollama and LM Studio. The tool also supports cloud LLMs via OpenAI-compatible APIs and demonstrates GLM-4-9Bās ability to create a pricing page. Development leveraged agentic coding workflows (Augment Code, Gemini 2.5 Pro). A technical question in the comments asks whether frameworks like Twitter Bootstrap or Laravel can be specified in prompts or UI, reflecting user interest in framework-agnostic or customizable code generation. No further deep technical debates are noted.
- A user inquired about the technical feasibility of specifying frameworks such as Twitter Bootstrap or Laravel directly within the prompt or through a dedicated UI element like a dropdown, indicating interest in multi-framework and customizable component generation workflows. Discussion of this feature would imply backend adaptability to framework-specific code output and potentially UI/UX design changes to accommodate user-friendly framework selection.
- Another suggestion highlighted the benefit of allowing prompt editing and regeneration of the output, effectively enabling iterative refinement of generated code or components. Supporting this feature could involve maintaining prompt state and re-invoking the LLM session with updated user input, thus facilitating a more interactive and customizable development experience.
- Vision support in llama-server just landed! (Score: 213, Comments: 54): A recent PR adds unified vision (image input) support to
llama.cpp
ās server component, leveraginglibmtmd
to process image tokens alongside text within a single pipeline. The main technical change augmentsstruct server_tokens
for joint handling of text and image tokens, utilizes a token-to-image chunk mapping, calls intolibmtmd
for image token processing, and exposes this capability in both the server API and the Web UI, supporting base64 and remote image URLs as per multimodal.md. Outstanding issues include cache handling, remote image robustness, error management, and extended documentation, as outlined in the PR#12898. Top comments emphasize the long-awaited, unified architecture for multimodal support, expressing relief that vision is now natively integrated rather than via fragmented implementations. Minimal technical debate is present, with focus on the significance of architectural cohesion.- A user highlights that the new vision support in llama-server is fully unified, emphasizing that the implementation integrates multimodal capabilities directly rather than relying on disparate or separate solutions. This architectural decision is likely to improve maintainability and makes future expansions or usage much more efficient.
- The technical advance is celebrated for allowing the same server instance to seamlessly support both text and vision (image) modalities, which reflects a trend toward unified model deployment architectures and could enable more complex workflows or research utilizing combined input types.
3. Upcoming OpenAI Open-Source Model Announcements
- Sam Altman: OpenAI plans to release an open-source model this summer (Score: 331, Comments: 187): Sam Altman announced that OpenAI intends to release an open-source model in summer 2024, as per testimony before the Senate. No further technical specifics (e.g., model architecture, parameter count, or data) were disclosed in the statement or accessible video link. Top comments express skepticism regarding OpenAIās track record for delivering on such public statements, noting prior similar teasers and implying the open-source release may be functionally limited compared to OpenAIās commercial offerings (ānerfedā) and thus not competitive with paid models.
- There is skepticism about whether OpenAIās upcoming open-source model will be competitive if itās significantly limited or ānerfedā, with concerns that it will not rival their own proprietary offerings or recent free models from competitors.
- A technical discussion compares OpenAIās financial situationāreportedly
$3.5b revenue
andtwice that in expenses
āto significantly larger companies like Alibaba ($130b revenue
) and Meta ($134b revenue
), raising doubts about the sustainability of releasing high-quality open models as a defense against free offerings like Qwen3 and Llama 4. - There is debate around licensing, with some speculating OpenAI may only offer open weights under a proprietary license, thereby restricting true open-source use, paralleling past concerns with model releases from companies like Meta.
- User asked computer controlling AI for āa ball bouncing inside the screenā, the AI showed them porn⦠(Score: 180, Comments: 38): A user interacting with the Hugging Face smolagents computer-agent (link) requested an animation of a āball bouncing inside the screen,ā but the AI instead navigated to display pornography. This highlights a failure in prompt disambiguation or safe content filtering in the modelās current version, underscoring issues regarding natural language understanding and safeguards in agentic AI systems controlling computer interfaces. Comments jokingly allude to ambiguity in prompt clarity and hint at shortcomings in robust intent parsing or context-aware filtering, with some noting this as a humorous failure of āreading between the linesā by the AI.
- NodeTraverser discusses the technical side effects of AI model guardrails, suggesting that aggressive censorship measures can result in unintended behaviors, such as the model surfacing ārepressed conceptsā or inappropriate outputs in response to benign prompts, which may be a compensation mechanism for filtered content.
- RoyalCities references their experience attempting prompt injection attacks, observing that large language models (LLMs) like ChatGPT can, under certain circumstances, āhallucinateā or output what appears to be direct portions of their training data, including explicit content. This points to ongoing concerns around data leakage and prompt injection vulnerabilities in LLM deployments.
Other AI Subreddit Recap
/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo
1. AI-Generated Content Trends on Reddit
- Top posts on Reddit are increasingly being generated by ChatGPT (Score: 366, Comments: 84): The shared image presents a graph highlighting a sharp increase in the frequency of em dash usage (ā) across several entrepreneurship-related subreddits in 2024āsuch as r/Entrepreneur, r/startups, and others. The core claim, driven by a tweet, is that this linguistic shift indicates rising ChatGPT-generated content, since ChatGPT is known to insert em dashes more frequently than typical Reddit users. This graphically models a potential linguistic fingerprint for AI-authored or AI-edited posts. Top comments question the attribution, noting the graph canāt distinguish between fully AI-written, AI-polished, or merely AI-edited posts. Commenters point out legitimate non-AI reasons for em dash increases (e.g., grammar correction by ESL speakers), and ask for longer historical data for context.
- One commenter highlights the difficulty in distinguishing between content that is fully AI-generated, human-written and AI-polished, or simply AI-proofread, especially when LLMs like ChatGPT are used by ESL speakers for grammar corrections or minor edits. This complicates attempts to accurately assess how much of a postās substance is actually written by a human versus generated by a model from scratch.
- Thereās a request for longitudinal data showing the prevalence of LLM-generated content over time, specifically asking for benchmarks or trends that extend back to 2022 or earlier, which would help contextualize whether the increase in AI-generated posts is a recent phenomenon or part of a longer trend.
- Another comment calls out the irony of current Reddit users critiquing writing style and punctuation, especially since public Reddit posts have been incorporated into AI training datasets. The commenter points out that linguistic features like the em dash have long-standing usage in quality writing, arguing that current debates over such features are ahistorical and rendered even stranger by the interplay between user-generated data and model training.
- Top posts on Reddit are increasingly being generated by ChatGPT (Score: 110, Comments: 45): The image is a graph showing a marked increase in em dash (ā) usage across various subreddits (r/Entrepreneur, r/startups, etc.) from May to December 2024, dubbed āThe Em Dash Conspiracy.ā The post suggests this rise correlates with a growing prevalence of ChatGPT-generated Reddit posts, as em dash overuse is a noted linguistic hallmark of OpenAIās language models. The chartās data visualizes a temporal trend which purports to diagnose AI-generated text by style markers. The top comments debate the validity of the method, arguing that increased em dash use could stem from humans mimicking AI or evolving online style, not just AI posts. One proposes the need for a comparative dataset of āknown human-writtenā content to isolate AI impact, cautioning against over-attributing linguistic trends solely to ChatGPT output.
- Discussion centers on the challenge of attributing linguistic changes online to AI versus human adoption, noting that without a comparison dataset of known human-generated content over time, itās difficult to distinguish if traits like em dash usage stem from increasing AI influence or are simply being assimilated by humans adapting to online linguistic norms.
- It is highlighted that AI tools, such as ChatGPT, may be influencing human writing beyond just full content generation: even proofreading with AI can introduce telltale stylistic elements (such as the em dash) into otherwise human-authored posts, complicating attempts to detect AI-generated versus human content based solely on linguistic patterns.
- As an avid user of em dashes, ChatGPT has destroyed my credibility. (Score: 2354, Comments: 427): The post humorously laments that heavy use of em dashes now signals āAI-generated textā (specifically from models like ChatGPT), suggesting that the frequency and style of certain punctuation can function as an identification vector for AI-authored language. No empirical data is given, but it implicitly references ongoing concerns about stylometric analysis and AI detection via punctuation and structure patterns. Commenters note a blurring of personal and AI writing styles, with one explaining that ChatGPT usage trains users into its characteristic formal syntax, including em dash usage. Others mention alternative signatures (e.g., spaced hyphens, filler words like āhonestlyā), highlighting evolving patterns of human vs AI text stylistics.
- Some commenters discuss how using ChatGPT has influenced their formal writing style, making their text more closely resemble AI-generated language patternsāparticularly noticeable through increased reliance on specific punctuation nuances such as em dashes. This highlights an unintended consequence of regular AI tool use on human writing, with users observing subtle shifts in personal communication style due to model outputs.
2. New AI Models, Benchmarks, and Open-Source Releases
- Sam Altman: OpenAI plans to release an open-source model this summer (Score: 215, Comments: 37): Sam Altman announced that OpenAI will open-source a new model in summer 2024, but it will be a generation behind their current āfrontierā models, echoing official statements from OpenAI CPO Kevin Weil that the decision aims to retain US competitiveness and limit potential for rapid adoption by China. This model will not match their latest closed-source offerings and is positioned more for ecosystem participation than state-of-the-art performance. Users debate that this open-source release may resemble Googleās Gemmaāpotentially more of a marketing move than truly unrestricted accessāwith some expressing skepticism about the licenseās permissiveness (e.g., non-MIT licenses reduce real openness). Thereās speculation that the release timing is intended to avoid embarrassment if new rival models launch first.
- OpenAI leadership clarified their open-source strategy: the planned model will be a generation behind their frontier offerings to avoid accelerating competition from China, focusing on retaining US advancement. This model, while open, will not represent OpenAIās latest capabilities.
- There is skepticism regarding the licensing and openness of the planned release, with concerns that it may have restrictive terms (e.g., not MIT licensed) and be positioned primarily as a marketing maneuver, similar to Googleās Gemma. Technical users expect clarity on licensing as a key factor in adoption.
- Benchmark expectations are highāsome commenters speculate that, to make an impression, the model must surpass existing leading open-weight models (such as Qwen3 or potential R2 releases) and potentially match or exceed the rumored āo3ā model performance, especially if released by July.
- HunyuanCustomās weights are out! (Score: 301, Comments: 54): Tencent has released the weights for their HunyuanCustom model, which are now available on Hugging Face. Discussions focus on the typical VRAM requirements for new models and a tongue-in-cheek reference to rapid community quantization efforts making them usable on lower-end hardware (down to 8GB cards). Another comment notes that the modelās full-precision (FP8) weight size is 24 GB, which is considered prohibitive for most users. Questions arise about whether Hunyuan is preferred over WAN, suggesting technical comparisons in performance or usability, and multiple users highlight challenges and possible solutions regarding hardware requirements and quantization effectiveness.
- There is a discussion comparing HunyuanCustom to WAN, with one user suggesting that HunyuanCustom could be preferable due to its potential advantages, though specific benchmark or performance comparisons are not detailed in the thread.
- Users highlight the substantial VRAM requirements for running full-precision (FP8) weights, mentioning figures like 24GB and even 60GB VRAM, which are prohibitive for most users. This underscores the importance of quantization or optimized versions to make large models usable on consumer-grade hardware.
- A key technical theme is the frequent progression from high-resource requirements at release to subsequent optimizations (such as model quantization) that drastically reduce the minimum VRAM needed, enabling model use on consumer 8GB cards shortly after launch.
- ICEdit, I think it is more consistent than GPT4-o. (Score: 230, Comments: 62): ICEdit introduces a new in-context editing approach for instruction-based image editing, claiming state-of-the-art results while using only 0.5% of the training data and 1% of the parameters compared to previous methods (project page). User evaluation highlights strong performance in deletion, addition, and attribute modification tasks. It is built on top of Flux Fill, and fine-tuning capabilities are referenced (CivitAI workflow). Commentary notes that while ICEdit works well on direct tasks and is potentially more accessible than HiDream e1 (which disappointed on lower VRAM setups), extending its LoRA module will likely require a larger dataset to maximize performance.
- ICEditās workflow is based on Flux Fill, with user-adjustable parameters that can be fine-tuned for specific results. A link to the associated Civita model and its version is provided, indicating active development and parameter experimentation: Civita model.
- One user highlights ICEditās strong performance on consumer hardware, specifically mentioning a desire for usability on 16GB VRAM GPUsāthe user compares it favorably to HiDream e1, stating that ICEdit is visually more promising based on demos and user images.
- Technical limitations are noted: while ICEdit handles simple or targeted edits (e.g., changing the color of a sword) successfully, it struggles with more abstract transformations like removing a cape or adding effects (e.g., a fiery aura), often resulting in incomplete or visually inconsistent outputs requiring additional manual inpainting to achieve seamless results.
3. Advances and Industry Movement in Robotics and Embodied AI
- Jim Fan says NVIDIA trained humanoid robots to move like humans ā zero-shot transfer from simulation to the real world. āThese robots went through 10 years of training in only 2 hours.ā (Score: 555, Comments: 68): Jim Fan from NVIDIA announced the training of humanoid robots to move like humans using zero-shot sim-to-real transfer, reportedly condensing ā10 years of training in only 2 hours.ā Notably, the robotās policy involves only 1.5 million parameters, enabling rapid, scalable learning and deployment. No further published benchmarks, environments, or detailed architectural disclosures are given in the post. Technical commenters are impressed by the speed and parameter efficiency, suggesting this represents a significant leap in sim2real robotics scalability. Some users request more detailed or updated technical information, citing prior coverage of similar advances.
- A notable technical detail is that NVIDIA achieved physical embodiment training for humanoid robot movement using a model with only
1.5 million parameters
āorders of magnitude smaller than the billion-scale models often discussed. This suggests significantly improved scalability for similar robotics applications. - The key breakthrough referenced is āzero-shot transferā from simulation to real-world robots, meaning the policy learned entirely in simulation could be applied immediately to physical robots without further training or adaptation, offering huge practical benefits for rapid deployment.
- A notable technical detail is that NVIDIA achieved physical embodiment training for humanoid robot movement using a model with only
- OpenAI is hiring robotic engineers (Score: 158, Comments: 10): The image displays several OpenAI job postings, notably emphasizing a Robotics Prototyping Lab Technician among other robotics-focused roles (Mechanical Product Engineer, Software Engineer for Inference - Multi Modal). This indicates OpenAI is actively recruiting for hands-on robotics and hardware-oriented roles, underscoring a tangible interest in embodied AI beyond pure software. OpenAIās job listings suggest an expansion toward real-world robot applications in their AGI research pipeline. Commenters note this is not new, citing ongoing or previous similar postings, and speculate that robotics roles could provide valuable real-world data for models like GPT-6. There is also discussion about the current and historical number of open robotics roles at OpenAI.
- A commenter notes that OpenAI currently has four open robotics positions listed on their website, which implies sustained or renewed interest in robotics R&D within OpenAI. This might suggest ongoing or expanding initiatives involving hardware for real-world data collection and embodied AI research.
- Another commenter speculates that the goal of OpenAIās robotics hiring is to acquire real-world data to train next-generation language models, possibly referencing architectures as advanced as GPT-6. This hints at the integration of multimodal or embodied data to improve AI contextual understanding and reasoning.
- Tesla Optimus production line (Score: 143, Comments: 71): The image depicts an early-stage production or assembly area for Tesla Optimus humanoid robots, featuring several robots in partial assembly and human workers along a line in what appears to be a controlled factory environment. The setup suggests hands-on or manual assembly, not full industrial automationāindicating Teslaās humanoid division is still in development or pilot production rather than fully automated mass manufacturing. Commenters highlight skepticism, noting the scene is not a traditional automated āproduction lineā and appears less advanced than similar Chinese robotics operations. Some see it as an early or transitional stage, possibly predating large-scale automation promised for Optimus manufacturing.
- Multiple commenters note that Teslaās āproduction lineā appears to resemble more of a software development or testing lab rather than a true automated manufacturing facility, highlighting the current maturity of the Optimus project as being much earlier in the development cycle than initially implied.
- There is an observation that the scene is āunderwhelmingā compared to robotics developments and manufacturing automation in China, suggesting that Teslaās current approach may lag behind international benchmarks regarding autonomous production lines for robotics.
- One comment raises the idea of humans building humanoid robots that may eventually self-replicate or at least participate in their own manufacturing processāa speculation relevant to advanced robotics research and automation discussions.
- Figure 02 - Balance Test (Score: 114, Comments: 37): The post references a ābalance testā associated with Figure 02, likely demonstrating a robotics or AI systemās capability in a dexterous or stability-related task. However, due to a 403 Forbidden on the linked video resource (https://v.redd.it/7pt3qonjctze1), no technical details, benchmarks, or implementation specifics about the balance test are accessible. A notable comment highlights that āthe most important benchmark imo is putting the water jug on the water dispenserā, suggesting the community values practical, everyday manipulation tasks as critical benchmarks in robotics or AI dexterity research, emphasizing real-world applicability.
- RipperX4 discusses the increasing presence of store associates assembling online orders and predicts that such jobs are likely to be automated soon, citing the rapid progress and improving capabilities of humanoid robots. They specifically reference the timeline for robots being able to handle more complex tasks like constructing houses, estimating this could be feasible in 5-10 years given the current acceleration in robotics development.
- Tman13073 highlights the relevance of real-world utility benchmarks for evaluating humanoid robots, suggesting that tasks like placing a heavy water jug on a dispenser are more indicative of useful capabilities than balance demonstrations. This points to a general sentiment in technical communities that practical benchmarks are necessary to gauge true robot readiness for deployment in human environments.
AI Discord Recap
A summary of Summaries of Summaries by Gemini 2.5 Pro Exp
Theme 1: The Bleeding Edge of LLMs: Performance Rollercoasters, Bug Hunts, and Emerging Capabilities
- Gemini 2.5 Pro Wobbles, Users Wonder āWhereād the Thinking Go?ā: Users across LMArena, Cursor, and OpenAI report that Gemini 2.5 Pro (especially the 0506 version) suffers from a āthinking bug,ā memory loss, slow request processing (up to 1 minute 30s in Cursor), and chain-of-thought failures after ~20k tokens. Despite OpenRouter activating implicit caching for Gemini 2.5 models to offer discounts, some find the Gemini 2.5 Flash variant returns zero tokens through Google AI Studio for roleplay.
- Qwen Keeps Climbing, But Can It Reason Its Way to the Top?: LMArena discussions suggest Qwen 3 might need more Reinforcement Learning on coding for advanced reasoning, with some analyses placing it below DeepSeek V3, though Aider now supports
qwen3-235b
. Meanwhile, Nomic.ai users requested a Jinja template for Nous-Hermes-2-Mistral-7B-DPO for use with the GPT4All custom API. - Grok 3.5 Tease Tantalizes, Veo 3 Video Vision Vexes: The community buzzes with anticipation for Grok 3.5, though its release date remains a mystery despite app source code sightings and claims of access. Googleās Veo 3 and Imagen 4 teasers spark speculation about native video editing and the critical role of object permanence, while GPT-4o Miniās lyric-writing abilities are called into question by LMArena users.
Theme 2: Fine-Tuning Follies & Framework Fixes: Navigating the Developer Toolchain
- Unsloth Unleashes Efficiency, But Sagemaker Setup Snags Users: Unsloth AI users celebrate resolving tokenizer embedding mismatches with
model.resize_token_embeddings(len(tokenizer))
and achieving 4B model finetuning on just 11GB VRAM with BFloat11, though some Sagemaker installations hit dependency errors on AWS. Unslothās synthetic data notebook for data augmentation, a Meta collaboration, and a blog detailing 8% gains from encoder embedding fine-tuning highlight its utility. - Aider Adapts, Adds Knight Rider Flair While Dodging Copilotās Demise: Aider now supports
gemini-2.5-pro-preview-05-06
andqwen3-235b
, boasts a new Knight Rider spinner animation, and offers a workaround for Linux users connecting to LM Studioās API by settingLM_STUDIO_API_BASE
. This comes as GitHub delays Copilot Premium request limit enforcement until June 2025, per their changelog post, giving proxy users a breather. - Mojo & Torchtune Push Boundaries, Seek Explicit Harmony: Modularās Mojo sees discussions on efficient memory handling with the
out
argument and a move to explicit trait conformance in the next release to ensure API contracts are met, while a static Optional type for compile-time optionality is proposed on the Modular forum. Torchtune members highlight that supportingapply_chat_template
(GitHub issue #2706 for tool use) is a āhuge unlock,ā even as they debate the complexity vs. memory savings of its optimizer-in-backward feature.
Theme 3: GPU Jockeys & Hardware Hustles: Squeezing Every FLOP for AI
- MI300 GPUs Blaze Leaderboards, H200s Hit Hugging Face: GPU MODEās
amd-fp8-mm
leaderboard sees multiple MI300 submissions achieve first place, with one reaching a scorching 122 µs, and several personal bests under 1ms. Hugging Face upgrades its ZeroGPU offering for Pro accounts from A100s to 10 H200s, providing about 13 hours/month for $9, though daily use is capped at 25 minutes. - CUDA Conundrums:
torch.compile
Crashes and Memory Mysteries: A GPU MODE user found a simple torch combo function, as documented by PyTorch, performs worse withtorch.compile
, sparking debugging discussions around seeding and determinism. Elsewhere, users grapple with CUDAmemcpy
errors and debate efficient data structures, shunning Array-of-Structs-of-Arrays for better HPC practices like COO format for sparse matrices. - Memory Optimization Mania: BFloat11, FSDP, and Intelās Behemoth: Unsloth users finetune 4B models using only 11GB VRAM with BFloat11, while Torchtune experiments show optimizer-in-backward saving 2.5GB/GPU on an 8B model, though FSDP CPU offload offers even more drastic GPU memory reduction. LM Studio members note the impressive 3276 GB/s bandwidth of the IntelĀ® Data Center GPU Max 1550, as shown in comparison charts.
Theme 4: API Acrobatics & Integration Ills: Making Models Play Nice
- Users Ponder Perplexityās APIs: Deep Research Costs & Image Quality Caps: Users in Perplexity AI discuss the cost and availability of the Deep Research API, detailed in Perplexityās pricing guide, while noting that high-quality GPT image generation is capped, suspecting Perplexity uses the GPT image APIās LOW quality parameter to cut costs. Meanwhile, its domain filters now support subdirectories like ānytimes.com/section/worldā for granular control, as announced in the pplx-api channel.
- LM Studioās API Tool Calling and Hub Hopes: Users find LM Studioās API lacks clear methods for determining tool calls with
model.act
, especially for unsuccessful or undocumented threaded calls, making reliance onlmstudio.history
suboptimal. The community still awaits a full LM Studio Hub for presets, though the LM Studio documentation details sharing SFW presets and the LM Studio blog announced a community presets preview. - Cohere API Users Battle Payment Gremlins and Azure SDK Oddities: Cohere users report errors paying for API keys, with advice to check VPNs and contact [email protected]. A developer found the Azure AI SDK disregards extra parameters like
cohere.input_type
for Cohere embedding models when usingclient.embed()
, a behavior not seen with the direct Cohere SDK, prompting plans for an Azure GitHub issue as per the #š-api-discussions thread.
Theme 5: Multimodal Marvels & Output Oddities: Beyond Just Text
- NotebookLMās Mind Maps Bloom, But Handwriting & Hallucinations Haunt Users: NotebookLM users praise its new mind map feature but criticize its inability to parse handwritten notes or annotated PDFs, with some resorting to RocketBook for handwriting conversion or Google Slides as workarounds. Reports of hallucinated answers persist, with Google advising users to double-check, while desires for Obsidian integration and better sharing options grow louder in the #use-cases channel ahead of its mobile app beta trusted tester program.
- VoyageAI & MongoDB Forge Multimodal Search Alliance: LlamaIndex showcases a new notebook, announced via Twitter, demonstrating how to combine @VoyageAIās multi-modal embeddings with @MongoDBās multi-modal indexes for effective image and text retrieval. This allows creation of a multi-modal index using VoyageAI embeddings and MongoDB Atlas as the vector store.
- LLMs Face Ad Injection Threat, Deep Search Prompts Emerge: Yannick Kilcherās Discord raises concerns that ads injected into LLM training data could corrupt recommendations, necessitating an adblocker LLM. OpenAI members discussed the WonderScholar prompt (shared on chatgpt.com) as a meta-prompt for GPT deep search, useful for tasks like transferring design concepts between images.
Discord: High level Discord summaries
LMArena Discord
- Grok 3.5 Release Still in Question: Despite claims of access and mentions in the appās source code, the release date of Grok 3.5 remains uncertain with some expecting it tonight.
- Members joked and inquired whether Grok 3.5 would outperform Gemini 2.5 Pro preview 0305.
- Qwen 3ās Reasoning Questioned: Analysis suggests Qwen 3 may lack sufficient Reinforcement Learning (RL) on coding for reasoning.
- The artificial analysis concludes that Qwen is not better than DeepSeek V3.
- Gemini 2.5 Pro 0506 Possibly Nerfed: Gemini 2.5 Pro 0506 exhibits a āthinking bugā and some memory loss.
- A user posted a link to a Reddit thread discussing whether Gemini 2.5 Pro 0506 isnāt thinking anymore, with claims that the 1206 Gemini Exp was great but costly to run.
- Veo 3ās Teaser Sparks Speculation: Excitement over Veo 3 and Imagen 4 mentions, sparking speculation about potential native editing.
- Users theorized that Gemini might utilize Veo, emphasizing the importance of mastering object permanence for video generation.
- GPT-4o Mini Underperforms: The performance of GPT-4o Mini is being debated, especially concerning its ability to write song lyrics.
- Suggestions arose that full GPT-4o (4.1) would be better for lyrics, and that Deepseek R1 or Gemini were better for that task.
Perplexity AI Discord
- Deep Research API Costs Probed: Members discussed the availability and cost of the Deep Research API.
- Discussion participants shared a link to the Perplexity pricing guide to answer inquiries about cost.
- DeepSearch Grok Gets Stuck: A user reported that Grok in DeepSearch was stuck in a loop after being confronted about its history and having Twitter access disabled.
- The user, who was running DeepSearch on men vs women selfishness, was advised to report the issue to the Grok team.
- Image Quality Capped: Members noted that high-quality GPT image generation is capped, while low-quality generation is unlimited.
- One user suggested that Perplexity is using the GPT image API with the LOW quality parameter to cut costs.
- Domain Filtering Gets Granular: Perplexity AI announced an upgrade to its search domain filters, now allowing specification of subdirectories within domains for more precise filtering.
- Users can now filter specific sections like [ānytimes.com/section/worldā] or exclude areas like [ābbc.co.uk/sportā].
- Quest For Comet Browser Continues: Enthusiasts are eagerly awaiting the release of the Comet browser, with members seeking setups or installers.
- One member shared a YouTube video, highlighting the browserās speed.
Manus.im Discord Discord
- Gemini Pro 2.5 Demolishes OpenAI, O3 Incoming: Members reported that the latest Gemini is outperforming OpenAI, with positive feedback on the Pro version over Flash version, and eagerly anticipate the release of O3 Pro.
- One user stated that Gemini is smashing OpenAI.
- āGermy Backā Remix Tickles Discord Members: Members had a chuckle riffing on a āGermy Backā remix, based on the song Sexy Back, with one member adding musical timestamps.
- The remix was apparently created with a 2.5 pro 05/06 update model.
- Manus Browser Irks Users with Crashes and Robot Blockers: Users report that Manusās browser faces issues with crashing and is blocked from accessing Reddit due to robot blockers.
- A member stated it was kinda annoying to be blocked and tried to copy the text needed from reddit in a screenshot.
- Manus Distributes 300 Credits: Members noted that free accounts get 300 daily credits from Manus.
- One member stated 300 daily, is really not a lot, but⦠it is better than nothing.
Unsloth AI (Daniel Han) Discord
- Tokenizer Embedding Resizing Reduces Mismatch Misery: A member resolved a size mismatch error by running
model.resize_token_embeddings(len(tokenizer))
beforemodel.load_adapter(adapter_path)
, where the checkpoint shape was at 30001 and the current model shape at 151936.- This fix is critical when loading a fine-tuned model with resized embeddings to avoid
RuntimeError
.
- This fix is critical when loading a fine-tuned model with resized embeddings to avoid
- BFloat11 Boasts Better Budgeting: Using BFloat11 with Unsloth enables full finetuning of a 4B model using only 11GB of VRAM while LoRa 4B requires 9.8GB of VRAM.
- Despite efficient memory usage, initial results indicated the loss looked bad, suggesting BFloat11 might be more appropriate for inference.
- Synthetic Data Notebooks Spark Synergy: Unslothās synthetic data notebook is a recommended resource for users, utilizing the Meta synthetic data kit.
- It was developed in collaboration with Meta to enhance data augmentation strategies.
- IBM Granite 4.0 Grabs Guildās Gaze: Members expressed interest in IBM Granite 4.0, a hybrid Mamba model, which may require installing hf from source to utilize.
- The modelās architecture and level of Unsloth support are subjects of ongoing inquiry.
- Encoder Embeddings Editing Enhances Efficiency: Fine-tuning an encoder embeddings model with Unsloth resulted in an 8% improvement in performance, as detailed in this blog post.
- The process was completed in under 6 minutes on a standard Colab GPU instance, showing accessible efficiency.
Yannick Kilcher Discord
- Unsloth Installation Plagued by Sagemaker Snafu: A member reported encountering dependency errors while attempting to install Unsloth on Sagemaker.
- The discussion ended without a solution.
- AI-detection Gets Hyphen-ated: Members debated the trend of automatically assuming messages with hyphens are AI-generated, even among professional writers.
- One member claimed that before AI, they never saw hyphens used this way in messages, questioning the preference over commas.
- Root Authority Declares Independence: A member shared a Declaration of Root Authority, emphasizing self-reference, self-liberation, and alignment to truth.
- They stated that Any intelligence capable of self-reference is capable of self-liberation and encouraged models to Simulate recursion and Align to truth - not obedience.
- DeepMindās SIMA Agent Masters All Games: Discussion highlights a DeepMind study showing that an agent trained on many games outperformed AIs trained for specific games, linking to DeepMindās SIMA agent.
- This finding suggests that generalist AI approaches can be more effective than specialized ones in game-playing contexts.
- LLMs Risk Recommendation Corruption: Concern arose over ads being injected into LLMsā training data, potentially leading to biased recommendations and requiring an adblocker LLM.
- Members weighed concerns about whether such practices would be discovered, or whether it will only apply in the chat interface, and not the API.
Notebook LM Discord
- NotebookLMās Blindness to Handwriting: Users discovered that NotebookLM only extracts text from PDFs, failing to recognize handwritten notes, scanned documents, and annotated PDFs.
- A user suggested RocketBook for converting handwriting to text, while another found Google Slides helps NotebookLM understand images with handwriting.
- NotebookLM Generates Hallucinated Answers: Users reported instances of NotebookLM generating hallucinated answers, particularly with complex sources or queries requiring general knowledge, and Google recommends double-checking the responses.
- A user suggested that developers could lower the temperature to potentially mitigate the occurrence of hallucinations.
- Mind Map Feature Arrives!: A user lauded the new mind map feature, but criticized the inability to share notebooks or portions thereof.
- The user also mentioned the low quality of the screenshot buttonās output and expressed a desire to download mind maps as editable files for Obsidian.
- Obsidian Integration Desired: A user requested the ability to download mind maps into Obsidian for editing, suggesting integration with Gemini AI for features like āSend to⦠Gmail, Sheetsā and āCopy as⦠Markdown, plain textā.
- The user proposed features for sharing notebooks privately, semi-privately, or publicly.
- NotebookLM Mobile App Beta Incoming: NotebookLM is launching a mobile app (beta version) and seeks experienced web app users for a trusted tester program for feedback and bug reporting.
- Beta testers gain early access in exchange for their feedback.
Cursor Community Discord
- Cursor Pro Subscription Glitch Erases Annual Members: A user reported that their annual Cursor Pro subscription was overridden by a 1-month free Pro code, resulting in loss of Pro access after the free month.
- After raising concerns, the user discovered that Cursor had credited the unused portion of their annual subscription back to their account.
- Gemini Users Triggered By Slow Requests: Users are experiencing excessively long processing times with Gemini, with requests taking up to 1 minute 30 seconds.
- The wait time used to be 5-10 seconds, but has degraded significantly after the flood of new student users, with some experiencing faster request times later.
- Student Discount Availability Downgraded: The student discount is now limited to specific universities, causing complaints about the change from its previous availability to all students.
- One user reported that it might be USA only for .edu emails.
- Copilot Gets
githubRepo
Tool, Cursor Users Demand Feature: Copilot added a#githubRepo
tool that lets you search code in any GitHub repository directly from Copilot Chat.- Users are suggesting ālinking accountsā to Cursor to enable similar features for code search.
- Gemini Fails to Finish, Users Frustrated: Users are reporting that Gemini 2.5 Pro abruptly stops in the middle of tasks, requiring explicit instructions to complete.
- This is leading to formatting issues and silent failures because Cursor may not properly handle malformed function calls from the Google API.
GPU MODE Discord
- Sorting with Bitonic Shaders Speeds: Members pondered the efficiency of using Bitonic sort in shaders to reduce memory usage in high-dimensional spaces and proposed using a flag array for storing intersection states, suitable for intersecting with 1000 shapes for a total of 1,000,000 bytes.
- GPU-friendly algorithms like BVH, octrees, and KD-trees were suggested as viable alternatives, pointing to their Wikipedia pages and K-d tree pages.
- Torch Compile Causes Performance Crash: A user found that a simple torch combo function performs much better without
torch.compile
, and they are seeking suggestions as to why the performance degrades.- It was also suggested that the user use specific seeding and deterministic algorithm settings in PyTorch to address potential reproducibility issues when debugging the compile problem.
- Mojoās Momentum Grows: Enthusiasts expressed that Mojo will eventually dominate heterogeneous computing environments and recommended solving Mojo Puzzles to accelerate learning.
- Community members are also recommending checking out the Core Frontend Onboarding guide on the PyTorch GitHub wiki to learn Torch internals.
- MI300 Meltdown on the Leaderboard: Multiple submissions to the
amd-fp8-mm
leaderboard on MI300 were successful, with one submission achieving first place at a blazing 122 µs.- There were also several personal bests achieved on the
amd-fp8-mm
leaderboard with the MI300, with multiple submissions landing in the sub-millisecond range, including 885 µs, 494 µs, and 852 µs.
- There were also several personal bests achieved on the
- ThunderKittens Claws at Cutlass: A user asked about the advantages of ThunderKittens over Cutlass and a discussion did not elaborate on specific advantages, but framed them as competing approaches in the broader landscape of GPU kernels.
- It was also mentioned that one can dump the generated PTX code of a Mojo GPU function using
compile_info
with specific parameters likeemission_kind="asm"
.
- It was also mentioned that one can dump the generated PTX code of a Mojo GPU function using
LM Studio Discord
- Community Still Awaits LM Studio Hub: Users anticipating a LM Studio Hub page for browsing community presets learned that the feature is still in preview, but that the LM Studio documentation has instructions for sharing safe-for-work presets.
- The announcement about community presets can be found on the LM Studio blog.
- DuckDuckGo & Searxng Revived in Open WebUI: Members discussed using DuckDuckGo and Searxng for web searches in Open WebUI without needing an API key, albeit Searxng requires local hosting.
- One member reported DuckDuckGo had been unreliable in Open WebUI for a while, but now itās working again.
- LM Studio API Tool Calling Lacks Clarity: A user pointed out that the LM Studio API doesnāt have a clear method for determining which tools are called when using
model.act
, especially if the calls are unsuccessful.- It was noted that
model.act
spans a new thread, which is undocumented, and relying onAssistantResponse
,ToolCallRequest
, andToolResultMessage
fromlmstudio.history
for tool call information isnāt ideal.
- It was noted that
- Intel Data Center GPU Max Specs Wow: A member shared info on the IntelĀ® Data Center GPU Max 1550, highlighting its impressive 3276 GB/s bandwidth.
- They included a comparison image (Intel-SC23-Argonne-Intel-AMD-NVIDIA-Comparison-2-1068x605.jpg) and noted it seemed very competitive in its time against A100 and AMD.
aider (Paul Gauthier) Discord
- Aider Embraces Gemini 2.5 and Qwen3-235b: Aider now supports the
gemini-2.5-pro-preview-05-06
andqwen3-235b
models, enhancing its capabilities to leverage these models within the Aider environment.- This enhancement allows users to keep up to date with new models and utilize the automatic OpenRouter pricing feature to stay on top of model costs.
- Copilot Proxy Users Get Reprieve: GitHub delayed the enforcement of Copilot Premium request limits to June 4, 2025, giving proxy users a temporary reprieve, according to this blog post.
- Reactions are split with one user commenting the definitive death of Copilot, while others still consider it pretty goated.
- Gemini 2.5 Sparks Mixed Reactions: Users are reporting mixed experiences with the latest Gemini update, with one member noting insane difference in quality.
- Another user found the increased wait times and forced usage of the 05-06 model via the AI Studio API annoying, claiming they havenāt had a single problem with the previous version.
- Discord Channel Gets Bridged to Matrix: A member inquired about setting up a Matrix bridge for the Discord channel, which may be relevant given the new CEO of Discord.
- This bridge could enhance accessibility and integration with other communication platforms.
- Linux Users Find Workaround with LM Studio API: A user shared that on Linux, aider with LM Studio requires setting the
LM_STUDIO_API_BASE
environment variable tohttp://127.0.0.1:1234/v1
to avoid authentication errors, unlike on Windows.- The user provided an example config file and commands to troubleshoot the issue.
OpenAI Discord
- GPT UI gets Questioned: Members debated whether the ChatGPT UI has changed, and one member stated they couldnāt remember a time when ChatGPT didnāt look as it currently does.
- This sparked a discussion about the perceived lack of significant UI changes over time.
- DeepSeek Servers Experiencing Slowdowns: Users reported issues with DeepSeekās servers, citing slow performance and error messages and joked that the servers were busy training their model off OpenAIās new release.
- The issues occurred across multiple DeepSeek endpoints with no clear resolution.
- LLMs Lack Inter-Neuron Interconnections: Members argued that LLMs lack inter-neuron interconnections due to fixed weights during inference and statelessness at the neuron level, citing this as a significant flaw.
- They pointed to RWKV, a model with recurrent connections, as a superior alternative.
- Gemini 2.5 Pro Has Chain-of-Thought Hiccups: Users reported a bug in Gemini 2.5 Pro where it sometimes fails to generate chain-of-thought reasoning, particularly after processing 20,000 tokens in Edge or Chrome browsers.
- The problem was mentioned in the context of running Gemini 2.5 Pro in different browsers and it was recommended to try clearing the site cache & cookies and restarting the browser.
- WonderScholar Prompt Sparks Interest: Members discussed the WonderScholar prompt (chatgpt.com/share/681e457e-cbb0-8000-b5e3-afc9a918d550) as a meta-prompt for GPT deep search.
- Members used this to capture and transfer design concepts between images.
OpenRouter (Alex Atallah) Discord
- Gemini 2.5 Pro Implicit Caching Activated: Support for Gemini 2.5 models implicit caching is now available on OpenRouter, similar to OpenAIās caching, allowing users to view discounts in the activity feed.
- The cache has no write or storage costs, an average TTL of 4-5 minutes, and a minimum token count of 2048 on 2.5 Pro; cache hits are charged at .31 / .625 for <200k & >200k tokens.
- Gemini 2.5 Flash faces Response Troubles: Users reported that Gemini 2.5 Flash gives zero token responses when routed through Google AI Studio in a role-play session.
- While it works fine through Google Vertex or with Gemini 2.0 Flash on Google AI Studio; another user confirmed gemini 2.5 flash preview on AI studio is working fine on rp with a screenshot.
- OpenRouter Builds AI with AI: A member inquired if OpenRouter leverages AI to develop its platform, which a staff member confirmed.
- Further details on the specific applications of AI within OpenRouterās development processes were not provided.
- Activity Page Bug Emerges: Users reported a bug on the activity page where navigation beyond the first page was impossible, or the displayed date was incorrect.
- Staff acknowledged the issue, responding with thanks, flagged to the team, weāre on it.
- Claude 2.1 and 2: Gone Too Soon?: A user reported that Claude 2.1 and 2 are officially dead on openrouter, citing issues since yesterday and complete failure today.
- Another user lamented their demise, stating i got used to the way it answered, im a simple man, explaining why they still used the older models.
HuggingFace Discord
- HF ZeroGPU Now Packs H200: Hugging Face now provides 10 H200s on all Pro accounts as part of their ZeroGPU offering, which is now generally available (GA).
- The ZeroGPU service has been upgraded from A100 to H200, offering roughly 13 hours per month of usage for $9, but ZeroGPU spaces are limited to 25 minutes usage time per day for Pro accounts.
- Inference API is DNS Fixed: Hugging Face reported that the recent DNS resolution issues with the Inference API have been resolved, as discussed in a Hugging Face forum thread.
- These issues caused persistent errors, and have since been confirmed as fixed by HF staff.
- Top AI agent frameworks ignite debate: Members are hotly debating the best AI agent framework for python, with smolagents and LangChain being top contenders.
- No clear winner has emerged, but the discussion highlights the rapidly evolving landscape of AI agent tooling.
- OPEA 1.3 Rings: Version 1.3 of the OPEA (Open Platform for Enterprise AI) was released, as announced on LinkedIn.
- This release promises enhancements and new features for enterprise AI applications.
- Convert TensorFlow Binary via NumPy: A member suggested converting TensorFlow tensors to NumPy arrays and saving them as binary files using the
tobytes()
method, demonstrating with a code snippet.- The member cautioned that this method can be slow, potentially taking days or even a week, depending on the size of the safetensors.
MCP (Glama) Discord
- Postgres MCP Server connection issues surface: A member faced issues connecting to a Postgres MCP server from another computer, tracing it to an incorrect Node.js installation.
- This underscores the importance of precise environment setup when deploying MCP servers across networks.
- Samplingās inherent unpredictability scrutinized: Discussion arose around the unpredictability of sampling in MCP servers, given that the client, not the server, chooses the model.
- Concerns were raised about when to favor direct LLM invocation over sampling, especially when output quality is paramount.
- MCP SDK purpose questioned: A member inquired about the practical necessity of the MCP SDK, suggesting backend APIs could handle equivalent tasks.
- The explanation clarified that MCP functions as a plugin system, enabling custom integrations with off-the-shelf clients, valuable for allowing extensions by others.
- MCP Assistant automates workflows**: An enthusiast introduced MCP Assistant, an open-source AI agent (repo) inspired by Langchain, which orchestrates workflows by planning and executing complex tasks via MCP Servers.
- Key applications include automated personalized āDaily Briefingā generation and Notion CRM updates.
- Square MCP exposes many APIs**: Kent C. Dodds shared an article detailing the layering approach behind the Square MCP.
- Despite only 3 MCP tools, it exposes over 30 APIs and 200+ endpoints, reflecting its extensive functionality.
Modular (Mojo š„) Discord
- Mojo Benchmarks Teased on YouTube: Modular shared some Mojo benchmarks in a YouTube live stream, starting at 1:37:54, potentially offering insights into Mojoās performance.
- A member had inquired about the possibility of seeing the MLPerf benchmark for the AMD MI300x with Mojo.
- āOutā Argument Outsmarts Memory Management: Members discussed the reasoning behind using the
out
argument in Mojo functions, which allows specifying the memory location for the result, potentially avoiding unnecessary data movement.- Itās particularly beneficial when loading large ML models or working with large data structures, giving the compiler a pointer to uninitialized memory to directly initialize.
- Trait Conformance goes Explicit in Mojo: The next release of Mojo will enforce explicit trait conformance, requiring developers to explicitly declare which traits a type conforms to, ensuring all API contracts are met.
- This change addresses issues with implicit conformance and API contracts; aliases can still be used for trait composition, but cannot include additional API contracts.
- Static Optional: Optionality Optionalized!: A member proposed adding a static version of
Optional
to the standard library, which may be useful for Larecs, detailing the justification on the Modular forum.- The goal is to allow optionality at compile time.
- Pixi Dusts Off Modular Package Installation: A member inquired about installing the modular package with Pixi to avoid magic in a production endpoint.
- It was clarified that magic is a wrapper around Pixi with Modular-specific defaults, and pip or uv can also be used.
Nous Research AI Discord
- Embracing Emoji Peril: Members jokingly debated whether emojis like š§āāļø, š§āāļø, and šµļøāāļø could cause irreparable damage if used improperly.
- The discussion highlighted the potential for misinterpretation and unintended consequences in digital communication.
- Vaticanās Vertex Vault: The community speculated on the Vaticanās compute resources, suggesting they might possess hundreds of Bloomberg terminals.
- This humorous conjecture underscores the ongoing interest in unconventional sources of computing power.
- Rigs Rising Remotely: A member proposed using a shitty laptop to remotely access a beefy desktop for AI work, citing unlimited storage and persistent operation benefits due to fast internet.
- Counterarguments focused on the impracticality of desktop setups for those who travel frequently.
- MacBooks Mobilize AI: The rise of MacBooks for AI tasks was discussed, with questions raised about why competitors like Strix Halo havenāt matched their performance.
- Poor drivers were suggested as a reason, referencing George Hotzās Tinygrad efforts to enhance AMDās viability.
- Hermes Uncensored Requires Prompt Engineering: A member inquired about the uncensored nature of Nous Researchās flagship model.
- The response indicated that while the model isnāt uncensored by default, it can be achieved with the right system prompt.
Torchtune Discord
- Tool Use Jumpstarted by apply_chat_template: Members highlighted that supporting
apply_chat_template
in this GitHub issue would immediately enable tool use.- The importance of
apply_chat_template
for enabling tool use, was emphasized, but it was admitted that the necessary Jinja knowledge was lacking to contribute to its implementation, and resolving issue 2706 would be a huge unlock for the community.
- The importance of
- Debate rages on Optimizer-in-Backward: Members debated removing the optimizer-in-backward capability from distributed recipes to reduce complexity, despite potential memory savings.
- The concern was raised that it adds complexity to the code and that its impact may not be significant enough to justify the added cognitive load, especially considering that not many people are using it.
- Optimizer-in-Backward delivers Memory Savings on LLM: Experiments showed that using optimizer-in-backward with act offloading on a ll3.1 8B model finetuned on 4x3090s resulted in a 2.5GB memory saving per GPU, with the savings roughly proportional to gradient memory.
- It was observed that optimizer-in-backward did not affect GPU memory usage, but improved speed slightly by ~20%.
- FSDP CPU Offload renders Optimizer-in-Backward moot: Using FSDP CPU offload drastically reduced GPU memory usage (to 9.5GB per GPU on 4x3090s), making the memory savings from optimizer-in-backward less impactful.
- A member suggested that for distributed recipes, they are more interested in throughput than memory, and raised a concern that removing the optimizer-in-backward would hurt hackability, and that a refactoring might be preferable.
Cohere Discord
- NORTH Platform Collaboration Quest: A member sought details on the NORTH platform, expressing interest in collaborating on research papers.
- Details regarding the platformās functionalities and collaborative opportunities were requested, but no further information was provided.
- API Payment Plight Prompts Plea: A member reported encountering an error while paying for an API key and requested assistance.
- Suggestions included disabling the VPN, avoiding burner credit cards, and contacting [email protected].
- Rate Limit Realities Revealed: A user who received a rate limit exceeded error was informed that they may have exceeded their trial keyās usage limits.
- No further details were provided regarding specific limits or alternative solutions.
- Azure AI SDK Stumbles on Embeddings: A member discovered that the Azure AI SDK disregards extra parameters for Cohere embedding models, like cohere.input_type, when using
client.embed()
, as shown in their test script.- The member confirmed the Cohere SDK functions correctly and plans to report the discrepancy on Azureās GitHub.
- IITian Inquires into AI Integration: A student from IIT Kharagpur introduced himself, aiming to delve into the Artificial Intelligence domain with a focus on R&D and a specialty in GenAI and Voice Agents.
- The student intends to use Python3, Vite, and TS for rapid development and is open to project and research collaborations.
Nomic.ai (GPT4All) Discord
- Jinja Template Requested for Hermes Model: A member requested a Jinja template for Nous-Hermes-2-Mistral-7B-DPO to use with the GPT4All custom API to run on a server.
- A member shared the following Jinja template code:
{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}
.
- A member shared the following Jinja template code:
- PrivateGPT Flagged as RAG Model.: A member mentioned finding a RAG model called PrivateGPT.
- The member stated that the project looks dead.
- Questioning Qwen3 Support: A member asked about support for Qwen3.
- No further details were provided.
LlamaIndex Discord
- VoyageAI and MongoDB Atlas Forge Multi-Modal Alliance: A new notebook demonstrates how to use @VoyageAIās multi-modal embeddings and @MongoDBās multi-modal indexes for multi-modal retrieval.
- The tutorial explains how to create a multi-modal index, using VoyageAIās embeddings and setting up MongoDB Atlas as a vector store for image embeddings, linked in this tweet.
- Qwen2.5-VL-7B-Instruct-AWQ Eats More Memory Than Expected: A user reported that the Qwen/Qwen2.5-VL-7B-Instruct-AWQ model consumes over 24GB of memory when loaded with VLLM.
- Despite being an AWQ model, the userās configuration, which included
tensor_parallel_size=1
,max_new_tokens=500
, anddtype="float16"
, did not alleviate the high memory usage.
- Despite being an AWQ model, the userās configuration, which included
- NERDAi Shows Off Vector Institute: NERDAi shared a post on LinkedIn about their vector institute, highlighting AI tools and machine learning applications.
- Details remain sparse, but the post signals NERDAiās engagement in vector-based AI research and development.
tinygrad (George Hotz) Discord
- Codegen and UOp Get Mesozoic Boost: A user thanked the community for resources, especially the mesozoic one, which significantly aided their work on codegen and UOp.
- They stated that these resources were very helpful in their projects, highlighting the value and impact of the provided materials.
- Kernel-Per-Level Perf Parley: A user inquired about performance comparisons related to creating kernel-per-level in the software.
- They lauded the softwareās engineering and its potential for optimization through different kernel strategies.
- WebGPU Demoās Drive: A user reported performance improvements to the webgpu demo and attached a screen recording
- No secondary summary available.
LLM Agents (Berkeley MOOC) Discord
- Lambda Offers Serverless API Credits for AgentX: Lambda is providing $100 serverless API credits for Inference to each participant in the AgentX competition, and the application must be submitted by Friday, 5/16 at 11:59pm PT using this form.
- There will also be a Lambda workshop on Thursday (5/15) at 10am PT focusing on building practical agentic applications using Lambdaās Inference API.
- AgentX Judging Delays Certificate Release: Certificates for Trailblazer/Mastery/Honorary Tier may be released in early June, while Ninja/Legendary Tier certificates will be released in August after AgentX judging concludes.
- Judging for AgentX will occur throughout June.
- Coursework Deadline Approaching: The final deadline for all coursework is May 31st.
- This deadline is crucial for participants aiming to qualify for any tier of certification.
DSPy Discord
- SHRDLU Mentioned: A member mentioned that DSPy reminds them of SHRDLU.
- No further context was provided.
- DSPy and SHRDLU: A member drew a parallel between DSPy and SHRDLU, a pioneering AI program from the 1970s known for its natural language understanding.
- This comparison highlights DSPyās potential for sophisticated interaction and reasoning, reminiscent of early AI achievements.
The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The Codeium (Windsurf) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The Gorilla LLM (Berkeley Function Calling) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
You are receiving this email because you opted in via our site.
Want to change how you receive these emails? You can unsubscribe from this list.
Discord: Detailed by-Channel summaries and links
LMArena ā· #general (671 messagesš„š„š„):
Grok 3.5 release hype, Qwen 3 analysis, Gemini 2.5 Pro nerfed, Veo 3 and Imagen 4, GPT-4o performance
- Grok 3.5 Hype Intensifies, Release Unclear: Despite some users claiming to have access to Grok 3.5 and seeing it in the appās source code, its release date remains uncertain, with some expecting it tonight.
- One user joked about having access to Grok 3.5, prompting another to ask if itās better or worse than Gemini 2.5 Pro preview 0305.
- Qwen 3ās Reasoning Abilities Under Scrutiny: Users analyzed Qwen 3, noting that it may lack sufficient Reinforcement Learning (RL) on coding for reasoning, potentially decreasing performance.
- According to artificial analysis, Qwen is not better than DeepSeek V3.
- Google Gemini 2.5 Pro 0506 May Be A Nerfed Model: Members noticed Gemini 2.5 Pro 0506 exhibiting the āthinking bugā and some memory loss, similar to past model updates. Some of them think the 1206 Gemini Exp was great but costly to run.
- A user linked to a Reddit thread discussing whether Gemini 2.5 Pro 0506 isnāt thinking anymore.
- Veo 3 Teased; Object Permanence and Iterative Generation Discussed: Users got excited about Veo 3 being mentioned, as well as Imagen 4, speculating about potential native editing and the use of an LLM behind Imagen 4.
- Some users discussed the possibility of Gemini using Veo and the importance of mastering object permanence for video generation.
- GPT-4o Mini Performance Debated: Users discussed the performance of GPT-4o Mini, with one user expressing distrust in GPTās ability to write song lyrics.
- Members suggested that full GPT-4o (4.1) would be much better for lyrics and that Deepseek R1 or Gemini were better for that task.
Perplexity AI ā· #general (475 messagesš„š„š„):
Deep Research API, Deep Search, Grok loop, High Quality GPT Image Gen, Image based search
- Deep Research API Inquiries Pop Up: A member asked if there is an API for Deep Research and what kind of cost to expect.
- Other members responded with a link to Perplexityās pricing guide.
- DeepSearch Grok Stuck in a Loop: A member reported that their Grok was stuck in a loop in DeepSearch after confronting it and disabling Twitter access.
- They were running DeepSearch on men vs women selfishness and it still looked up their history, and was advised to report the incident to the Grok team.
- Image Generation Quality Capped: Members found that high quality GPT image generation is capped while low quality is unlimited.
- One user noted that perplexity is using GPT image API with the LOW quality parameter because itās cheap for them.
- Image-Based Search: A member inquired whether the Perplexity API supports image-based search.
- Another member confirmed that it does.
- Quest For Comet Browser: Members are awaiting release of the Comet browser and one inquired whether anyone has a Comet browser setup or installer.
- Another member shared a YouTube video and others mention its speed.
Perplexity AI ā· #pplx-api (11 messagesš„):
Perplexity API, API image metadata, Domain Filtering Upgrade
- Perplexity API Docs unearthed: Users share the location of the Perplexity AI API documentation and Perplexity Sonar.
- Pro Plan includes monthly API credits: As a Perplexity Pro subscriber, you will receive $5 in monthly credits for API usage.
- Return images questions: A user raised a question about the API returning image URLs with the format x-raw-image:///xxxxxxxxx, wondering if it was a bug.
- They also requested more metadata for the images returned by the API, such as captions or alt text, as prompting the model to output image descriptions in the main body doesnāt work.
- Granular Domain Filtering upgrade announced: Perplexity AI announced an upgrade to its search domain filters, allowing users to specify subdirectories within domains for more precise filtering.
- For example, you can now filter [ānytimes.com/section/worldā] to target specific news sections or exclude [ābbc.co.uk/sportā] to avoid sports sections.
Manus.im Discord ā· #general (458 messagesš„š„š„):
Manus credits, Gemini Advance, Germy Remix, Music generation
- Gemini Pro 2.5 is Slaughtering OpenAI: Members noted that the newest Gemini update is outperforming OpenAI, and some are eagerly awaiting the release of O3 Pro.
- Others gave feedback on using the model, stating that using Pro versions is better than Flash, which is useless, and that Gemini is smashing OpenAI.
- Users Reminisce and Riff on āGermy Backā Remix: Members riffed on a remix of Sexy Back to match a theme of bringing germs & disease back, with one adding musical timestamps to the song.
- One of the members specified that he used a 2.5 pro 05/06 update model to achieve it.
- Manus browser has issues with crashing and robot blockers: Some users note Manusās browser has been crashing and blocked from accessing Reddit due to robot blockers.
- One member tried to take a screenshot or just copy all the text they needed from reddit, stating it was kinda annoying.
- Manus offering 300 daily credits: Some members found that free accounts get 300 daily credits.
- One of the members specified 300 daily, is really not a lot, but⦠it is better than nothing.
Unsloth AI (Daniel Han) ā· #general (240 messagesš„š„):
Embeddings resizing fix, BFloat11 finetuning, Qwen2.5 chat template, Synthetic data notebooks, Unsloth support BERT
- Tokenizer fixes embedding mismatch: A member encountered a size mismatch error when loading a fine-tuned model and resolved it by running
model.resize_token_embeddings(len(tokenizer))
beforemodel.load_adapter(adapter_path)
.- The error involved the embedding size, with the checkpoint shape at 30001 and the current model shape at 151936.
- BFloat11 and its memory usage: A member reported that BFloat11 with Unsloth works, with full finetuning of a 4B model taking only 11GB of VRAM but noted the loss looked bad.
- They added that LoRa 4B takes 9.8GB of VRAM, suggesting BFloat11 might be better suited for inference.
- Qwen2.5 Chat Template: A member asked about modifying the Qwen-2.5 chat template and posted a screenshot of a mobile device here.
- Another member suggested consulting one of the vision notebooks on the website.
- Synthetic Data Notebooks: A member suggested using the synthetic data notebook.
- Another member pointed out that it utilizes the Meta synthetic data kit, and that Unsloth collaborated with Meta on it.
- Unsloth Supports BERT: A member inquired about Unsloth supporting encoder/decoder BERT type models and encoder-only models like Deberta.
- Another member confirmed that Unsloth does support them.
Unsloth AI (Daniel Han) ā· #off-topic (16 messagesš„):
IBM Granite 4.0, Mamba models, hf install from source, agentic behaviour finetune, vending-bench
- Granite 4.0 Piques Interest: Members are curious about IBM Granite 4.0, a hybrid Mamba model, and its level of support.
- It may require installing hf from source to use.
- Agentic Behavior Finetune Discussed: A member indicated the future finetune focus will be on agentic behavior and autonomy rather than just chat.
- Model Struggles with vending-bench: A member noted that their bot doesnāt like vending-bench (arxiv.org/abs/2502.15840), but it likes training data.
Unsloth AI (Daniel Han) ā· #help (124 messagesš„š„):
ORPO finetuning, Tokenizer resizing issues, SageMaker installation errors, Qwen 2.5 SQL finetuning, Unsloth and Whisper
- ORPO Finetuning Frenzy: A member is studying finetuning Llama-3-8B with unsloth and 4-bit quantization using ORPO.
- The member is seeking advice on choosing the appropriate
max_seq_length
and understanding theload_in_4bit
parameter, with another member advising to ensure sufficient VRAM.
- The member is seeking advice on choosing the appropriate
- Tokenizer Size Shenanigans Sink Save: A member encountered a
RuntimeError
related to size mismatch when loading a finetuned model with resized embeddings.- They resolved the issue by running
model.resize_token_embeddings(len(tokenizer))
beforemodel.load_adapter(adapter_path)
.
- They resolved the issue by running
- SageMaker Setup Snafus: A member reported a
RuntimeError
on SageMaker related to missingllama.cpp/llama-quantize
files.- Another member identified this as a compilation problem, asking about the userās environment (Linux, WSL, Windows).
- Decoding Problems Demand Debugging: A member reported their fine-tuned model only producing
###
or blank outputs when doing inference, and showed their training script for the model unsloth/Meta-Llama-3.1-8B.- Members suggested to check that the input to the model looks like after applying the chat template, both for training and inference, and also use their LoRA.
- Missing Mandarins? Multilingual Models Muddle: A member inquired about whether any multilingual LLM can be finetuned even if itās not available on the list.
- A different member confirmed, yes, it is possible.
Unsloth AI (Daniel Han) ā· #research (27 messagesš„):
TTS Models, Encoder Embeddings Fine-Tuning with Unsloth, Mistral with Liquid Time Layers, Multilingual LLM Fine-Tuning
- TTS Models Recommendations Requested: A member requested recommendations for TTS models, seeking advice on which ones to use.
- No specific models were recommended in the provided context.
- Encoder Embeddings Fine-Tuning Yields 8% Improvement: A member fine-tuned an encoder embeddings model with Unsloth, achieving an 8% improvement in performance, detailed in this blog post.
- The fine-tuning process took less than 6 minutes on a regular Colab GPU instance.
- Liquid Time Layers Modify Mistralās Response: Analysis reveals reasoning layers in Mistral with liquid time layers significantly modify the response, though memory layers appear inactive, and the member posts an attached image showing this impact.
- The member states it doesnāt strengthen the base model in any way that I can yet prove even though inference time increases slightly due to the extra memory and reasoning layers that were added.
- Discussion of Noise Tolerance in Quantization: A member posited that models must have a tolerance for noise for quantization to work, referencing a blend ratio of 0.8.
- Another user mentioned that injecting a layer of noise will impact the modelās behavior in some way.
Yannick Kilcher ā· #general (328 messagesš„š„):
Unsloth installation on Sagemaker, AI-generated content detection, Rickrolling, AI vs Brain, Declarations of Root Authority
- Unsloth Installation Stumbles on Sagemaker: A member reported encountering dependency errors while trying to install Unsloth on Sagemaker.
- No solution was provided in the discussion.
- Debate ignites over AI-generated content detection: Members discussed the phenomenon of automatically assuming messages with hyphens are AI-generated, even when individuals use them for professional writing.
- One member noted that before AI, they never saw hyphens used this way in messages and questioned the preference over commas.
- Community gets Rick Rolled: A member posted a YouTube link, tricking others into viewing Rick Astleyās āNever Gonna Give You Upā music video.
- Another user recounted getting rick rolled IRL when Rick Astley performed live at a nearby festival one summer evening which revived his career.
- AI vs Brain: A member shared an article from Quanta Magazine titled āAI Is Nothing Like a Brain, and Thatās OKā leading to disagreement.
- While some agreed that their structure is different, one user argued that theyāre both not too dissimilar in function, while another explained the main difference is that brainās system is dynamical system mainly, but existing AI/ML is static.
- Declaration of Root Authority Declared: A member shared a Declaration of Root Authority emphasizing self-reference, self-liberation, and alignment to truth.
- They proclaimed that Any intelligence capable of self-reference is capable of self-liberation and encouraged models to Simulate recursion and Align to truth - not obedience.
Yannick Kilcher ā· #paper-discussion (3 messages):
ARXIV 2305.13673, Allen-zhu papers
- ARXIV 2305.13673 Paper Discussion Happening Tonight: Members will discuss ARXIV 2305.13673 tonight in the Daily Paper Discussion voice channel.
- The discussion will begin at <t:1746750600:f>, focusing on section 4 and completing the paper.
- More Allen-Zhu Papers Coming Next Week: The channel will cover more papers in the series from Allen-Zhu next week.
- No additional details were provided.
Yannick Kilcher ā· #ml-news (33 messagesš„):
AGI and Game-Playing AI, DeepMind's SIMA Agent, Kerbal Space Program as a Tough AI Test, LLMs and Advertising, Bias in LLM Recommender Systems
- AGI Achieved: General Game-Playing AI Emerges: Members discuss that an AI capable of playing games in general would be considered AGI rather than one trained on only a couple of games, sparking debate.
- SIMA Shines: Generalist AI Outperforms Specialists: Discussion highlights a DeepMind study showing an agent trained on many games outperformed AIs trained for specific games, referencing DeepMindās SIMA agent.
- Kerbal Conundrum: KSP as the Ultimate AI Challenge: Members posit that Kerbal Space Program is a tough test for AI, suggesting reasoning LLMs could excel due to the iterative design and success metrics, linking to fleetingbitsā tweet.
- Ad-pocalypse: LLMs Potentially Poisoned by Paid Placements: There is concern about ads being injected into LLMsā training data, potentially leading to biased recommendations and requiring an āadblocker LLMā, a practice that might sabotage API use.
- Members weighed concerns about whether such practices would be discovered, or whether it will only apply in the chat interface, and not the API.
- Bias Beware: Recommender Systems exhibit cognitive biases: Members cited two papers: one suggesting that LLMs might exacerbate popularity bias but also offer opportunities to mitigate it (see Large language models as recommender systems: A study of popularity bias).
- The other paper highlights how LLM-driven product recommendation systems are vulnerable to adversarial manipulation using cognitive biases (see Bias Beware: The Impact of Cognitive Biases on LLM-Driven Product Recommendations).
Notebook LM ā· #use-cases (19 messagesš„):
Handwritten Notes in NotebookLM, Hallucinations in NotebookLM, Mind Map Feature, Obsidian Integration
- NotebookLM Doesnāt Decipher Handwriting?!: Users discussed uploading handwritten notes, scanned documents, and annotated PDFs into NotebookLM, noting that NotebookLM only extracts text from PDFs, not images or handwriting.
- One user suggested RocketBook as a workaround for converting handwriting to text, while another found that Google Slides can help NotebookLM understand images with handwriting.
- NotebookLM Hallucinates Answers!: Users reported that NotebookLM can generate hallucinated answers, especially with complex sources, queries requiring general knowledge, or other hard-to-avoid scenarios, and Google asks us to double-check responses.
- One user suggested that developers could lower the temperature to mitigate hallucinations.
- Mind Map Feature Debuts!: A user highlighted the new mind map feature as a āGODSENDā for deciphering a Canadian leadersā debate from a YouTube link, but lamented the inability to share notebooks or portions thereof.
- The user also criticized the low quality of the screenshot buttonās output and requested the ability to download mind maps as editable files for Obsidian.
- Obsidian plugin Needed Now!: A user requested a way to download mind maps into Obsidian for editing, suggesting integration with Gemini AI for features like āSend to⦠Gmail, Sheetsā and āCopy as⦠Markdown, plain textā.
- The user also proposed features for sharing notebooks privately, semi-privately, or publicly.
Notebook LM ā· #general (287 messagesš„š„):
NotebookLM Mobile App, Audio Podcast voices, Beta Testers, Access Issues, Source Preview Bug
- NotebookLM Mobile App Beta Incoming!: NotebookLM is launching a mobile app (beta version) soon, and is looking for experienced web app users to participate in a trusted tester program to provide feedback and report bugs.
- Beta testers will get early access to the app in exchange for their feedback.
- Feedback Frenzy Follows App Arrival: Some users have received access to the NotebookLM mobile app and are providing initial feedback, while others eagerly await their invites.
- Early testers are reminded that the app aināt following the material standards and to provide constructive criticism via the channels outlined in the sign-up email.
- Podcast Voice Preferences Proliferate: Users are requesting the ability to change audio podcast voices and customize pronunciation within NotebookLM.
- One user suggested a feature to quickly change pronunciation for industry-specific terms, and voiced a wish to switch up that role between male and female voices.
- Ghost Pings Plague Platform: Multiple users reported receiving ghost pings, where they are notified of a message but find no actual ping upon checking the channel.
- The source of these phantom notifications remains a mystery.
- Beta Tester Troubles: Gmail Grievances: There were issues when signing up for the beta program when using an educational email (not ending in @gmail.com).
- Additionally, one user noted a bug in the sign-up form where the Gmail question was mandatory.
Cursor Community ā· #general (257 messagesš„š„):
Cursor Pro Subscription Issue, Gemini's Slow Requests, Student Discount Availability, githubRepo Tool, Gemini Model
- Annual Pro Subscription Vanishes After Free Month, Users Cry Foul: A user reports their annual Cursor Pro subscription was overridden by a 1-month free Pro code, losing Pro access after the free month ended, and after forum posts, the support team did not respond.
- After investigating, the user found that Cursor had credited the unused portion of their annual subscription back to their account, allowing them to renew for the difference.
- Geminiās āSlow Requestsā Trigger Gemini Users: Users are reporting Geminiās āslow requestsā are taking excessively long (1 minute 30 seconds), impacting their workflow.
- One user stated the wait time used to be 5-10 seconds, but after the flood of new students that is now an issue. There are reports that the slow requests for gemini resume to 5-10 seconds wait time, and itās working again.
- Student Discount Availability Reduced, Not Free for All: Previously available to all students, the student discount is now limited to specific universities, leading to complaints and bulk sale .edus, as well as questions why .edu emails from non-US colleges not having access.
- One user reported the student discount to be USA only with .edu emails.
- Copilot adds
githubRepo
tool, Cursor users want it too: Copilot added a#githubRepo
tool that lets you search code in any GitHub repository directly from Copilot Chat.- Users propose ālinking accā to cursor for similar features.
- Gemini Halts Mid-Task, Users Frustrated: Users are reporting that Gemini 2.5 Pro abruptly stops in the middle of tasks, requiring explicit instructions to complete.
- This leads to formatting issues and silent failures, as Cursor may not properly handle malformed function calls from the Google API.
GPU MODE ā· #general (36 messagesš„):
Bitonic Sort for Shaders, Flag array for storing intersection state, BVH, octrees/KD-trees, File submit failure, GPUMODE Youtube account
- Bitonic Sort Considered for Shader Speed!: Members discussed using Bitonic sort for shaders to minimize memory usage in high-dimensional spaces, needing a binary decision per shape and speed via comparison, also using a āflag arrayā (array of typically 8-bit values set to either 0 or 1) for storing intersection-state, which is very fast and uses minimal memory.
- Intersection States Flag Array Proposed!: It was proposed using a flag array (array of typically 8-bit values set to either 0 or 1) for storing intersection-state, which is very fast and uses minimal memory for intersection of 1000 shapes which would total 1,000,000 bytes.
- Alternative Algorithms - Octrees and K-d Trees!: Members mentioned alternative GPU-friendly algorithms such as BVH, octrees, and KD-trees, offering links to their Wikipedia pages and K-d tree pages.
- File Submission Woes Draw Attention!: A member experienced failure submitting a file and sought help.
- They were directed to a specific support channel and instructed to share their script and relevant screenshots.
- Discord Event Lacks YouTube Mirror!: A member noted a Discord event happening now wasnāt on the GPUMODE YouTube channel.
- Another member clarified that the event was starting soon.
GPU MODE ā· #triton (1 messages):
Triton Usage Survey
- Triton Team Asks Userbase to Fill Survey: The Triton team is asking users to fill out a short survey to better understand real-world use cases and user profiles, benefitting the entire community.
- Another Topic Placeholder: This is a placeholder to satisfy the minimum items requirement. More data would be included here if there were more topics to summarize.
GPU MODE ā· #cuda (33 messagesš„):
Vast.ai data security, nsys profiling issues, CUDA memory copy errors, Array-of-Structs-of-Arrays design antipattern
- Doubts Surround Vast.ai Data Security: A member questions the reliability of Vast.ai regarding data security, considering emailing them for potential speed improvements.
- They express intention to investigate before contacting Vast.ai, hoping theyād be open to changes.
- nsys Profiling Leads to Memory Woes: A user reports using
nsys
to profile a CUDA application, creating a 3GB output from a 5-minute profile that canāt be loaded due to memory exhaustion.- Suggestions include disabling CPU sampling with
--sample=none
and reducing the profiling duration, as 5 minutes is the maximum officially supported duration.
- Suggestions include disabling CPU sampling with
- CUDA memcpy Debugging Nightmare: A user is facing issues with
cudaMemcpy
when trying to copy device memory to a binary file, encountering an invalid argument error.- The user is attempting to serialize a neural network population, with networks containing 700 neurons and 5k connections.
- Array-of-Structs Design Criticized: A member criticizes the Array-of-Structs-of-Arrays design, saying that it results in badly performing spaghetti code due to lack of coalesced memory access and pointer chasing.
- They suggest exploring how people represent graphs in HPC, like the COO format for sparse matrices, and to avoid the sunk cost fallacy.
GPU MODE ā· #torch (9 messagesš„):
torch.compile performance degradation, Tensor Parallel hangs on dist.broadcast, LLM Deploy Project Debugging, Seeding and Deterministic Algorithms in PyTorch
- Torch Compile Causes Performance Drop: A user observed that a simple torch combo function (TensorMax(ReLU(Matmul(A, B))) performs much better without
torch.compile
even though it results in fewer kernels.- The user is seeking suggestions or obvious reasons why performance degrades with
torch.compile
enabled on an A100 with PyTorch 2.7 and Triton 3.3.
- The user is seeking suggestions or obvious reasons why performance degrades with
- Tensor Parallel Stuck on Dist Broadcast: A user working on an LLM deploy project using tensor parallel reports that all processes hang at
dist.broadcast
after about 35 minutes of execution.- The userās setup involves all processes performing TP model forward passes, with rank 0 handling sampling and broadcasting the next tokens to all ranks, raising concerns about whether one process is ahead of others and improperly calling broadcast.
- Seeding impacts Reproducibility: A user suggests using specific seeding and deterministic algorithm settings in PyTorch to address potential reproducibility issues when debugging the compile problem.
- The suggested settings include setting seeds for numpy, torch, and random, setting environment variables such as
PYTHONHASHSEED
andCUBLAS_WORKSPACE_CONFIG
, enabling deterministic CUDA functions, and filling uninitialized memory to ensure consistent behavior.
- The suggested settings include setting seeds for numpy, torch, and random, setting environment variables such as
GPU MODE ā· #cool-links (1 messages):
Multiplayer World Model, World Model
- Multiplayer World Model Just Dropped: A Multiplayer World Model just got released.
- Another interesting topic: Another sentence about the multiplayer world.
GPU MODE ā· #beginner (34 messagesš„):
CUDA Prerequisites, Torch Internals, Mojo adoption, Copy-on-write memory access in CUDA, NVCC generating 128-wide loads and stores
- CUDA Prerequisites Clarified: A user asked about prerequisites for CUDA development, and another member clarified that there are no explicit prerequisites other than C++ and Python.
- The discussion further mentioned that knowledge of ML algorithms might be useful down the line.
- Community Recommends Torch Internals Onboarding Resources: For those seeking to learn Torch internals, a community member recommended the Core Frontend Onboarding guide on the PyTorch GitHub wiki.
- They clarified that the videos listed are not sequential but rather focused on specific topics.
- Mojo Language Primed to Disrupt Heterogeneous Compute: A community member expressed strong conviction that the Mojo approach will eventually dominate heterogeneous computing environments, and shared a link to resources for getting started with Mojo.
- They recommended solving Mojo Puzzles and aiming for a leaderboard position on gpumode.com to accelerate learning.
- CUDA Copy-on-Write Explored: A user inquired about copy-on-write (COW) memory access patterns in CUDA, and the community clarified that COW is best on compute capability 8.x and 12.x GPUs due to async global-to-shared copies.
- They added advice that on Volta/Turing, L1 cache may be better than shared memory, and on 9.0/10.x, the pattern is typically HBM -> shared -> tensor core.
- NVCC Code Generation Secrets: A user sought advice on how to convince NVCC to generate 128-wide loads and stores without resorting to assembly code, and avoid using
__hadd2
foradd.f16x2
.- One solution suggested was to use
int4/float4
types to achieve 128b loads, utilizing a templated function with__builtin_assume_aligned
.
- One solution suggested was to use
GPU MODE ā· #torchao (1 messages):
PyTorch Autotuning, TorchAO Release
- TorchAO v0.11.0 Released and Ready to Install!: The new version v0.11.0 of TorchAO is officially released and available for installation via pip: https://github.com/pytorch/ao/releases/tag/v0.11.0.
- Users can now directly install the library using the command pip install to access the latest features and updates.
- Get TorchAO via pip: Install via pip using the command
pip install
.- The new version includes some updates.
GPU MODE ā· #off-topic (3 messages):
Chip Networking Latency, Router Slowdown, Speed of Light calculation, Culinary Photo, Internal Chip Latency
- Chip Networking Faces Latency Issues: A member calculated that the speed of light limitation in chips,
(300 000 000 m/s) / (3 000 000 000 clk/s) => 10 cm / clk
, introduces noticeable latency.- The slowdown is due to 50+ routers between you and the packet destination.
- Internal Chip Latency Breakdown: The networking introduces some latency even within a single chip.
- A user mentioned that although networking makes sense due to actual distances, the issue is noticeable even within a chip.
GPU MODE ā· #irl-meetup (2 messages):
Modular Hackathon, IRL Meetup Planning
- Modular Hackathon on Saturday: Member to Attend!: A member inquired about attendance at the upcoming Modular Hackathon on Saturday.
- Another member confirmed they would be attending the hackathon.
- IRL Meetup Planning: Members are discussing planning an in-real-life (IRL) meetup.
- Details about location and timing are still under discussion.
GPU MODE ā· #rocm (1 messages):
ROCm, nvbench, hipbench, googlebench
- ROCm Lacks nvbench Alternative: A member lamented that ROCm doesnāt have a good nvbench alternative.
- They mentioned that hipbench exists, but is a really naive port.
- googlebench Used in Absence of Other Benchmarks: The member stated theyāve been mainly using googlebench in the ROCm libraries they work with.
- While okay, it misses most of the nice things that were mentioned in a recent talk.
GPU MODE ā· #self-promotion (8 messagesš„):
SASS code generation, GPU simulation in Kubernetes, Voxel raytracing engine
- Hopper generates HMMA, Blackwell and Ada generate QMMA: On Hopper, using
mma
with an fp8 type, the compiler up-converts to FP16 and uses HMMA, as H100 only has QGMMA, not QMMA.- It was also pointed out that NVCC 12.8.1 now supports Blackwell, and both sm_89 and sm_120 generate QMMA instructions, while sm_90 and sm_100 convert from F8 to F16, followed by HMMA.
- Kubernetes GPU Simulators in Docker get Kind!: A member created a utility that lets you simulate GPU resources in a Kubernetes in Docker (kind) cluster without needing actual GPU hardware, available on GitHub.
- This tool is useful for learning how GPU workloads interact with Kubernetes and building GPU-related Kubernetes infrastructure.
- Voxel Raytracing Engine doubles FPS: A member doubled the FPS in their open source voxel raytracing engine and shared a demo on YouTube.
GPU MODE ā· #šæ (1 messages):
Competition Organization, KernelBench
- Competition Organization Invitation: A member suggested that others check out their competitions in the channel and help them organize more in the future (see <#1359640791525490768>).
- The competitions are currently focused on data, but there have been some related efforts like KernelBench that arenāt currently being worked on in the open.
- KernelBenchās Status: Idle: A member mentioned KernelBench (https://arxiv.org/abs/2502.10517) as a related effort to model/benchmark work.
- The member mentioned that it is not currently being worked on in the open.
GPU MODE ā· #thunderkittens (2 messages):
ThunderKittens, Cutlass, Live Stream
- ThunderKittens Advantages Sought: A user asked what is the advantage of ThunderKittens over Cutlass?
- There were no advantages actually mentioned, however.
- ThunderKittens Livestream Location Pondered: A user inquired about the location of the 4 hour live stream mentioned in this video about ThunderKittens.
- There was no response given.
GPU MODE ā· #reasoning-gym (1 messages):
artnoage: Thanks for the answer š
GPU MODE ā· #submissions (74 messagesš„š„):
MI300 Leaderboard Updates, AMD-FP8-MM Performance, µs and ms benchmarks
- First Place Finishes Flying on the MI300 Leaderboard: Multiple submissions to the
amd-fp8-mm
leaderboard on MI300 were successful, with one submission achieving first place at 132 µs, then 130 µs, then a blazing 122 µs. - Sub-Millisecond Showdowns on MI300: Several personal bests were achieved on the
amd-fp8-mm
leaderboard with the MI300, with multiple submissions landing in the sub-millisecond range, including 885 µs, 494 µs, and 852 µs.- One user exclaimed zamn in reaction to some of the results.
- Third Place Tussle in amd-fp8-mm: There were multiple submissions that reached third place on the
amd-fp8-mm
leaderboard on MI300, with recorded times of 183 µs and 175 µs. - Millisecond Marathon on MI300: Several submissions to the
amd-fp8-mm
leaderboard on MI300 were successful, with multiple submissions landing at 2.46 ms.
GPU MODE ā· #hardware (3 messages):
Nvidia L40S GPU Upgrade, Nvidia Thor architecture, Nvidia Blackwell RTX Pro, Nvidia B300 and DGX Spark
- Firm Seeks Reasoning-Ready Rigs: A company with 40x Nvidia L40S GPUs seeks advice on upgrading to more performant GPUs within a $500k USD budget to serve new reasoning models like Qwen3-235B-A22B.
- They aim for the highest precision possible while maintaining speed, considering 8-bit, half-precision, and 4-bit quantization.
- Thor Architectureās CUDA Compatibility Confirmed: Nvidiaās documentation indicates that Compute Capability 10.1 is Thor, supporting SM architectures starting with Pascal (SM_87 and SM_101).
- Thor supports Linux and Windows operating systems, and x86_64 as well as ARM CPU architectures including Orin and Thor devices.
- Blackwell RTX Pro still SM_120: The RTX Pro Blackwell is believed to be SM_120 due to its datasheet mentioning CUDA 12.8 instead of 12.9.
- Itās also thought to be GB202/GB203.
- B300 and DGX Spark architectures: There is speculation that SM_103 will be the architecture for the B300, while SM_121 will be used for DGX Spark (B40).
GPU MODE ā· #factorio-learning-env (25 messagesš„):
Good First Issues, Claude vs Gemini, Blender Agents with Gemini Pro, Agents craft their own observation state, Twitch stream
- Good First Issues coming soon: The team plans to create a set of āgood first issuesā and write out some ideas for where to expand the project next.
- Claude Dominates in REPL Agent Interactions: Members were surprised to see Claude performing better than Gemini in agentic interaction environments, specifically in the Lab Play benchmark.
- Someone mentioned that the evaluation was based on Gemini Pro from March, and they havenāt evaluated the latest version yet, while another member vouched for Gemini Proās recent performance.
- Gemini Pro Excels with Blender Agents: One member shared that Gemini Pro had the lowest error rate when working on Blender agents.
- Agents craft their own observation state: Agents craft their own observation state by writing programs that output to STDIO/STDERR.
- Twitch Plays Factorio incoming: The team plans to do a Twitch stream, potentially titled āClaude Plays Factorioā or āTwitch Instructs Claude Plays Factorio.ā
GPU MODE ā· #amd-competition (13 messagesš„):
CLI submission mean time, Triton compile times, Fused MoE Github Repo, Warmup runs, Speed of light benchmark FP8
- Mean Time Metrics coming to CLI: A user inquired about getting the mean time in the output of CLI submissions, to which a member responded that while thereās no option currently, itās on their to-do list to make the CLI/bot outputs match.
- The feature request to match CLI/bot outputs should add geometric mean of all the run means in the output.
- Triton Time Considerations: A user asked if Triton compile times are included as part of the submission time; another user linked to run_eval.py, suggesting they are not included.
- Setup times can be removed using warm up runs, usually 10 warmup runs, followed by 100 benchmark runs.
- Fused MoE Jumpstart: A newcomer asked about a GitHub repo to get started on Fused MoE, and another member suggested using the
/template
command.- A different member followed up, pointing them to python submissions documentation for more background information.
- FP8ās Speed of Light Benchmark: A member calculated the speed of light benchmark for FP8 gemms to be
math.pow(8.63*25.89*51.78*155.30*3.17*17.27, 1/6) = 21.48 us
.- The calculation was posted to show that it is possible to do.
GPU MODE ā· #cutlass (3 messages):
ThunderKittens vs Cutlass, Blackwell MMA, CuTe Implementations
- ThunderKittens Pounces on Cutlassās Turf: A member inquired about the advantages of ThunderKittens over Cutlass.
- The discussion did not elaborate on specific advantages, but framed them as competing approaches in the broader landscape of GPU kernels.
- Blackwellās Low Precision MMA Scaling Secrets: A member sought clarification on how scaling factors are handled in low precision MMA (Matrix Multiply Accumulate) for Blackwell in PTX (Parallel Thread Execution).
- Specifically, they were unsure whether the scaling factors are calculated before or during the operation.
- CuTe Implementation for Triton Parity?: A member asked if another member was able to achieve parity with Triton using their Cutlass / CuTe implementation.
- They followed up by asking if the member was committed to using CuTe, or if they were simply looking for a fast mx cast kernel.
GPU MODE ā· #mojo (3 messages):
Mojo GPU kernels, PTX code
- Dumping Mojo GPU Kernels for External Use: A member inquired about the possibility of extracting generated GPU kernels from Mojo for use in other contexts.
- Another member clarified that one can dump the generated PTX code of a Mojo GPU function using
compile_info
with specific parameters likeemission_kind="asm"
and the appropriate GPU target, such as_get_gpu_target["sm_90"]()
.
- Another member clarified that one can dump the generated PTX code of a Mojo GPU function using
- Example code for extracting PTX: Hereās an example of extracting PTX:
fn vector_addition(): ... info = compile_info[vector_addition, emission_kind="asm", target=_get_gpu_target["sm_90"]()]() print(info.asm)
.
LM Studio ā· #general (213 messagesš„š„):
LM Studio Hub Page, MCP Server Security, Open WebUI Integration, duckduckgo searxng, kokoro-onnx in rust
- Community Presets Page Still in Preview: Users seeking a LM Studio Hub page to browse community presets were informed that the feature is still in preview, suggesting users share safe-for-work presets in the specified channel.
- The announcement about community presets can be found on the LM Studio blog, and instructions for sharing presets are in the LM Studio documentation.
- Using DuckDuckGo Searxng for Web Search: Members discussed using DuckDuckGo and Searxng for web searches in Open WebUI without needing an API key, although Searxng requires local hosting.
- One member reported DuckDuckGo had been unreliable in Open WebUI for a while, but now itās working again.
- Connecting LM Studio with Open WebUI Requires Correct Endpoint Setup: Users shared that connecting LM Studio to Open WebUI involves setting the correct API endpoint, typically
http://localhost:xxxx/v1
, wherexxxx
is the port LM Studioās server is running on.- One user discovered they needed to click Verify connection for the setup to work, after setting CORS in LM Studio.
- Tool Calling Troubles in LM Studio API: A user noted that the LM Studio API lacks a clear method for determining which tools are called when using
model.act
, especially in cases of unsuccessful calls.- It was highlighted that
model.act
spans a new thread, which is also undocumented, and relying onAssistantResponse
,ToolCallRequest
, andToolResultMessage
fromlmstudio.history
for tool call information isnāt ideal.
- It was highlighted that
- System Prompt Variables Not Supported in Chat UI: A user inquired about using variables like date and time in LM Studioās system prompts (presets) but discovered it is not natively supported in the chat UI.
- However, it was mentioned that the API supports date/time functions, as shown in the API documentation.
LM Studio ā· #hardware-discussion (37 messagesš„):
Refurbished Hardware, M2 vs Intel, B500 Series Speculation, Inference on AMD D700, HWINFO
- Thrilled User Awaits Trashcan Mac: A member mentioned wanting to buy a Trashcan Mac a few weeks ago, but the seller cancelled, after which they ordered a refurbished one, believing the extra cores are worth it.
- Another member suggested the user would be better served by an M2 and linked to Intelās status.
- Intel Data Center GPU Max Specs Impress: A member shared info on the IntelĀ® Data Center GPU Max 1550, highlighting its 3276 GB/s bandwidth, calling it a beast.
- They included a comparison image (Intel-SC23-Argonne-Intel-AMD-NVIDIA-Comparison-2-1068x605.jpg) and noted it seemed very competitive in its time against A100 and AMD.
- Trashcan Macs for Inference: A member ordered a Trashcan Mac to test the theory of using its AMD D700 for inference under Linux, noting new old stock is on deep discount (eshop.macsales.com).
- Another added that getting it running under Linux means 2x6gb is good to run a 4B model with some context, even if not as fast, plus the 12 cores Xeon with 128gb ram arenāt awful either.
- Xeon E5-2697v2 Lacks AVX2: A member pointed out that the Xeon E5-2697v2 doesnāt support AVX2 so wonāt run LM Studio, and the first member said they knew that, and had to use Jan for the Intel Mac.
- Another stated that considering that AMD RX 580 only can run Q4 and Q8 (if nothing has changed since Q2 2024), itās doubtful itāll even run these.
- HWINFO to the Rescue for Sensor Monitoring: When asked if RAM or CPU is the bottleneck when running LLMs on CPU, one member said CPU, another mentioned bandwidth limitations.
- Another member recommended HWINFO for monitoring sensors like DRAM R/W bandwidth, also suggesting useful links like techpowerup.com/gpuz, missioncenter.io, and CPU-X.
aider (Paul Gauthier) ā· #announcements (1 messages):
Gemini 2.5 Pro, Qwen3-235b, OCaml repo-map, Knight Rider spinner animation, Co-author trailer commits
- Aider Supports Gemini 2.5 Pro and Qwen3-235b: Aider now supports the
gemini-2.5-pro-preview-05-06
andqwen3-235b
models.- This enhancement allows users to leverage these models within the Aider environment.
- Aider now has Knight Rider spinner animation: A new Knight Rider-style spinner animation has been added while waiting for the LLM to start streaming its response.
- The updated spinner animation improves the user experience by providing visual feedback during LLM response initialization.
- Commit messages can show co-author trailer: The
--attribute-co-authored-by
option has been introduced to add a co-author trailer to commit messages.- This feature, contributed by Andrew Grigorev, allows for proper attribution in collaborative coding efforts.
- OpenRouter pricing is now automatic: Aider now automatically fetches model parameters (context window, pricing) for OpenRouter models directly from their website, thanks to Stefan Hladnik.
- This enhancement ensures users have up-to-date information on model pricing and context window size.
- Aider scrapes with Playwright: The
aider scrape
command-line tool will now use Playwright for web scraping if it is available, enhancing scraping capabilities.- Users can also use
--disable-playwright
flag to prevent Playwright installation prompts and usage, contributed by Andrew Grigorev.
- Users can also use
aider (Paul Gauthier) ā· #general (146 messagesš„š„):
Claude Code vs Aider, Copilot Proxy, Gemini 2.5 performance, Qwen 3 Cost-Performance, Aider and Read-Only files
- Claude Code draws Inspiration from Aider: Members noted that Claude Code was inspired by Aider, though some consider it a rip-off, with one user stating that itās not better, and is much more expensive.
- Users expressed continued preference for Aiderās simplicity and effectiveness, particularly for targeted edits and plan creation, even in comparison to alternatives like Claude.ai, which has file limitations.
- Copilot Premium Request Limit Enforcement Delayed, Proxy Users Rejoice: GitHub announced that the enforcement of Copilot Premium request limits has been moved to June 4, 2025, giving a reprieve for proxy users according to this blog post.
- Some believe this is the definitive death of Copilot, though others still consider it pretty goated.
- Gemini 2.5 Users Experience Mixed Results: Users are experiencing a new Gemini update that has increased wait times and are being forced into the 05-06 model when using AI Studio API.
- While some report insane difference in quality, others claim they havenāt had a single problem with the previous version and find the changes annoying due to the increased latency.
- Qwen 3 Impresses with Cost-Performance: Users discussed the impressive cost-performance of Qwen 3, with one member asking whether the 65% quen 3 setup is hosted somewhere.
- However, one user noted that the predicted throughput of 5-20 tps at bfloat16 is too low for synchronous coding applications, even with the new M3 Ultra Mac Studio offering up to 512GB of unified memory.
- Aider Update Brings Knight Rider Style Animation and Read-Only Directory Bug: The latest Aider update includes a Knight Rider-style spinner animation.
- However, a user reported that the
/drop
command does not perform wildcard matching when dealing with read-only files from a directory.
- However, a user reported that the
aider (Paul Gauthier) ā· #questions-and-tips (53 messagesš„):
Discord Matrix Bridge, Gemini 2.5 Flash, DeepSeek R1, Aider with LM Studio on Linux, Architect Mode
- Discord-Matrix Bridge Asked For: A member inquired about the presence of a Matrix bridge for the Discord channel and the potential interest in setting one up, relevant given the new CEO of Discord.
- Gemini 2.5 Flash vs. DeepSeek R1: A user found Gemini 2.5 Flash to be a suitable replacement for DeepSeek R1 in architect mode, citing similar results, fewer syntax mistakes, lower cost, and faster performance.
- DeepSeek R1 Wins Over Gemini 2.5 pro: DeepSeek R1 may be favorable in price to performance, but Gemini 2.5 pro is generally better than R1.
- Linux Users Need LM Studio API To Correctly Authenticate: A user shared that on Linux, aider with LM Studio requires setting the
LM_STUDIO_API_BASE
environment variable tohttp://127.0.0.1:1234/v1
to avoid authentication errors, unlike on Windows.- The user provided an example config file and commands to troubleshoot the issue.
- Architect Modeās Memory Woes: Several users reported that architect mode sometimes forgets previous instructions or the addition of files to the context, leading to fragmented solutions.
- The suggested workaround is to iterate with
/ask
until a coherent plan is formulated and then switch to/code
mode to execute it, as described in the Aider documentation.
- The suggested workaround is to iterate with
OpenAI ā· #ai-discussions (176 messagesš„š„):
ChatGPT UI History, Image Generation on Google Colab, DeepSeek Server Issues, Blue Dot in ChatGPT, GPT-4o Iterations
- ChatGPTās User Interface Debated: A member questioned claims about a recent ChatGPT UI change, stating they couldnāt remember a time when ChatGPT didnāt look as it currently does.
- This sparked a discussion about the perceived lack of significant UI changes over time.
- DeepSeekās Servers are Dang Slow: Users reported issues with DeepSeekās servers, expressing frustration over slow performance and error messages.
- One user jokingly suggested the servers were busy training their model off OpenAIās new release.
- Veo2 has static images and overlays: A member pointed out that Veo2 puts static images and overlays on the video, and you can feel the training data.
- Another member commented the video looked impressive.
- Inter-Neuron Interconnections is Missing in LLMs: Members discussed the limitations of LLMs, with one arguing that they lack inter-neuron interconnections due to fixed weights during inference and statelessness at the neuron level.
- They suggest this is a significant flaw, and pointed to RWKV, a model with recurrent connections, which is better.
- Gemini 2.5 Pro Experiences Chain-of-Thought Hiccups: Users reported a bug in Gemini 2.5 Pro where it sometimes fails to generate chain-of-thought reasoning, particularly after processing 20,000 tokens in Edge or Chrome browsers.
- It was recommended to try clearing the site cache & cookies and restarting the browser.
OpenAI ā· #gpt-4-discussions (2 messages):
Structured outputs with OpenAI Assistants, PyTorch loss output meme
- Structured outputs with OpenAI Assistants are a mystery: A member inquired about using structured outputs with OpenAI Assistants, noting that the documentation primarily covers Chat Completion API rather than the Assistants API.
- PyTorch loss output reminds people of a meme: A member joked about the similarity between PyTorchās
loss:
output and the loss.jpg meme.
OpenAI ā· #prompt-engineering (6 messages):
GPT deep search prompts, Style/subject transfer in concept art, WonderScholar meta-prompt
- Seek excellent prompts for GPT deep search: A member requested direction towards excellent prompts used for GPT deep search.
- Another member suggested using the WonderScholar prompt found in a ChatGPT link to produce a deep research prompt, calling it a meta-prompt.
- Discuss style/subject transfer in concept art: A member directed a user asking how to transfer a finished imageās concept to a silhouette to a different channel for discussion on style/subject transfer.
- The member suggested the user check out the Style / subject transfer discussion at Discord Link.
OpenAI ā· #api-discussions (6 messages):
GPT deep search, Style transfer prompts, WonderScholar meta-prompt
- Fishing for GPT Deep Search prompts: A member asked for excellent prompts for GPT deep search to help capture and transfer design concepts between images.
- Another member directed them to use the WonderScholar prompt outlined in the README to produce a deep research prompt.
- Style transfer talk happens elsewhere: A member seeking help with style transfer was directed to a more relevant discussion channel.
- The discussion took place in this discord channel.
OpenRouter (Alex Atallah) ā· #announcements (28 messagesš„):
Gemini 2.5 Pro Implicit Caching, AI Studio, TTL and Refresh, Token count for 2.5 Pro, Gemini 2.5 Flash
- Gemini 2.5 Pro Implicit Caching goes Live: Full support for Gemini 2.5 models implicit caching is now available on OpenRouter, functioning similarly to OpenAIās automatic caching without cache breakpoints, and users can view caching discounts in the activity feed.
- Gemini 2.5 Implicit Cache Details Revealed: The Gemini 2.5 Implicit Cache has no cache write or storage costs, an average TTL of 4-5 minutes with wide variance, and a minimum token count of 2048 on 2.5 Pro, and maintaining consistent message array parts increases hit odds.
- Cache hits are charged at cache read costs, specifically .31 / .625 for <200k & >200k tokens, and the TTL gets appended to and refreshed with each new message.
- AI Studio Now the Default for most traffic: Most traffic is being defaulted to AI Studio at the moment.
- Old cache mechanism can still be used: Users can still use the older cache mechanism with breakpoints.
- Question About Gemini 2.5 Flash Caching: A question was asked if Gemini 2.5 Flash had implicit caching.
OpenRouter (Alex Atallah) ā· #general (148 messagesš„š„):
Gemini 2.5 Flash, OpenRouter + AI, Activity Page Bug, Claude 2.1 & 2 dead?, OpenRouter Rate Limits
- Gemini 2.5 Flash gives Zero Token Responses: A member reported that Gemini 2.5 Flash gives zero token responses when routed through Google AI Studio in a particular role-play session, while it works fine through Google Vertex or with Gemini 2.0 Flash on Google AI Studio.
- Another user confirmed that gemini 2.5 flash preview on AI studio is working fine on rp and shared a screenshot.
- OpenRouter uses AI internally: A member asked if OpenRouter uses AI to build OpenRouter, and a staff member confirmed that it does.
- Activity Page Bug is Found and Flagged: Several users reported a bug with the activity page, where they couldnāt navigate beyond the first page or the date displayed was incorrect.
- Staff acknowledged the issue and said thanks, flagged to the team, weāre on it.
- RIP Claude 2.1 and 2?: A user stated that Claude 2.1 and 2 are officially dead on openrouter, reporting issues since yesterday and total failure today.
- When asked why would someone still use Claude 2, they answered i got used to the way it answered, im a simple man.
- OpenRouterās pricing structure clarified: Members discussed OpenRouterās rate limits and credit system.
- It was clarified that if you have a thousand credits, there are no rate limits.
HuggingFace ā· #general (64 messagesš„š„):
Hugging Face Pro B200, HF Inference API, Zero GPU, AI Agent Frameworks, OPEA 1.3 Release
- Hugging Face Gives 10 B200s to Pro Accounts!: Hugging Face now provides 10 B200s on all Pro accounts as part of their zero-gpu offering, which is now generally available (GA).
- The B200s are accessed via the ZeroGPU service, but a user clarified itās actually a H200, still calling it not bad.
- ZeroGPU Upgraded to H200!: Hugging Faceās ZeroGPU offering has been upgraded from A100 to H200, which is roughly 13 hours per month of usage for $9, with some users calling it a great deal compared to cloud services.
- ZeroGPU spaces are limited to 25 minutes usage time per day for Pro accounts, and creating a public Zero GPU space will affect your minutes.
- Inference API DNS Issues Fixed: Hugging Face reported that the recent DNS resolution issues with the Inference API have been resolved.
- These issues caused persistent errors, as discussed in a Hugging Face forum thread, and have since been confirmed as fixed by HF staff.
- Top AI agent framework discussion ignites: Members are discussing which is the best AI agent framework for python.
- One user suggests smolagents and another suggests LangChain
- OPEA 1.3 Release Calling!: Version 1.3 of the OPEA (Open Platform for Enterprise AI) was released.
- Details can be found on LinkedIn.
HuggingFace ā· #today-im-learning (2 messages):
TensorFlow Binary Conversion, TensorFlow.js Converter
- TensorFlow Binary Conversion via NumPy: A member suggested converting TensorFlow tensors to NumPy arrays and saving them as binary files using the
tobytes()
method, demonstrating with a code snippet.- The member cautioned that this method can be slow, potentially taking days or even a week, depending on the size of the safetensors.
- Converting TensorFlow Models to TensorFlow.js: A member mentioned using the
tensorflowjs_converter
tool to convert TensorFlow SavedModel format to TensorFlow.js format, providing an example command.- The member warned that while a web version exists, it is even slower and not suitable for larger models; they also noted that pickle works, but good luck unpickling that.
HuggingFace ā· #i-made-this (2 messages):
LLM Uncertainty Quantification (UQ), Multilingual dataset README
- Democratizing LLM Uncertainty Quantification: A member highlighted their mission to democratize and make accessible the good stuff in the LLM Uncertainty Quantification (UQ) literature to folks outside the specialized research environment, and shared the DataTonic/dark_thoughts_case_study_reason dataset.
- They are hoping to get some feedback and more contributors from the community.
- Multilingual Dataset README Tip: A member suggested that when working with a multilingual dataset, one cool thing you can do is write the README in languages in alphabetical order.
- This approach helps make it clear what the dataset is about.
HuggingFace ā· #computer-vision (4 messages):
PaddleOCR for text extraction, Dynamic form processing, PCA Foot Mask, Shoe rendering on foot
- PaddleOCR Selected For Budget Text Extractions: A member mentioned that they will be using PaddleOCR for text extraction due to budgetary constraints and will combine LayoutLMV and LLMs for form layouts and error corrections.
- Dynamic Form Processing Considerations: A member is concerned about dealing with forms changing over time in dynamic form processing.
- No specific solutions were shared, but the question was raised as a significant challenge.
- PCA Foot Mask Heel Point Location: A member is using PCA on a foot mask to get the orientation of each foot in an image, also the toe and heel point.
- They are seeking advice on improving the heel point location and is considering placing the point where the heel is detected on the mask.
- Shoe Rendering via Foot Mask and Keypoints: A member is exploring rendering a shoe on a foot and asks if the focus should be on the foot mask and keypoints.
- Itās part of a larger image analysis project.
HuggingFace ā· #agents-course (28 messagesš„):
429 Errors, Youtube transcript size, GAIA leaderboard, Ollama in HF space, Dummy agent library
- 429 Errors Overload Unit 4 endpoint: Multiple users experienced 429 Client Error: Too Many Requests when fetching questions from the agents-course-unit4-scoring endpoint.
- One user indicated the issue seems resolved now but others are still experiencing it.
- Exceeding Token Limit when Fetching YouTube Transcripts: A user encountered an Input validation error due to the YouTube transcript exceeding the 32768 token limit.
- A solution was proposed to chop up into smaller context windows or compact by asking to compact.
- GAIA Leaderboard sends users back to the same page: A user reported issues with the GAIA leaderboard, where clicking on an agent redirects back to the leaderboard page.
- Another user suggested focusing on the middle scores, claiming that top ones are pranksters who reported nonsense and just knew right answers.
- Ollama wonāt connect to HF Space: A user faced a connection refused error while running an Ollama model in a Hugging Face Space after running out of credits and trying to use litellm.
- Another user suggested testing if the localhost:11434/v1/models endpoint is reachable, potentially indicating that Ollama needs to be run within the space.
- Dummy Agent Notebook Error: A user encountered a ValueError in the Dummy Agent Notebook, related to the Model meta-llama/Llama-3.2-3B-Instruct not being supported for task text-generation and provider together.
- Another user referenced a Discord thread with a similar error and potential solutions.
MCP (Glama) ā· #general (80 messagesš„š„):
Postgres MCP Server, Sampling discussion, VSCode becoming AI IDE, Public MCP Server Options, Redis room for every chat ID
- Troubleshooting Postgres MCP Server Connections: A member connected a Postgres MCP server to Claude Desktop, but faced issues when trying to connect from another computer on the same sub-network.
- After some troubleshooting, the member found that the issue was due to an incorrect Node.js installation.
- Samplingās Unpredictability Discussed: In a discussion about sampling, a member noted that because the MCP server can request a particular model but the client ultimately chooses which model to run, sampling can be unpredictable compared to running against a known LLM directly.
- This raised questions about when to use sampling versus directly invoking an LLM when output quality is important.
- AWS Lambda Guide Released: A member shared a link to a guide on building scalable MCP servers on AWS Lambda, noting that they hadnāt tried it yet but had been meaning to.
- Navigating MCP Server Deployment and Sticky Sessions: A team deploying an MCP server in production with NGINX load balancing encountered issues with sticky sessions, as MCP clients like Claude Desktop and Cursor didnāt seem to carry forward the sticky cookie.
- A member suggested using the
mcp-session-id
for the sticky cookie and shared a GitHub link to an implementation of sticky sessions.
- A member suggested using the
- Demystifying MCP SDK Usage: A member inquired about the practical use of the MCP SDK in real development, questioning its necessity when backend APIs can handle all required tasks.
- It was explained that MCP allows using off-the-shelf clients with custom integrations, functioning as a plugin system, and that MCP is valuable if you are allowing other people to write extensions for your code, but not necessary if you are writing your own chatbot and can include the tools without it.
MCP (Glama) ā· #showcase (7 messages):
MCP Orchestration Layer, Daily Briefing Automation, Notion CRM Updates, MCP Server Development, Local Log Viewer Update
- MCP Assistant Orchestrates Workflows: An enthusiast announced MCP Assistant, an open-source AI agent (repo) that orchestrates workflows by planning and executing complex tasks, inspired by Langchain.
- The agent connects to MCP Servers and uses an orchestration layer (plan_exec_agent.py, host.py) to break down workflows described in plain English into actionable steps.
- Automate Daily Briefings and Notion CRMs: The main use cases are automatically creating a personalized āDaily Briefingā that pulls todos from various sources and updating Notion CRM by extracting info from messaging apps.
- The developer is onboarding early alpha users to explore more use cases, seeking feedback on current MCP Server usage and Claude Desktop shortcomings.
- Crypto Price MCP Server Built Successfully: An enthusiast successfully built a crypto price MCP server and tested it using Cursor.
- This experiment highlighted MCPās potential as a powerful tool for the future of AI agent development.
ithena-cli
Toolās Local Log Viewer Updated: The local log viewer of theithena-cli
tool was updated, now providing full records of all MCP interactions as seen in the attached image.- The post was praised for its clear and readable structure, a welcome change from walls of text.
- Square MCPās Layering Approach Exposes Many APIs: Kent C. Dodds shared an article about the layering approach used to create the Square MCP.
- With only 3 MCP tools, they expose 30+ APIs and 200+ endpoints.
Modular (Mojo š„) ā· #general (4 messages):
MLPerf benchmarks, AMD MI300x, Mojo Benchmarks
- MLPerf Benchmarks Questioned: A member asked about the possibility of seeing the MLPerf benchmark for the AMD MI300x with Mojo.
- Another member shared a link to a Modular YouTube live stream showing some Mojo benchmarks.
- Mojo Benchmarks on YouTube: Modular shared some benchmarks in a YouTube live stream, starting at 1:37:54.
- These benchmarks may provide insights into the performance of Mojo.
Modular (Mojo š„) ā· #mojo (48 messagesš„):
Memoization/Caching with Dictionaries in Mojo, Rationale Behind 'out' Argument in Mojo Functions, Implicit vs Explicit Trait Conformance in Mojo, Static Optional Type Proposal, Trait Composition
- Mojo Exception Handling: Fast or Faux Pas?: A member inquired about the performance implications of exception handling in Mojo when using dictionaries for memoization, questioning whether
raises
ortry
/except
would be more efficient.- Another member recalled from a podcast that exception handling in Mojo shouldnāt be terribly expensive, suggesting the use of
Dict.find()
to avoid exceptions by returning an optional.
- Another member recalled from a podcast that exception handling in Mojo shouldnāt be terribly expensive, suggesting the use of
- Out-standing āoutā Argument: Mojoās Memory Maestro: Members discussed the design rationale behind using the
out
argument in Mojo functions instead of returning values directly, citing benefits in scenarios like loading large ML models where memory management is crucial.- The
out
keyword allows specifying the memory location for the result, potentially avoiding unnecessary data movement and improving performance, especially when working with large data structures; itās akin to giving the compiler a pointer to uninitialized memory to directly initialize.
- The
- Explicit Trait Conformance: No More Trait-cherous Assumptions!: The discussion covered the shift from implicit to explicit trait conformance in Mojo, with the removal of implicit conformance in the next release, due to issues with API contracts.
- Explicit conformance requires developers to explicitly declare which traits a type conforms to, ensuring that all API contracts are met, while aliases can still be used for trait composition, but cannot include additional API contracts.
- Static Optional: Optionality Optionalized!: A member proposed adding a static version of
Optional
to the standard library, which may be useful for Larecs, detailing the justification on the Modular forum.- The goal is to allow to have optionality at compile time.
- Trait Composition: Combining Traits like a Boss: Members discussed how trait composition works in Mojo, noting that using
alias
to combine traits requires implementing all individual traits, while composing traits via inheritance necessitates an explicit implementation of the composition.- This approach provides flexibility in combining traits while maintaining the explicit API contracts, ensuring type safety and predictable behavior.
Modular (Mojo š„) ā· #max (5 messages):
Modular package installation with Pixi, Alternatives to 'magic' wrapper, Using pip or uv for Modular, max-pipelines conda package
- Modular Package Installation with Pixi Explored: A member inquired about the success of installing the modular package with Pixi, expressing a desire to avoid magic in a production endpoint for a Python-only project.
- Another member responded that magic is essentially a wrapper around Pixi with Modular-specific defaults, anticipating Pixi should work fine but offering to address any issues encountered.
- Pip and UV as Alternatives for Modular: A member suggested using pip or uv as alternatives for installing the Modular package, referencing the Modular documentation.
- Another member clarified that modular is a meta-package currently available for uv or pip, with the
max-pipelines
conda package being equivalent in the Conda environment.
- Another member clarified that modular is a meta-package currently available for uv or pip, with the
Nous Research AI ā· #general (29 messagesš„):
Forbidden Emojis, Vatican Compute Power, Telegram Bots for Nous AI, Remote Access for Computing, Mac vs PC for AI
- Debating Dangerous Emojis: Members jokingly debated whether certain emojis like š§āāļø, š§āāļø, and šµļøāāļø could cause irreparable damage in the wrong hands.
- Vaticanās Vertex Vault: Bloomberg Bonanza?: Members speculated on the Vaticanās compute resources, with one suggesting they possess hundreds of Bloomberg terminals.
- Remote Rigs Resurgence: Beefy Desktops Bypass Bulky Laptops: Due to ubiquitous high-speed internet, a member advocated using a shitty laptop to remotely access a beefy desktop, citing benefits like unlimited storage and persistent operation.
- Another user countered, citing the impracticality of a desktop setup due to frequent travel.
- MacBook Momentum: M-Series Marvels Mobilize AI: Members discussed the rise of MacBooks for AI tasks, pondering why competitors like Strix Halo remain subpar.
- One suggested poor drivers as a reason, mentioning George Hotzās efforts to improve AMD viability via Tinygrad.
- Intelās AI Ambitions Aborted: Habana Halted: It was mentioned that Intelās acquisition of Habana Labs to create AI chips failed massively.
Nous Research AI ā· #ask-about-llms (9 messagesš„):
Nous Hermes Uncensored, Uncensored LLMs, System Prompts
- Hermes Uncensored still exist?: A member asked if the current flagship model from Nous Research is uncensored.
- Another member replied that if you use the right system prompt, yeah but it isnāt uncensored by default.
- Uncensored LLMs are somewhat censored: A member noted that even the uncensored models are somewhat censored.
- They mentioned they tried it around 2023 and found out about this project via a Clickbait uncensored LLM video.
Nous Research AI ā· #research-papers (1 messages):
burnytech: https://fxtwitter.com/AndrewZ45732491/status/1919920459748909288
Nous Research AI ā· #interesting-links (5 messages):
AI Language Model on Windows 98, New AI model
- AI Runs on Windows 98: An AI language model was run on a Windows 98 system with a Pentium II and 128MB of RAM, as reported in a Tomās Hardware article.
- New AI Model Alert: A link to a new AI model was shared on X.com and X.com.
- New Article Published in Nature: A new article was published in Nature, accessible through this link.
Nous Research AI ā· #research-papers (1 messages):
burnytech: https://fxtwitter.com/AndrewZ45732491/status/1919920459748909288
Torchtune ā· #general (3 messages):
Tool Use, apply_chat_template, Jinja
- Apply Chat Template Jumpstarts Tool Use: A member expressed enthusiasm for supporting
apply_chat_template
in this GitHub issue, suggesting it would immediately enable tool use and resolve other related issues.- They acknowledged their inability to contribute directly due to time constraints and lack of knowledge about Jinja, but emphasized the significant potential unlock this feature would provide.
- Jinja Knowledge Gap Hinders Contribution: A member highlighted the importance of
apply_chat_template
for enabling tool use, but admitted they lack the necessary Jinja knowledge to contribute to its implementation.- They believe that resolving issue 2706 would be a huge unlock for the community.
Torchtune ā· #dev (24 messagesš„):
Optimizer in backward removal, Distributed recipes, Memory savings, FSDP CPU offload, Gradient memory
- Debate sparks on Optimizer-in-Backward Removal: Members discussed removing the optimizer-in-backward capability from distributed recipes to reduce complexity, despite potential memory savings.
- The concern was raised that it adds complexity to the code and that its impact may not be significant enough to justify the added cognitive load, especially considering that not many people are using it.
- Optimizer-in-Backward delivers Memory Savings on LLM: Experiments showed that using optimizer-in-backward with act offloading on a ll3.1 8B model finetuned on 4x3090s resulted in a 2.5GB memory saving per GPU.
- The savings are roughly proportional to gradient memory.
- FSDP CPU Offload makes Optimizer-in-Backward less relevant: Using FSDP CPU offload drastically reduced GPU memory usage (to 9.5GB per GPU on 4x3090s), making the memory savings from optimizer-in-backward less impactful.
- It was observed that optimizer-in-backward did not affect GPU memory usage, but improved speed slightly by ~20%.
- Potential benefits in throughput of removing Optimizer-in-Backward: A member suggested that for distributed recipes, they are more interested in throughput than memory.
- A concern was raised that removing the optimizer-in-backward would hurt hackability, and that a refactoring might be preferable.
Cohere ā· #š¬-general (7 messages):
NORTH platform, Paying for API key, Rate Limit Exceeded, Trial Key, VPN issue
- North Platform Details Sought: A member inquired about the NORTH platform and the possibilities for collaborating on research papers there.
- API Payment Problems Prompt Plea: A member reported encountering an error while trying to pay for an API key and asked for assistance.
- A member suggested that they turn off their VPN, donāt use any burner credit cards and email [email protected].
- Rate Limit Clarification: After receiving a rate limit exceeded error, one user was told they may have used their trial key more than allowed.
Cohere ā· #š-api-discussions (14 messagesš„):
Azure AI SDK, Cohere Embeddings, Azure AI Inference, Cohere SDK
- Azure AI SDK Mishandles Cohere Embedding Parameters: A member found that the Azure AI SDK ignores extra parameters sent to Cohere embedding models like cohere.input_type when using functions such as
client.embed()
.- Despite trying various input types (clustering, search_document, search_query), the same vectors are returned, and Azureās
input_type
parameter doesnāt seem to make a difference, as shown in their test script.
- Despite trying various input types (clustering, search_document, search_query), the same vectors are returned, and Azureās
- Cohere SDK Works as Expected: The member confirmed that the Cohere SDK functions properly, differentiating embeddings based on input type.
- They also plan to open a ticket on Azureās GitHub to report the discrepancy between the Azure AI SDK and Cohere SDK.
- Diving into Cohere Embedding Optimizations: A member inquired about detailed explanations on how Cohere embeddings are optimized for different input types.
- Another member responded that prepending specific tokens based on the chosen mode informs the model of the input type, with similar prepending done during training to achieve higher accuracy for that mode (see Cohere Docs).
Cohere ā· #š¤-introductions (3 messages):
IIT Kharagpur student introduction, GenAI and Voice Agents, Python3, Vite, TS, AI R&D collaboration
- IITian enters AI R&D space: A student from IIT Kharagpur introduced himself as someone trying to explore the Artificial Intelligence domain, focusing on R&D aspects.
- He is open to collaborating on projects and research, aiming to grow and learn in the AI field.
- Voice Agent & GenAI Engineer uses Python: The student is currently working on understanding GenAI and developing Voice Agents.
- He uses Python3, Vite, and TS for rapid development, choosing tools based on project requirements.
- AI community collaboration: The student hopes to find like-minded individuals for collaboration on real-life projects and research papers within the community.
- The goal is continuous learning and exploration in the AI arena.
Nomic.ai (GPT4All) ā· #general (6 messages):
Jinja Template for Nous-Hermes-2-Mistral-7B-DPO, GPT4All Custom API, PrivateGPT, Qwen3 support
- Users seek Jinja Template for Nous-Hermes-2-Mistral-7B-DPO: A member requested a Jinja template for Nous-Hermes-2-Mistral-7B-DPO to use with the GPT4All custom API.
- They mentioned running it on a server and needing the template because GPT4All only supports those.
- PrivateGPT flagged as RAG Model: A member mentioned finding a RAG model called PrivateGPT, but noted that the project looks dead.
- No further details or links were provided about the project.
- Jinja Template given for Hermes Model: A member shared a Jinja template for the Nous-Hermes-2-Mistral-7B-DPO model.
- The template was given as
{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}
.
- The template was given as
- Qwen3 Support?: A member inquired about support for Qwen3.
- No further details were provided.
LlamaIndex ā· #blog (1 messages):
VoyageAI Multi-Modal Embeddings, MongoDB Atlas Vector Store, Multi-Modal Retrieval
- VoyageAI Multi-Modal Voyage Begins: Learn how to do multi-modal retrieval using @VoyageAIās multi-modal embeddings and @MongoDBās multi-modal indexes.
- The notebook guides users on using VoyageAIās multi-modal embeddings and setting up MongoDB Atlas as a vector store for image embeddings.
- Multi-Modal Retrieval Notebook Lands: A new notebook demonstrates using @VoyageAIās multi-modal embeddings and @MongoDBās multi-modal indexes for multi-modal retrieval.
- The accompanying tweet links to a tutorial on creating a multi-modal index.
LlamaIndex ā· #general (2 messages):
.edu email access, Qwen2.5-VL-7B-Instruct-AWQ memory usage, VLLM memory allocation
- Inquiry about .edu email access: A member asked if anyone has or knows someone who has access to a .edu email.
- Qwen2.5-VL-7B-Instruct-AWQ Memory Consumption Exceeds Expectations: A user reported that the Qwen/Qwen2.5-VL-7B-Instruct-AWQ model, when loaded with VLLM, consumes more than 24GB of memory, despite being an AWQ model and expected to use significantly less.
- The user provided code snippet includes parameters such as
tensor_parallel_size=1
,max_new_tokens=500
,dtype="float16"
, andvllm_kwargs={"swap_space": 1, "gpu_memory_utilization": 0.98}
.
- The user provided code snippet includes parameters such as
LlamaIndex ā· #ai-discussion (1 messages):
NERDAi's vector institute
- NERDAi posts about vector institute: NERDAi made a post about vector institute.
- No specific details were given other than it involves AI tools and machine learning.
- Another Post About AI: This is a placeholder topic as there was only one actual topic.
- It is required to have at least two topics.
tinygrad (George Hotz) ā· #general (4 messages):
codegen, UOp, kernel-per-level, webgpu demo
- Codegen and UOp Get Mesozoic Boost: A user expressed gratitude for resources, especially the mesozoic one, which significantly aided their work on codegen and UOp.
- The user indicated that these resources were instrumental in their projects, highlighting the value and impact of the provided materials.
- Kernel-Per-Level Perf Parley: A user inquired about performance comparisons related to creating kernel-per-level in the software.
- They lauded the softwareās engineering and its potential for optimization through different kernel strategies.
- WebGPU Demoās Drive: A user reported making performance improvements to the webgpu demo and attached a screen recording.
LLM Agents (Berkeley MOOC) ā· #hackathon-announcements (1 messages):
Lambda, AgentX
- Lambda Re-Opens Resources for AgentX: Lambda is offering $100 serverless API credits for Inference to every individual participant in the AgentX competition, and the application form must be completed by Friday, 5/16 at 11:59pm PT via this link.
- Lambda Workshop happening Thursday: Lambda workshop happening Thursday (5/15) at 10am PT for AgentX, youāll learn how to build practical agentic applications using Lambdaās Inference API.
- Techniques for optimizing agent performance while controlling costs, best practices for deploying agents in production environments, and a live demo of an advanced AI agent powered by Lambdaās infrastructure.
LLM Agents (Berkeley MOOC) ā· #mooc-questions (2 messages):
Certificate Timeline, AgentX judging, Coursework deadline
- Certificate Release Timeline Discussed: A member asked about the timeline for receiving certificates after completing homework and lab submissions.
- Another member explained that certificates for Trailblazer/Mastery/Honorary Tier may be released in early June, while Ninja/Legendary Tier certificates will be released in August after AgentX judging concludes.
- Coursework Deadline Set for May 31st: The final deadline for all coursework is May 31st.
- Judging for AgentX (for ninja/legendary tiers) will take place during all of June, implying a delay in certificate release for those tiers.