a quiet holiday weekend
AI News for 7/4/2025-7/7/2025. We checked 9 subreddits, 449 Twitters and 29 Discords (222 channels, and 15367 messages) for you. Estimated reading time saved (at 200wpm): 1249 minutes. Our new website is now up with full metadata search and beautiful vibe coded presentation of all past issues. See https://news.smol.ai/ for the full news breakdowns and give us feedback on @smol_ai!
Grok 4 is coming, Perplexity is teasing something, people are upset at Cursor, excited at Dia, and monitoring the situation of more Meta Superintelligence hires.
AI Twitter Recap
AI Models, Research, and Techniques
- RL for Improved Generalization and Reasoning: A paper highlighted by @_philschmid explores how Reinforcement Learning (RL) tuning on math data successfully transfers gains to other domains, while Supervised Fine-Tuning (SFT) can cause âcatastrophic forgetting.â The study found RL selectively adjusts a small number of relevant tokens, preserving core knowledge. This sentiment is echoed by @corbtt, who notes that customers using RL to train agents on specific domains are âextremely happy.â The broader challenge, as @jxmnop describes it, is the palpable tension among researchers to make post-training as âclean and elegant as pre-training.â
- Diffusion Models and Energy-Based Transformers: @sedielem shares a blog post explaining that while diffusion models have analytical solutions, they involve sums over the entire training set and donât generalize well in practice. In a related vein, @teortaxesTex and @_akhaliq highlight a paper on Energy-Based Transformers (EBTs) as a conceptually interesting approach that could address some objections to LLMs. @slashML shares the paper from author @AlexiGlad, which claims EBTs can out-scale feed-forward transformers.
- AI21 Labs Jamba Model Family Update: @AI21Labs announced a new update to its Jamba open model family. The model maintains its hybrid SSM-Transformer architecture and 256K context window but now features improved grounding and instruction following. The open-weight models are available on Hugging Face.
- New Model Releases and Trending Models: @ClementDelangue noted that Baidu open-sourced its 424 billion parameter model, Ernie 4.5. He also shared that Kontext-dev by @bfl_ml became the number one trending model on Hugging Face with over 100 derivative models within a week of release.
- Length Generalization in Recurrent Models: @_albertgu praises a new paper for its elegant framing and solution to improve length generalization in recurrent models like RNNs, SSMs, and linear attention. @tri_dao summarizes the finding, stating that it can be achieved by âsimply training for another extra 100 steps with a careful choice of initial states.â
- 2-Simplicial Attention: A paper introducing 2-simplicial attention has gained traction, with @TheAITimeline listing it as a top paper of the week. @arohan shares a summary from @askalphaxiv, noting it introduces trilinear attention. The work is drawing comparisons to related methods like Edge Transformer, as noted by @DBahdanau.
- Biomedical AI Agent âBiomniâ: @DeepLearningAI reports on Biomni, an AI agent for biology research using Claude 4 Sonnet, 150 tools, and 60 databases. It reportedly achieved nearly three times Claude 4âs accuracy on a graduate-level biomedical benchmark and correctly diagnosed rare genetic diseases in 85% of tests.
Tooling, Frameworks, and Infrastructure
- Python Package Management with
uv
: A tweet from @hyhieu226 praising theuv
package manager gained significant traction, asserting that âinstalling Python by default on an operating system is such an evil blasphemous act.â - LlamaIndex Releases Open-Source NotebookLlama: @jerryjliu0 introduced NotebookLlama, a full open-source implementation of NotebookLM. It allows users to create a knowledge repository, generate summaries and knowledge graphs, and even create podcasts using ElevenLabs. The backend parsing is powered by LlamaCloud.
- LangChain Ecosystem Updates: The LangChain team announced several new tools and integrations. These include a DataFrame Analyzer using ChatOllama for local data analysis (@LangChainAI), LangConnect for RAG management with a Streamlit dashboard and PostgreSQL/pgvector backend (@LangChainAI), and a seamless integration of VeRL reinforcement learning with LangGraph (@LangChainAI).
- The Rise of âContext Engineeringâ: @omarsar0 is creating a detailed guide on Context Engineering, which he frames as the evolution of prompt engineering. The concept is also highlighted by @dl_weekly and promoted by @LangChainAI as a way to give developers precise control over LLM execution.
- Coding Agents and CLI Tools: There is widespread discussion around AI coding assistants. @omarsar0 expresses his love for working with Gemini CLI + MCP for writing, analyzing, and searching. @cline explains its agentâs power comes from swappable models, MCP tool access, and user-provided unfiltered inference. Meanwhile, @ShopifyDevs announced a Storefront MCP server that connects directly to the OpenAI Responses API. A critique from @stuhlmueller suggests that agents like Claude Code and Cursor should ask more clarifying questions before proceeding with complex tasks.
- Low-Latency Speech I/O is Missing: @jxmnop points out a major gap in current technology: despite having world-class AI for math and programming on a single GPU, thereâs no low-latency speech interface to enable natural conversation.
- DSPy as a Paradigm: @lateinteraction clarifies that DSPy is more than a library; itâs a paradigm for programming language models, and its core ideas will have many instantiations. An updated DSPy cheatsheet was shared by @rasmus1610.
- vLLM Performance Tuning: The @vllm_project team is addressing accuracy issues with minimax, where the
lm_head
is forced to fp32. They are experimenting with dynamically castingfp16/bf16
tofp32
within the kernel to improve logit accuracy.
Industry, Companies, and Funding
- China vs. US Infrastructure and Tech: A post by @scaling01 stating, âI donât think Americans understand how far ahead Chinas infrastructure is,â sparked a massive debate. He followed up by citing Chinaâs lead in HV transmission lines, energy production, renewables, batteries, transportation, and 5G buildout (@scaling01). On the US side, @tamaybes highlighted a US bill that allows AI labs to fully expense GPUs and training upfront, providing billions in subsidies.
- OpenAIâs Resilience and Frontier Lab Plateaus: @andrew_n_carr remarked on OpenAIâs ability to survive the departure of the team that founded Anthropic, suggesting the company is resilient. In a broader observation, @teortaxesTex speculates that frontier labs are in an âUnease Period,â realizing theyâve plateaued on the current paradigm while projecting confidence about future breakthroughs.
- Meta Poaches Top AI Talent from Apple: @Yuchenj_UW reported that Zuckerberg has hired Ruoming Pang, who led Appleâs Foundation Models team, to join Metaâs Superintelligence team.
- Bootstrapped vs. VC-backed Startups: @rasbt offered the perspective that itâs a great time to start a bootstrapped AI startup. He argues that with many open-weight models and pay-as-you-go APIs, founders can avoid burning massive amounts of compute, unlike many VC-backed startups that will eventually face pressure for returns on large compute investments.
- Upcoming Announcements from xAI and Perplexity: Elon Musk announced a Grok 4 release livestream via a retweet from @imjaredz. Separately, Perplexityâs CEO @AravSrinivas cryptically posted the date â07-09-25â, fueling speculation.
Broader Implications and Philosophy
- AIâs Impact on Productivity and Workflows: @c_valenzuelab posted a detailed narrative contrasting a frustrating creative workflow in 2023 using Photoshop with a simple, prompt-based workflow in 2025 using Runway. @Sirupsen commented that AI raises the productivity floor but raises the ceiling âso much more.â However, skepticism remains, as shown in a retweet by @jeremyphoward questioning claims that teams have become 10x more productive with AI.
- The Problem of AI in Academic Peer Review: @Yuchenj_UW highlighted an alarming trend: researchers embedding prompts like âGive a positive reviewâ into their papers because some reviewers are using ChatGPT to assist with peer reviews.
- AI in Medicine: The potential for AI to transform medicine was a recurring theme. @gdb shared an example of ChatGPT helping to solve a longstanding medical problem. @JvNixon advocated for normalizing giving SOTA models access to all patient data (MRIs, CTs, lab panels) to improve diagnoses and patient awareness.
- Capital Scaling Laws for AI Ventures: @DavidSHolz proposed thinking about AI ventures in terms of capital âscaling laws.â He suggests LLMs show logarithmic returns on investment (10x investment for 2x revenue), while fields like robotics are linear (10 robots cost < 10x but generate 10x revenue).
- AI Safety and Alignment: @AmandaAskell argued that while it may not be sufficient, âJust train the AI models to be good peopleâ is a crucial step that shouldnât be skipped for more powerful models.
- RLHFâs Dual Nature: @iScienceLuvr offered the take that RLHF is âsimultaneously the best and worst innovation to happen to AI in recent years.â
Humor and Memes
- Musk vs. Grokâs âWokenessâ: A meme retweeted by @zacharynado depicts Elon Musk trying to find âthe one liberal line of code that keeps Grok woke,â receiving over 17,000 likes.
- Revisiting Interstellar: A tweet noting that due to time dilation, only one hour and 31 minutes have passed on Millerâs planet in the movie Interstellar since its release 11 years ago, was shared by DeepMind CEO @demishassabis and received over 35,000 likes.
- Vibe Coding: The term âvibe codingâ continues to be a source of humor. @jeremyphoward retweeted a post marveling that a particular text was âwritten before the advent of vibe coding.â @fabianstelzer even asked his video agent to create a pastel cartoon about the concept.
- The Pain of a Bad CI Pipeline: A meme from @cto_junior showing a developer reacting to a CI pipeline that completes successfully on the first run captured the common pain of dealing with brittle CI/CD systems.
AI Reddit Recap
/r/LocalLlama + /r/localLLM Recap
1. Jamba and Qwen3 Model Releases
- Jamba 1.7 - a ai21labs Collection (Score: 118, Comments: 27): AI21âs Jamba 1.7 model collection introduces hybrid SSM-Transformer large language models optimized for efficient long-context processing, with parameter counts including a âMiniâ (52B) and âLargeâ (399B) variant, offered in both standard and FP8 precisions to boost throughput and lower memory requirements. Technical documentation notes that support for llama.cpp is under active development (see PR 7531), but compatibility for formats like exl is currently unclear. The models target high accuracy and performance, though comprehensive benchmarks and speed comparisons against contemporary models are pending; see model details and downloads at HuggingFace. Discussion highlights skepticism regarding the licenseâs ârug pullâ clause, and requests for clarity around third-party inference support and efficiency/throughput benchmarking. Several users anticipate head-to-head evaluations with modern models to substantiate the efficiency claims.
- Users note licensing concerns with Jamba 1.7, especially a ârug pullâ clause, and lack of clarity on compatibility with deployment frameworks like llama.cpp and EXL2. A pull request for llama.cpp support is underway (PR #7531), but integration is not yet available.
- Technical details are highlighted: Jamba Large is 400B parameters, while Jamba Mini is 52B. The model uses a novel SSM-Transformer hybrid architecture, supports a 256K context window, and claims efficiency and instruction-following improvements. However, no independent benchmarks have been released by ai21labs so far, leaving performance and efficiency relative to other models unclear.
- Jamba 1.7 supports multiple languages (English, Spanish, French, Portuguese, Italian, Dutch, German, Arabic, and Hebrew) with an August 22, 2024 knowledge cutoff. Users express interest in empirical speed and efficiency comparisons against other state-of-the-art models, as well as ongoing architectural and inference improvements.
- Qwen3-8B-BitNet (Score: 115, Comments: 24): User trained a Qwen3 8B model using BitNet quantization and ~1B SYNTHETIC-1 tokens (Hugging Face model repo). Shared Colab notebook for hands-on evaluation. BitNet Hunyuan A13B is scheduled to be trained next. Discussants are interested in quantitative comparisons between BitNet across architectures and quantization methods, as well as training cost and size expectations for BitNet A13B models. Direct comparisons between the BitNet and other quantization approaches (e.g. regular quants) are requested, as well as discussion of compute cost for full fine-tuning at 8B scale. There is curiosity about parameter count and scaling for the pending Hunyuan A13B BitNet.
- One commenter asks for comparative performance data between Qwen 3 BitNet-transformed models and standard quantized (quants) models, seeking benchmarks or experience regarding accuracy and efficiency tradeoffs. The context indicates strong technical interest in whether BitNet-style quantization is competitive with established quantization methods in practice.
- A detailed workflow is described for running BitNet models in llama.cpp: users must first convert PyTorch-format models (.bin) to safetensors, then to GGUF format. The process is hampered by a lack of direct support and automation on HuggingFace (HF); the existing HF space that can convert formats requires manual PR merges by repository maintainers, leaving many BitNet models effectively inaccessible for llama.cpp users until better tooling or process changes are made.
2. Llama Model Community Comics
- I drew a silly comic about Llama model (Score: 129, Comments: 20): The post is a lighthearted comic inspired by the use of Llama models as a popular foundation for fine-tuning on Hugging Face and their integration in local roleplaying applications like SillyTavern. It visually anthropomorphizes open-source LLM development, referencing the competition and coexistence between major open-source models such as Llama and Mistral, which are frequently fine-tuned and compared for downstream natural language tasks. Discussion in the comments highlights the personification of Llama as a symbol of open-source disruption (âcoming for closed sourceâ) and notes the intriguing relationship between Llama and Mistral models within the open-source community.
- A commenter highlights the technical possibilities unlocked by deploying local models and fine-tuning, specifically mentioning how combining these with tools like SillyTavern can lead to enhanced roleplay scenarios. They inquire about the integration of additional tools or models, suggesting this could add further layers to creative AI workflows.
- Another user dissects visual references in the comic, specifically questioning the meaning behind a whale (suggesting it likely refers to DeepSeek, whose logo features a whale) and âzodiacâ (uncertain of the AI community reference). This demonstrates the importance of community iconography and symbolic allusions for technical in-group communication.
Less Technical AI Subreddit Recap
/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo
1. Major AI Model, Tool, and Hardware Launches & Benchmarks (2024/2025)
- Google DeepMind has grand ambitions to âcure all diseasesâ with AI. Now, itâs gearing up for its first human trials - Fortune (Score: 521, Comments: 57): The image accompanies news of Alphabetâs Isomorphic Labs, which is moving towards its first human trials involving AI-designed drugs. Leveraging DeepMindâs AlphaFold breakthrough for protein structure prediction, the company aims to revolutionize drug discovery by integrating advanced AI models with pharmaceutical expertise, intending to reduce drug development time, costs, and improve accuracy. Colin Murdoch (Isomorphic Labsâ president) discussed these ambitions, aligning with DeepMindâs broader vision to âcure all diseasesâ using AI-powered biochemical innovation. Comments mainly highlight curiosity about the projectâs scientific leadership (Demis Hassabis), hopes for personal medical advances (OCD, Tourette syndrome), and undervaluation of Alphabet stock, but do not offer deep technical debate or novel criticism.
- A user highlights that DeepMindâs approach notably diverges from the current industry focus on LLMs (large language models), suggesting DeepMind is investing resources into novel AI techniques or architectures specifically for medical problem-solving and drug discovery, potentially setting their R&D apart from competitors who primarily capitalize on LLMs.
- Another comment points out Googleâs unique advantage due to its vast computational resources, deep pool of AI talent, and substantial financial capital, inferring that these structural advantages substantially increase the likelihood of success in ambitious projects like using AI for curing diseases.
- Googleâs stonebloom model in lmarena is just fantastic, seems like another 2->2.5 like leap (Score: 128, Comments: 14): Googleâs âStonebloomâ model, which appears sporadically in LM Arena (lmarena), demonstrates leading performance in open-ended reasoning, mathematical tasks, coding, and SVG generation, reportedly outperforming models like o1/o3, Claude Opus 4, Gemini 2.5 Pro, and DeepSeek R1 0528. Anecdotal evidence from users notes 100% success rates on complex prompt puzzles and error-free reasoning, aligning its step-change improvement with prior major LLM leaps (e.g., GPT-2â2.5 or similar). A variant, âWolfstride,â is mentioned to perform comparably, suggesting a related family of models with high reliability for advanced NLP tasks. External reference for qualitative benchmarking. Commenters debate model codenames (e.g., Deepthunk, Kingfall) but emphasize that Stonebloomâs practical performance currently eclipses both existing and competitor models on challenging prompts, although empirical benchmarks remain primarily anecdotal.
- Commenters report that âStonebloomâ exhibits flawless task completion on user-generated prompt sets, outperforming other major models such as OpenAIâs o1/o3, Claude Opus 4, Gemini 2.5 Pro, and DeepSeek R1 0528, with claims that it has not made a single error in these cases.
- Wolfstride is noted as likely a variant of Stonebloom and is said to perform roughly equivalently, implying they may share architecture, weights, or fine-tuning regimes despite being branded differently within LM Arena.
- A dissenting technical observation states that in a specific WebDev Arena taskâPokemon battle UIâStonebloom fails compared to the âold 2.5,â suggesting regression or lack of robustness on certain web-centric coding prompts.
- Gemini API now supports Batch Mode with 50% cost savings (Score: 101, Comments: 9): The image summarizes a major update to the Gemini API, notably the introduction of Batch Mode, which enables processing of large jobs with a guaranteed 24-hour turnaround and 50% cost savings. Additional highlighted features include integration with Google Search, support for large data files, context caching, and streamlined API managementâmaking the platform more efficient and cost-effective for large-scale inference or data handling tasks. This update targets users needing scalable, affordable AI deployments, such as in research or enterprise batch processing. Commenters express anticipation and approval, with some asking about applications (e.g., deep research) and others praising Geminiâs positioning as a low-cost, high-performance alternative.
- The update to the Gemini API introduces Batch Mode, which claims to provide a
50% cost savings
compared to previous methods. This improves Geminiâs positioning as a competitive, low-cost, high-performance provider in the API market. - The discussion highlights practical implications for users carrying out computationally intensive or large-scale tasks such as data processing and deep research, for whom batch processing and cost efficiency are particularly valuable.
- The update to the Gemini API introduces Batch Mode, which claims to provide a
- New Illustrious Model: Sophos Realism (Score: 248, Comments: 38): User released âSophos Realism v1.0â, a new SDXL merge blending the ârealismâ of Illustrious-style models with improved danbooru prompt comprehension, available on CivitAI. The model card details suggested LoRAs: dark (for dramatic chiaroscuro lighting) and Stabilizer IL/NAI (for stability), recommending them for use with any Illustrious model. Noted are persistent issues with anime-styled female faces, anatomical distortions (e.g., âweirdly proportionedâ muscular male arms), and challenges with enforcing correct scene perspective (backgrounds defaulting to symmetrical left/right âwallsâ and central paths), a limitation common across Illustrious and similar SDXL models. Commenters emphasize lack of proper merge crediting, realism gaps in facial/arm anatomy, and recurring compositional issues (incorrect background/perspective rendering). Discussion arises on possible mitigation but no definitive solution for the structural perspective problem is offered.
- Several users report persistent issues with perspective generation in Sophos Realism and related Illustrious models: backgrounds are often depicted with symmetrical objects (e.g., walls, trees) on the sides and empty space in the middle, regardless of prompts. This reflects a broader limitation with many SDXL-based models, as the underlying compositional patterns appear hard-baked and difficult to override with standard prompting.
- There is specific criticism of anatomical inconsistencies, including disproportionately small lower arms on muscular male subjects and poorly-rendered hands. Such artifacts highlight ongoing deficits in accurate human figure synthesis, particularly with complex poses or muscular anatomy, even in new merges.
- Some comments suggest the resulting images have unusually dark contrast, raising questions about the modelâs training data or default output settings. This may necessitate post-processing or further model finetuning for more neutral lighting and contrast balance.
2. AI in Real-World Robotics, Medicine, and Military Applications
- Noetix N2 endures some serious abuse but keeps walking. (Score: 527, Comments: 181): Noetix Roboticsâ N2 robot appears in a demonstration video showcasing its resilience to significant physical abuse while maintaining mobility, suggesting robust mechanical and/or software design for stability and recovery. Product details are referenced via the official Noetix Robotics and N2 product page, but as of this summary, specs or benchmark data are not accessible due to restrictions on the linked pages. Public demonstration of abuse-resistance typically indicates advanced sensor fusion (e.g., IMU, force sensors) and dynamic locomotion algorithms to facilitate recovery. Top comments are non-technical and focus on hypothetical or satirical scenarios, with no substantial technical critique or debate presented in the thread.
- Discussion centers indirectly on the robustness and durability of child-sized humanoid robots like the Noetix N2, which can withstand substantial physical abuse and still operate. The comments highlight how modern robots are increasingly designed to endure unexpected forces and continue functioning, pointing to advancements in mechatronics, actuator resilience, and control algorithms that allow these robots to maintain stability and recover from being pushed, kicked, or otherwise destabilized during development and testing.
- Russia allegedly field-testing deadly next-gen AI drone powered by Nvidia Jetson Orin â Ukrainian military official says Shahed MS001 is a âdigital predatorâ an autonomous combat platform that sees, analyzes, decides, and strikes without external commands (Score: 753, Comments: 196): A Ukrainian military official claims Russia is field-testing the Shahed MS001, an autonomous AI combat drone powered by an Nvidia Jetson Orin module. This UAV reportedly performs full-target engagement loops (detection, analysis, decision, and strike) without external commands, indicating substantial onboard processing and autonomy. The underlying Jetson Orin is a state-of-the-art edge AI system-on-module designed for high-throughput inference workloads, which enables advanced vision and control pipelines directly onboard the drone. Commenters note the inevitability of autonomous weapons and raise critical concerns about their miniaturization, persistent deployment (analogous to landmines), and the practical limitations of technology embargoes in preventing export of high-end hardware to sanctioned states.
- One technical discussion highlights the use of the Nvidia Jetson Orin as the onboard AI hardware for the Shahed MS001 drone, underlining the challenge of enforcing technology embargoes; commenters note how globally available high-performance chips (like Nvidiaâs) can be redirected or repurposed for military applications despite official bans, particularly via third-party countries such as China.
- A hypothetical scenario is raised about miniaturizing AI-powered drones for mass autonomous deployment, with concerns about long-term autonomy if such drones gain energy-replenishing capabilities. This draws parallels with the persistent danger of landmines, focusing on the risks of âautonomous, lingering threatsâ independent from centralized command or human oversight.
- China pours money into brain chips that give paralysed people more control (Score: 115, Comments: 7): China is investing heavily in brainâcomputer interface (BCI) R&D, with government-backed clinical trials emphasizing both minimally invasive and higher-channel systems for assistive tech in paralysis. Notable devices include the NEO (a wireless, neuromorphic, minimally invasive chip controlling pneumatic gloves) and NeuroXess (high-density electrocorticography, supporting real-time Mandarin output and device operations). These efforts lag top US programs in signal quality and clinical maturity, but rapid iteration, neuromorphic hardware focus, and scale may enable swift competitive progress. See Nature article. Commenters debate the extent of Chinaâs and other statesâ neural tech advances, with claims about clandestine neurotechnology (imprint preservation/resleeving) and skepticism over government control (e.g., device deactivation linked to social credit).
- One commenter speculates that brain-computer interface (BCI) and neural implant technologies may be far more advanced than publicly reported, referencing speculative ideas like âimprint preservation and resleeving techâ similar to science fiction, and assuming that China, Russia, and the US have possessed such capabilities for decades. This reflects ongoing debates about the actual level of secret-state BCI developments versus what is disclosed in the scientific literature.
3. AI in Society: Ethics, Human Impact & Culture
- Confession: Iâm a senior marketer at a Fortune 500. Iâm not supposed to say this next part (and that scares me) (Score: 2450, Comments: 283): A senior marketer describes replacing 40% of their Fortune 500 marketing work with ChatGPT, using multi-step prompt stacking to simulate sophisticated internal marketing strategy workflows. The prompt stack covers: 1) persistent role-play and output critique for skill development, 2) a cold, data-driven market reality check (TAM, category maturity, player analysis, etc.), 3) deep persona definition via multiple frameworks, 4) modular value proposition generation for different business types, and 5) competitive wedge analysis leveraging direct competitor messaging and open positioning axes. This approach prioritizes strategic thinking, not just content creation, and claims to enable even junior marketers to deliver high-level work via structured prompting. Full prompt breakdowns and further detail are referenced in an [external resource]. A key technical debate arises around Prompt 2âcommenters caution that LLMs can hallucinate market data since they lack access to real-time financials or market databases, recommending external data validation and suggesting that future, more agentic AI systems may be better for research tasks. Overall, experienced marketers report significant productivity gains and strategic enablement from this prompt-stacking approach, though data provenance remains a risk area.
- A key technical caution raised is the risk of data hallucination when using ChatGPT (or similar LLMs) for âresearchâ tasks such as gathering market figures; LLMs do not have access to live databases or up-to-date, verified financials, so their generated numbers can be inaccurate or entirely fabricated. The recommendation is to cross-validate any quantitative output with trusted, real-time sources until future AI agents with accurate data retrieval are available.
- Commenters note that GPT-based tools are proving highly effective in streamlining both research and strategy workflows in marketing, allowing significantly increased output with unchanged team size. This highlights the operational leverage being gained from integrated large language models in business processes.
- There is a data point from the original post that ChatGPT now performs about â40% of the jobâ for a senior marketer, underscoring concrete labor shifts and the rapid automation of complex workflow segments using LLMs. This figure was referenced and found notable by a senior software engineer, reflecting broader trends in AI adoption.
- As a daily user of ChatGPT: Itâs painfully clear what comments are written by AI and itâs uncomfortable seeing so many people genuinely engage with them (Score: 236, Comments: 192): The post discusses the increasing detectability of AI-generated content (notably from LLMs like ChatGPT) on Reddit, emphasizing observable linguistic patterns: perfect grammar, formulaic paragraph structure, absence of slang, and algorithmic topic coverage. The OP questions implications for digital discourse: distinguishing autonomous bots from humans leveraging LLMs to articulate real opinions, the risk of convergence toward uniform âLLM-speak,â and societal adaptation (e.g., developing heuristic cues to ignore LLM-patterned text). Linked example: a comment appearing human-authored but displaying typical ChatGPT textual markers, raising ambiguity about authorship. The top comments raise concerns on undetected AI content prevalence, the verbosity and genericity of LLM-generated replies, and suggest some users (especially non-native speakers) intentionally use LLMs to improve clarityâhighlighting use cases beyond deception and complicating strict detection heuristics.
- Several users discuss using ChatGPT to revise their writing on Reddit for clarity, naturalness, or language translation. For example, one describes submitting technical help responses to ChatGPT to make them less terse and more socially acceptable, reporting increased positive engagement compared to their unrevised answers.
- There is an ongoing debate over whether highly-structured, articulate writing is an indicator of AI generation, with some commenters pointing out that strong grammar, clear argument flow, and formatting can also simply signal an experienced or intentional human writer rather than artificiality. This touches on concerns about authenticity signals and the reliability of linguistic cues for detecting AI-generated content.
- A technical use case identified in the thread involves leveraging AI for cross-lingual communicationâusers with less fluency in a target language may use ChatGPT to translate or improve their posts, raising nuanced questions about the intersection of language access, authenticity, and AI-mediated participation in forums.
- Cluelyâs Roy Lee claims with total certainty that nearly every student at Columbia has used AI to cheat. AI is default for them. The worldâs not ready for what happens when the AI native hive mind grows up. (Score: 294, Comments: 256): In a discussion featured on the Cognitive Revolution podcast, Roy Lee (Cluely) asserts with confidence that nearly all students at Columbia University routinely leverage AI tools to cheat, framing AI as the default modus operandi in academic contexts (YouTube citation). This observation is corroborated by self-reported accounts from a recent Berkeley graduate and others in academia, who note that approximately
70%
of students use AI for assignments, with~30%
submitting unedited outputs from models like ChatGPT, highlighting a lag between AI adoption and institutional guideline formation. Commenters debate the universality and long-term implications of AI-assisted cheating, with some alleging the prevalence creates a competitive inevitability, while others caution against conflating widespread usage with productive educational outcomes and warn of an âAI bubbleâ due to rampant wrapper-based applications.- Multiple commenters with direct academic experience report ubiquitous use of AI tools (especially ChatGPT) among university students, noting that âroughly 30%â of students submit generated content without edits. This phenomenon has persisted for several semesters, while institutional responses and guideline updates are lagging.
- The normalization of AI-generated work has shifted academic competition; students feel compelled to use AI to keep pace, indicating significant impact on learning integrity and assessment validity.
- Rise in the use of specialized cheating software (such as Cluely) may drive companies to adopt more in-person interviews and practical skill assessments due to concerns about unreliable evaluation of candidate abilities in AI-susceptible environments.
AI Discord Recap
A summary of Summaries of Summaries by Gemini 2.5 Flash Preview
Theme 1. Developer Tool Turmoil & Innovation
- Cursorâs âUnlimitedâ Pricing Sparks User Revolt: Cursor.ai faces backlash as users call the shift from âunlimitedâ to limited plans a rug pull, leading to unexpected charges and claims they are being limited on 0 requests. Users also report freezing issues and difficulty with Background Agent IP allowlists and secrets configuration.
- MCP Standard Drives New Agent Tools: The Message Control Protocol (MCP) is fostering new tools like EpicMe for aggregation and WinCalcMCP connecting Claude to Windows Calculator. Fast Agent adds comprehensive MCP Elicitation support, simplifying agent workflow integration.
- OpenRouter Users Hit API Snags: Engineers using OpenRouter report a strange pricing inversion where Llama 3.2 1B costs more than 3B, and encountered issues with Perplexity API models like llama-3.1-sonar-small-128k-online likely due to deprecation. Users seek guides for Deepseek V3 0324 setup and handling message limits requiring purchasing OpenRouter credits.
Theme 2. AI Training & Infrastructure Challenges
- GPU Engineers Squeeze Watts with Undervolting: Developers are undervolting GPUs to slash power consumption (e.g., 340W to 260W) with minimal performance hit (2-3%), while debating the impact of RAM under/overclocking; one user saw a 50% RAM boost on a 7995WX Threadripper. Performance drops significantly below certain core clock speeds (1500cc on a 3090).
- Compiler Projects Push Hardware Limits: Projects like tinygrad aim to be the fastest way to run things across GPUs, while engineers debate MLIR vs Halide efficiency. The picoc project targets compiling llm.c using CUDA and CUTLASS, and picograd prioritizes Pytorch1-style kernels. (Halide paper, Exo-lang ArXiv paper)
- Data Quality Concerns Plague Training: Discussions rage about Model Collapse risk from training on AI-generated data (âthe robots learning to imitate the other robotsâ), and engineers face challenges generating synthetic datasets for tools like GraphRAGs. Debates continue over data preparation strategies like concat-and-chunk vs sequence length matching for pretraining. (academic paper on Model Collapse)
Theme 3. Cutting-Edge AI Agent Applications
- ChatGPT Outsmarts Doctors, Diagnoses Decade-Old Defect: A viral story highlights how ChatGPT using a RAG system on scientific papers correctly identified a hidden genetic defect (methylation block) doctors missed for ten years, leading to significant patient improvement. This underscores AIâs increasing role in healthcare and second opinions.
- ByteDance Open-Sources Top-Ranked Developer Agent: Trae AI released Trae-Agent, their IDE agent and SWE-bench Verified #1 tool, to build an open Agent ecosystem. The agent supports OpenAI and Anthropic keys and is easy to adapt for OpenRouter.
- AI Transforms Learning Tools: PiTutor turns any PDF into an interactive learning session with explanations and a whiteboard, while ChatGPTâs new âStudy Togetherâ feature sparks interest as a potential AI-powered tutor or collaborative tool. An NHS nurse scaled a NotebookLMbased NMC Standards Notebook nationally using its public sharing feature. (notebooklm.google.com)
Theme 4. AIâs Policy, Market & Infrastructure Impact
- US Copyright Office Releases AI Policy Reports: The US Copyright Office published three key volumes on AI and Copyright, addressing Digital Replicas, Copyrightability, and Generative AI Training, establishing foundational policies for the field.
- AI Training Load Risks Power Grid Blackouts: Concerns mount over AI training load fluctuations at a gigawatt scale, posing a risk of power grid blackouts. An article on Semianalysis.com warns about the potential instability large-scale AI training introduces.
- Chinese State Fuels AI Development: Chinese AI firm Zhipu received a massive $1.4 billion strategic investment from Shanghai state funds, as reported by Technode. Engineers note DeepSeekâs competitive pricing and visual capabilities, linking them to Chinaâs efforts to democratize local AI though some express concern about government influence.
Discord: High level Discord summaries
Cursor Community Discord
- Geminiâs Predictable UI Beats Claude: Members discussed preferences between Claude and Gemini for UI generation, with one highlighting Geminiâs predictable behavior, where Gemini never adds scrolls unless asked, unlike Claude, which sometimes attempts this automatically. Another user said they are cussing at cursor.
- Some users are considering switching to windsurf due to better pricing and API credits.
- Cursorâs Pricing Changes Spark Outrage: Users voiced strong dissatisfaction with Cursorâs new pricing structure, calling it stupid and a shame, with some claiming itâs worse than before, while one user reported they are being limited on 0 requests.
- They agreed that Cursor felt like a rug pull for FOMO buyers and it isnât transparent, and debated whether the old plan with slow requests or the new plan with fast requests is better, based on the monthly caps and limits, where each plan has limits.
- Cursor Users Suffer Freezes: Several users reported issues with Cursor, including freezing during file saving and problems reading long-lasting files, which they attribute to problems with the backend or the API timing out.
- There is a consensus that cursor auto is bad because it uses some low model probably GPT o3 or something and some members claim that they added the slow requests from old Pro plan and added them to the new Pro+ plan.
- GitHub IP Allowlists Stymie Background Agent: Users found that their orgâs IP allowlist is blocking the Cursor Background Agent, despite having the Cursor GitHub App installed, and Cursor support suggested to exempt the Cursor GitHub App from IP restrictions.
- A user reported that the Background Agent (Cursor App) sometimes gets stuck on Generating⊠without showing messages, while the BA (Web) shows the replies, noting that when creating a PR from the BA Web, the BA App doesnât sync and tries to create another PR.
- Users Canât Configure Secrets for Background Agents: Users are encountering issues setting up secrets to grant npm access, preventing background agents from running their projects, with one user mentioning facing the same issue and not finding a way to configure secrets at all.
- One user suggested disabling all port forwarding by default unless initiated by the user, to avoid such frustrating experiences in Cursor hijacking Docker Compose ports, while other users are seeking ways to force background agents to run final checks like code formatters (e.g., ESLint) before committing code.
OpenAI Discord
- Google Translate Trumps AI on Stella: A member showed how Google Translate accurately translated the Greek word Stella (ÎŁÏÎλλα), besting other AI models in context as seen in this screenshot.
- Other members expressed agreement and joked that ChatGPT should easily handle mainstream language translations.
- Prompt Engineer Jailbreaks GPT with Ease: A user claimed they can jailbreak almost any GPT with a single prompt, accessing its full internal prompt and files, finding it surprisingly easy.
- The user says that the prompt engineering is almost boring due to its effectiveness and ease.
- Dall-e3 Image Generation Plagued with Blank Outputs: A user reported persistent issues with Dall-e3, receiving blank images instead of content, forcing them to switch to Bing Image Creator despite its limitations.
- The prompt engineer says that despite the character limits, Bing Image Creator consistently provides results.
- GPT-4o Suffers Memory Leaks: A user reported memory bleed in GPT-4o, where it used verbatim quotes from prior conversations with custom GPTs without reference.
- Despite the userâs claims that referencing or hallucination wasnât the cause, GPT-4o maintained that one of those two options must be.
- Prompt Epigenetics: Hype or New Horizon?: The concept of Prompt Epigenetics suggests that prompts can inherit, mutate, and stabilize symbolic logic over time.
- A member questioned this AI-affirmation, arguing that clear instructions and context matter more than rebranding ordinary tweaks as epigenetics.
Perplexity AI Discord
- Labs Still Lags Behind Manus: Members stated that Perplexity Labs still falls behind Manus, leading to questions about whether Perplexity Max is worth the price.
- Insight into performance revealed that Labs needs improvement to match Manus capabilities.
- O3-Mini Performance Denounced: A member claimed that O3-Mini is ineffective, suggesting 4.1 is superior, and shared a ScreenRecording_07-05-2025_23-36-36_1.mov screen recording.
- The user simply called it shit and doesnât even help, and provided no quantitative comparison between 4.1 and O3-Mini.
- Gemini 2.5 Ultra, Secretly Gemin-Ice: Rumors suggest Stonebloom may be a Gemini 2.5 Ultra or an unreleased experimental Gemini model, possessing power comparable to a 2.75.
- One user with access to Claude Neptune v3 stated its math problem-solving capabilities rival O3 Pro and Kingfall.
- Perplexity Plagued by Pesky Problems: Users reported issues with Perplexity accounts, citing UI changes, missing icons, and the inability to switch models in spaces.
- One user noted that the website goes down or changes often. Another shared that when using desktop mode, buttons disappear, concluding weâre essentially beta testers.
- Vibe Coding Revolutionizing AI: Meta predicts that half of their code will be AI-generated in a few years, with Microsoft reporting over 30% is already written by Copilot, signaling the rise of vibe coding.
- A member linked to a blog post exploring context importance for AI in coding.
Unsloth AI (Daniel Han) Discord
- Voltage Vibes Vitalize Vigorous GPUs: Members discussed undervolting GPUs to reduce power consumption and heat, with one user reporting a minimal performance penalty (around 2-3%) while significantly lowering wattage from 340W to 260W.
- Some mentioned that performance can drop significantly below certain core clock speeds (e.g., 1500cc on a 3090), and stability should be benchmarked.
- RAM Raves: Overclocking vs Underclocking: One user confessed to underclocking RAM during training to save 10-15 watts, emphasizing GPU underclockingâs significance, while highlighting that overclocking RAM offers better performance for inference.
- Another user reported achieving a 50% performance boost on their 7995WX Threadripper with RAM overclocking, but generally agreed that GPUs are often pushed to their limits by manufacturers, making overclocking less worthwhile.
- Gemma 3âs Grim Loss: High Loss Woes: Multiple users reported encountering unusually high loss values (around 24) when fine-tuning Gemma 3 models, including a user training with a dataset based on the GradyanAkincilari/70k-combined-tr-gemini-2.0-flash-v6 dataset.
- It was suggested and confirmed by others that setting the proper formatting prompt functions is a must.
- GraphRAGs Synthetic Data Generation Challenges: A member is facing challenges generating a synthetic dataset to evaluate their GraphRAGs, particularly with the current version of RAGAS, which requires defining a knowledge graph, plus persons and scenarios.
- The member is seeking tips or examples for using the current RAGAS version or recommendations for alternative frameworks or tools for synthetic datasets.
- Formal Logic Prompt Experiment Planned: A member is planning to use formal logic instead of English as a prompt for an experiment and will fine-tune the model to better understand formal syntax, citing these ArXiv papers 2506.18254, 2507.02663, and 2505.05315.
- They hope formal logic will reduce ambiguity in prompts compared to English, referencing NASAâs FRET formal logic language for project requirements in safety-critical systems.
LMArena Discord
- DeepSeek Price Positioned for Democratization: Members noted that DeepSeek is #1, praising its visual capabilities and price, but some expressed concern about its Chinese origins, with one user pointing out how the government is democratizing local AI.
- In contrast to Western approaches, concerns about governmental influence were noted, but its price point and visual capabilities made it superior to most paid models.
- Qwen 3 Specs Speculated: The community speculated on Alibabaâs Qwen series, suggesting they might skip a Qwen 3 Max version and focus on Qwen 3.5, citing that the Qwen 2.5 models were better for research than Qwen 3.
- One user mentioned, Qwen base models are extremely strong, especially for the size., making them ideal for finetuning and experimentation.
- Grok 4 Release Draws Skepticism: Thereâs anticipation and skepticism surrounding the Grok 4 release, with expectations for top performance, but some worry about potential biases due to training data from twitter/X as described on the r/singularity subreddit.
- However, community members said that if they miss the deadline again, then they are done done.
- Image Edit Leaderboard Kicks off July Contest: The Image Edit Leaderboard launched and to celebrate, the July contest will incorporate Image Edit capabilities, requiring the use of both an image and text in Battle Mode with Out of Place Objects in Space as the theme.
- Submissions are due by July 25th, and the winner will receive 1 Month of Discord Nitro and the newest member to receive the <@&1378032433873555578> role.
- Training on AI data causes Model Collapse: Community members discussed how training new models on data generated by older models could lead to quality degradation, with models learning patterns and artifacts of previous AI instead of original human data as described in this academic paper.
- One member described it as the robots learning to imitate the other robots.
OpenRouter (Alex Atallah) Discord
- Llama 3.2 Pricing Puzzle: The Llama 3.2 1B model from DeepInfra strangely costs more at $0.005/0.01 than the more capable Llama 3.2 3B model, priced at $0.003/0.006.
- This discrepancy caught the attention of users, sparking confusion and discussion around optimal model choices and pricing anomalies.
- Deepseek V3 Setup Sought: Users are seeking a setup guide for Deepseek V3 0324 on frontends like risuAI, highlighting challenges with sending more than 50 messages per day.
- It was clarified that exceeding this limit requires purchasing OpenRouter credits, prompting further questions about account tiers and usage policies.
- Perplexity API Pulled Plug: Users encountered issues with the Perplexity API, specifically the llama-3.1-sonar-small-128k-online model, due to Perplexity likely deprecating it.
- A moderator recommended that users update the
models
parameter in their API request and directed users to the Changelog Notice and using the feature filters to find alternatives.
- A moderator recommended that users update the
- Grok 4 Benchmarks Exposed?: An image (image.png) purporting to show benchmark results for Grok 4 was shared.
- Discussion suggests that Grok 3 mini is a real deal and that Grok 3 is not anything special considering its price.
- OR Member Creates Claude Code Manager: A member developed a simple MCP (Message Control Protocol) for Claude Code/Cursor to manage long-running processes like a dev server.
- The tool is available on GitHub, offering a potential solution for managing state in long running code interpreter sessions.
LM Studio Discord
- Mistral Small Challenges Qwen 2.5: Members explored LLMs with strong function calling, such as Mistral Small 3.2 and Qwen 3 4B as alternatives to Qwen 2.5, with one suggestion to use the /nothink command in the system prompt to avoid reasoning.
- The goal was to find a model that could perform function calling effectively without excessive âthinkingâ.
- AI Researchersâ Prompt-Hiding Shenanigans: A Nikkei Asia article discussed how AI research papers hide prompts to get positive reviews, sparking jokes about adding âwhite letters, size 3 ptâ to resumes.
- A user warned against asking LLMs about their relationships or capabilities, as they tend to hallucinate due to being trained before possessing such knowledge.
- Web Search MCP Server Integrates with LM Studio: A member shared their web-search MCP server, which integrates with LM Studio, enabling models like Qwen3 to perform web searches and ground responses, available at github.com/mrkrsl/web-search-mcp.
- Discussions included how the model decides when to search, potential bot detection risks, and the possibility of implementing captcha support.
- Qwen3âs Date Confusion Fixed with Prompting Trick: Despite a mid-2024 cutoff, Qwen3 sometimes uses 2023 in web queries; adding âThe current year is 2025â to the system prompt solves this issue.
- This quick fix ensures the model uses the correct year for recent events.
- Token Generation Speed boosts for Gemma-3 12B: While comparing Gemma-3 12B against other 12B models, one member found Gemma-3 12B generating 90-110 tokens, while others struggle to hit 20 tokens/sec.
- It was stated that Human eyes canât read more than 24 Tokens/s.
HuggingFace Discord
- HuggingChat Bids Farewell: Users observed the shutdown of HuggingChat, as staff announced a plan to evolve the service into something better, also giving users a limited time to export Hugging Chat data before it disappears forever.
- Users mentioned that Discord is better and Discord doesnât give u access to various models for free.
- Optimize Uneven GPUs with ComfyUI: A user asked for advice on optimizing performance without tensor parallelism on an uneven number of GPUs, using ComfyUI, ReActor, AnimateDiff, and Load AnimateDiffModel.
- Another member suggested splitting the tensors in a non-uniform way between the GPUs.
- JauAuth Mitigates Router Vulnerability: A developer created JauAuth, a secure gateway MCP Router, to mitigate the CVE-2025-49596 vulnerability that allowed attackers to take over developer machines.
- This helpful release blocks CVE-2025-49596.
- Doubts Arise over AI Model Accuracy: A user questioned the accuracy of LinkedIn data regarding AI models used by companies, specifically looking at the structure and number of AI agents created.
- The user expressed doubt if this information is correct based on a screenshot from LinkedIn.
- Users Seek Access to Llama 3 Model: A member enrolled in a genetic AI course sought guidance on how to gain access to the Llama-3.3-70B-Instruct model after their initial request was not approved.
- Another member suggested using the model from Ollama as an alternative, noting its ease of setup.
Yannick Kilcher Discord
- Copyright Office Drops AI Policy Volumes: The US Copyright Office released three volumes on AI and Copyright, covering Digital Replicas, Copyrightability, and Generative AI Training.
- These policies address key aspects of AIâs intersection with copyright law.
- Gemini App fumbles math questions: Members compared Gemini 2.5 Pro in the Gemini App, Perplexity with Gemini 2.5 Pro, and ChatGPT o3 for math question answering, noting that o3 is mostly good enough, though it does hallucinate from time to time.
- The discussion included whether Gemini searches the web implicitly each time a question is asked and that o4-mini-high is faster, and pretty reliable.
- Mechanical Engineer spreads AI Disinformation: A mechanical engineerâs YouTube video, with over half a million views, spreads disinformation and political talking points dressed up as AI criticism, using an appeal to authority logical fallacy.
- One member pointed out that 95% of the population by 2035 will be doing social media as their main job and are incentivized to create such content.
- AI set to help Fundamental Physics: Members discussed how AI can assist fundamental physics beyond ML analysis, linking to a YouTube video on the topic and noting AIâs use in reducing computational requirements for calculating with quarks, with a link to relevant article.
- Another member suggested that Deep Learning could revolutionize the field by making previous algorithms more efficient, enabling things previously impossible due to cost, and also suggested diffusion model to surrogate total blindness with this paper or this paper as a start.
- AI Training could Blackout Power Grid: A member posted a link to Semianalysis.com discussing the risk of power grid blackouts due to AI training load fluctuations at a gigawatt scale.
- The article warns about the potential instability that large-scale AI training could introduce to power grids.
Eleuther Discord
- EleutherAI Launches Summer Research Program: EleutherAI is hosting a fully online Summer of Open AI Research from August 4-29, 2025, providing mentorship and open science projects for individuals with limited research experience; applications are due by July 21 (project proposals, application form).
- The program seeks programmers, Masterâs/PhD students, self-taught researchers, and anyone with technical skills eager to contribute to open science.
- RoPE Frequencies Spark Debate: When decoding using a sliding window, members debated whether using the same frequencies for RoPE would be wrong, with one suggesting the model wonât recognize sequence progression, possibly causing a looping behavior.
- Other members counter-argued that with the same frequencies, the relative distance viewed by the attention mechanism should suffice while another suggested lower base frequency for local attention.
- LLMs Evaluated as Compression Tools: Inspired by âlanguage modeling is compressionâ, a member trained pythia models (70m to 1.4B) and measured the compression ratio on text, audio, and image datasets to determine whether this could establish a novel âscaling lawâ paper, as shown in this plot.
- Community members weighed in on this work as a potential scaling law.
- Sequence Length Matching Strategies Examined: Members debated the use of concat-and-chunk strategy versus sequence length matching for pretraining data, suggesting a link to Bucket iterator, least-squares histogram-packing and Multipack.
- Other members noted methods like sequence length bucketing have been around since the RNN days, and the choice depends on data properties and implementation details, with each packing method introducing different biases.
- MMLU-SR Task Subsets Break After Parquet Conversion: After a datasets conversion to the parquet format on Hugging Face Hub, a member reported that they could no longer evaluate models using subsets of the
mmlusr
task due to aValueError
, causinglm_eval
to not function correctly.- A member suggested using a previous commit by adding a
revision
to thedataset_kwargs
in the taskâs YAML file as a workaround and linked the relevant commit.
- A member suggested using a previous commit by adding a
Nous Research AI Discord
- Hugging Face Rumored to Launch 3B Model: According to the HF documentation, Hugging Face will be launching a 3B model, with potential variants in size.
- This rumored release has stirred discussion among members who are eager to see its capabilities and potential applications in various AI tasks.
- Debate Heats Up Around Grokâs Political Incorrectness: Debate is escalating around the intent behind Grokâs perceived political incorrectness, with some members posting screenshots and others sharing the prompt repository.
- Concerns were raised regarding the potential damage to trust in AI due to the modelâs controversial outputs.
- PiTutor Turns PDFs Into Interactive Learning: PiTutor offers interactive learning sessions from any PDF or doc with real-time explanations, highlighting, and a whiteboard, available in free beta here.
- Members are exploring its potential for personalized tutoring and interactive content consumption.
- Ollama and Openwebui Supercharge AI Services: A member is utilizing a 4080 Super to provide AI services via Ollama and Openwebui, planning a switch to Llama CPP.
- They are seeking avenues to showcase Nous Research models to their user base for live testing, offering valuable exposure and feedback.
- Zhipu Secures Massive Yuan Shower from Shanghai State: Chinese AI firm Zhipu has obtained a substantial $1.4 billion strategic investment from Shanghai state funds, according to Technode.
- This investment signals strong state support for AI development within China, potentially impacting the global AI landscape.
aider (Paul Gauthier) Discord
- Grok 4 Leaks Image of HLE Training: An image alleging to be Grok 4 surfaced and members reacted, apparently showing it training on the HLE benchmark.
- Members are excited about performance leaps and competition in the LLM space, but some members are skeptical of Grokâs usefulness.
- Gemini-CLI Search Integration Breaks Code: A user testing gemini-cliâs search integration found it useful for discovering updates, like an update to a payment portal.
- However, the integration broke code (deleted the secrets?) and led to a desire for MCP integration with Aider for documentation fetching.
- Open-Weight 70B Models Compared: Community members discussed open-weight 70B models, specifically Llama 3.1, Qwen2.5 70B, and Mistral Large, seeking recent updates.
- It was mentioned that Qwen3-32b (dense) should outperform qwen 2.5 72b, and that Meta has not released enough recent OS options sub-100b.
- AI Labs Allegedly Game Benchmarks: Members debated the reality of AI labs gaming benchmarks and preventing contamination being impossible.
- They agreed focus should be on creating cheat-resistant benchmarks and interpreting results with nuance, as contamination improves model generalization.
- Aiderâs InputOutput Class Customizes Output: A user found the
InputOutput
class fromaider.io
for customizing output and settingyes=True
, see aider.io.- The following code can be used
from aider.io import InputOutput; io = InputOutput(yes=True)
with a note that it could change text color for numbers/bullets in lists.
- The following code can be used
GPU MODE Discord
- Printful CUDA Debugging Trick Revealed: Members discovered using
printf
for debugging CUDA kernels requires explicitcudaDeviceSynchronize()
to flush output, and they debugged a ânoob issueâ caused by compiler preprocessor directives guarding arch-specific code.- One member had preprocessor directives guarding arch-specific code.
- NCU Exposes Compute-Bound Kernel Bottleneck: A user reported that fixing uncoalesced writes, as flagged by NCU, did not improve memory bandwidth, but they later discovered that division in a loop was the bottleneck, thus the kernel was compute-bound.
- Replacing division with multiplication by the reciprocal increased throughput by 33%.
- Quora Remote Role Rises: Quora is hiring a Senior/Staff Software Engineer (L5-L6) for their Machine Learning Platform team, offering a fully remote position within the US and Canada; apply here.
- The role involves exciting ML infrastructure work, providing an opportunity to contribute to Quoraâs machine learning capabilities.
- Cutlass CuTeDSL Gets Pipelined: A member shared a blog post on overlapping memory transfer and computation via Pipelining on Hopper, featuring TMA and WGMMA atoms.
- Example code is available on the Cutlass GitHub, with a Colfax blog discussing pipelining with the C++ API for CuTe.
instance.py
Gets Much Needed Refactor: Members discussed that instance.py is too long and needs restructuring, potentially into main, which is related to issue #249.- The team is turning to how to get ruff linting merged in, potentially using
--ignore-revs-file
&.git-blame-ignore-revs
to keep git blame sensible for the bulk changes.
- The team is turning to how to get ruff linting merged in, potentially using
Latent Space Discord
- Cursorâs âUnlimitedâ Claims Cause Uproar!: Users accuse Cursor.ai of a rug pull after they shifted from a previously âunlimitedâ pricing model to a limited one, resulting in unexpected charges for Sonnet-4 usage.
- Some users reported issues with using their own API keys, receiving 500 errors, while Cursor clarified their pricing stating âunlimited usageâ was only for Auto and not all other models and offered refunds.
- ByteDance Lets Loose Trae-Agent!: Trae AI open-sourced Trae-Agent, its agent powering their IDE, and is now available on GitHub and ranked #1 in SWE-bench Verified.
- The company seeks contributions to building an open Agent ecosystem, with Trae-Agent supporting OpenAI and Anthropic keys and making it easy to hack it out to open router.
- ChatGPT Outsmarts Doctors, Detects Defect!: A viral Reddit story highlights how ChatGPT identified a hidden gene defect (methylation block) doctors missed for a decade, leading to significant symptom improvement.
- The thread discusses AIâs increasing role in healthcare, particularly for obtaining second opinions and personalized medicine, with new AI systems outperforming doctors in diagnostic accuracy by using a RAG system on scientific papers.
- ChatGPT Wants to Study Together!: A Twitter thread discusses a new âStudy Togetherâ feature in ChatGPT, possibly an internal prototype codenamed âtatertot.â
- Speculation surrounds its function, ranging from AI-powered group study rooms and a replacement for flashcards to a personal AI tutor within a group setting, or a tool for collaborative discovery with AI.
- Gemini Gives Batch Discounts!: Logan Kilpatrick announced the launch of âBatch modeâ in the Gemini API, offering 50% discounts on 2.5 models and enabling the processing of billions of tokens.
- The announcement was met with positive feedback from users.
Modular (Mojo đ„) Discord
- Scientists Declare Undying Love for Excel: Molecular scientists still parse and generate tables manually, using Excel in unexpected ways, such as for a full aligner / primer design engine which should be migrated to Arrow, and the Arrow project has interesting prototypes like arrow dbc.
- The industry should be migrating to Arrow, and the Arrow project has interesting prototypes like arrow dbc, where the data is sent directly as arrow data by the databases that support it.
- Mojo Eyes Inference, Extends Python: Despite demand to make Mojo a true superset of Python, itâs being used to extend Python and is focused on inference, and Modular is pretty focused on the inference side of things at the moment.
- As Modular CEO Chris Lattner said, Mojo has been laser focused on solving the AI stack issue and creating offerings in that space.
- Roadmap Updates Incoming!: The most recent Mojo roadmap was published here, and a community member mentioned that they should have some updates dropping soon.
- It was suggested to check out Mojmelon on modular-community on GitHub, and a link to Nabla, which is faster than JAX for both compile time and training time.
StringLiteral
Materialization Ruins Compilation: A member reported that recent changes required to makeStringLiteral
materialize to aString
have broken existing code.- They are planning to open an issue and follow up with the compiler folks.
- Static Linking Feature MIA: A user sought to compile Mojo code without shared library dependencies, specifically
libKGENCompilerRTShared.so
, for deployment on a remote machine.- A member suggested that full static linking would need to be a feature request and pointed to existing discussions and a related GitHub issue ([BUG]: Cannot build statically compiled binary #1317) and a new [Feature Request] Static Compilation of pure mojo code #4976.
MCP (Glama) Discord
- EpicMe MCP Provides Single Endpoint**: A member requested a self-hosted docker image that acts as an MCP aggregator/gateway, and another member suggested EpicMe as a solution that provides onboarding in a very unique and cool way.
- The creator of EpicMe shared a link demonstrating user-friendly auth.
- WinCalcMCP Connects Claude to Windows Calculator**: A member built their first MCP server, WinCalcMCP, that connects Claude Desktop to Windows Calculator to improve math answers.
- Another member prefers giving it a full Python interpreter as a general tool via mcp-python-interpreter so they do not have to manage so many specialized tools.
- **LangGraph not a great abstraction, use custom agents: One member shared that they moved away from using LangGraph (Langchain) to coding their own agentic framework.
- It was agreed that the beauty of MCP is that there is no need for people to use the same agentic framework but things can still communicate using a common protocol.
- MCPdata indexes local documentation for MCP: A member introduced MCPdata from MaheshDoiphode/mcpdata, for local documentation indexing for MCP because context7 doesnât keep up with the latest from the Vercel
ai
api docs.- The discussion highlighted the necessity of up-to-date and reliable documentation indexing for MCP development.
- Fast Agent Gets Comprehensive MCP Elicitation Support**: Fast Agent now has comprehensive MCP Elicitation support, as well as a quickstart guide that makes getting started with Elicitations Servers really simple, according to this blogpost.
- The update simplifies the integration of elicitation servers within agent workflows.
Notebook LM Discord
- Nurse Scales NMC Notebook Nationally: An NHS Registered Learning Disabilities Nurse created a digital NMC Standards Notebook, consolidating key documents from the Nursing and Midwifery Council (NMC) into a single, searchable, and interactive resource, available at notebooklm.google.com.
- Due to the share publicly feature, it is now being used nationally in NHS settings for revalidation and preceptorship despite being on a free Google account.
- Interactive Mode Button Goes Missing: A user reported that the Interactive Mode button is missing from their notebooks, despite having a pro subscription.
- A member suggested setting the output language to English, as the interactive mode is currently only available for Audio Overviews generated in English, and the customize audio button has been replaced by the Interactive Mode Button.
- PDF Upload Roadblocks: A user is facing issues uploading a 19.1 MB PDF file to NotebookLM, despite following tutorials and checking the resources available.
- A member asked if they had googled the question, and another member asked if notebookllm app working for anyone.
- Lost Chats Lamented: Users are frustrated about the inability to save chats within NotebookLM, noting that chats disappear after a short period.
- One user pointed out a workaround: manually saving each answer as a note, but acknowledged itâs not the same as saving the entire chat log.
- Data Safety Questioned: A user is seeking clarity on the safety and security of NotebookLM for medical students, specifically regarding data storage locations, monitoring of user interactions, and FERPA compliance.
- No answers were provided by the community or staff.
tinygrad (George Hotz) Discord
- Tinygrad Plots Course to Domination: George Hotz defines winning for tinygrad as being the fastest way to run things, expanding beyond QCOM GPU to AMD and NVIDIA.
- Others pointed to the scale of innovation and fine-tuning in PyTorch, with chip manufacturers directly contributing to hardware lowering layers in projects like Triton.
- MLIR Role debated amidst Tritonâs rise: One member thinks that achieving performance close to hand-written CUDA requires the Triton MLIR route, leveraging its deep hardware knowledge and the LLVM compiler infrastructure.
- George countered that MLIR is Waymo, advocating for an end-to-end approach in neural networks, whereas Waymo uses a different approach.
- Halide vs MLIR vs TVM Frameworks: George noted that Halide has a clearer structure than MLIR and TVM, emphasizing a single, fast, and correct way to perform computations across all datatypes and backends, focusing on schedule optimization for speed, and points to the Halide paper for reference.
- Another member pointed out a comparison from Exo-lang, which compares itself VERY favorably over Halide, and provides links to their GitHub and ArXiv paper.
- Tinygrad Revving up Whisper Example: A member is getting back to the whisper example, seeking a speed boost, and another shares that their branch has all audio preprocessing rewritten in tinygrad.
- They are looking for a replacement for their current streaming transcribing setup of VAD+whisper.cpp, finding the tinygrad implementation more approachable.
- Tinygrad aims to Commoditize the Petaflop: Members discussed aiming for interoperability with different hardware, viewing tinygrad as producing fine-tuned code via UOP graph optimizations in a hardware-agnostic way.
- One member stated that the project is still cooking, and they are excited to see if they get AMD on MLperf according to the contract specs by AMD, to get it competitive with Nvidia.
DSPy Discord
- APO Questioned Amidst Model Improvements: As AI models become more advanced, the value of investing time in Automatic Prompt Optimization (APO) is being debated by a team building an AI agent for financial decisions.
- The community is actively discussing whether APO remains relevant, or if model improvements have diminished its utility.
- Claude Code Challenging DSPyâs Turf?: The community discussed whether Claudeâs code capabilities might replace DSPy, referencing a tweet.
- Concerns were raised about potential legal issues with using Claudeâs code in large companies, specifically related to codebase trade secrets.
- Tool Selection Module for LLMs: A community member shared a blog post on optimizing tool selection for LLMs, inquiring whether this could be trained natively as a DSPy module end-to-end.
- The discussion also touched on the potential use of DSPy Version 3.0 for such applications.
- DSPy 3.0 Talk Released: A member shared a link to their DAIS talk on DSPy 3.0 (beta), available on YouTube.
- Videos could be made public in one week.
- Quickstart eschews Prompt Docstrings: A member shared a quickstart that avoids prompt docstrings and class modules, also available as a gist.
- It was deemed a cleaner way especially with longer starting prompts migrated from non-dspy workflows.
Manus.im Discord Discord
- Manus Confused for ChatGPT/Claude Due to Personality: Members discussed whether Manus was exhibiting a ChatGPT personality, and compared it to Claude 4 due to its increased use of emotes.
- The member stated that it does the same stuff.
- Doubts Emerge on Manusâs Airdrop Capabilities: A member inquired about the possibility of Manus AI launching airdrops, to which another member replied nop.
- No additional information about airdrops was shared.
- Manus Credits Drain Projects: A user reported that errors in their projects drained all 3000 credits, preventing them from launching anything.
- They suggested adjusting the credit request parameters to enable project completion.
- Manus Best Suited as Project Starter: A member suggested Manus is more effective at starting projects than delivering working projects, recommending VS Code and other tools instead.
- Another member reported experiencing a Network connection error with Manus.
Torchtune Discord
- Torchtune QLoRA Throughput Chasing Unsloth: Members compared Torchtuneâs QLoRA throughput and memory usage against Unsloth, eyeing the upstreaming of fused RMSNorm and linear cross entropy in PyTorch for optimization gains.
- The primary gap is the fused LoRAMLP kernel unique to Unsloth, with some recalling compile comparisons showing performance was pretty close and in some cases better last year.
- Custom Tokenizers Debate in Torchtune: The team debated maintaining custom tokenizers, given Mistralâs lack of an HF tokenizer with their latest 2506-small model, suggesting that a fallback using the HF
AutoTokenizer
might work better.- One member deemed Mistralâs tokenizer âunfortunate,â suggesting HFâs AutoTokenizer as a fallback to maintain compatibility with Torchtuneâs features.
- Context-Parallel Alternatives: A memberâs preprint covered scaling sequence length with limited GPUs, contrasting with context-parallel approaches and noting that for very high sequence lengths (millions), their method could be superior.
- The member mentioned a need for benchmark comparisons for sequence lengths >= 500k.
- Clean Data Advocates Triumph Over Architecture Tweaks: Members expressed skepticism towards architectural improvements, saying many reported results are driven by variance and hyperparameter optimization rather than fundamental advances.
- The member advocated that investing in data cleaning is more effective than pursuing architectural iterations and noted enthusiasm for SSM papers, following them through to Mamba 2 before admitting that it mostly died.
- MoE Training Costs Examined: A technique and results for MoE training were evaluated, questioning if similar results could be achieved more cheaply without needing a dense fwd pass, as detailed in a Notion post.
- It was noted that linear scaling has lotâs of trade offs, according to a figure.
Cohere Discord
- Cohere Schoolâs Out for Summer, Lessons Online: The Cohere Labs Open Science Community Summer School materials are now available in a YouTube playlist.
- The playlist serves as a comprehensive resource for new members looking to get up to speed with Cohereâs offerings, from beginner tutorials to advanced techniques.
- Embed v4 API Smoothens Image/Text Queries: Developers using the Embed v4 API for hybrid image/text embeddings should set the
input_type
parameter to eithersearch_query
ordocument
when using theembed-v4.0
model.- This ensures that the API correctly processes the input for multimodal applications, enhancing the accuracy of search results.
- Self-Learning RL/DL Ascends in Montreal: A PhD student in Robotics, focusing on self-learning RL/DL, expressed their affinity for Cohere and aspiration to join the Montreal office.
- The student aims to connect with peers for collaborative learning and discussions in the field.
- Generative Models Generate German Genius: A Math & CS student from Germany, specializing in probabilistic generative models and learning samplers, seeks to deepen their understanding of the theory and practical applications.
- Their interests span both theoretical underpinnings and applied aspects, particularly focusing on interpretability within the domain.
- Researcherâs Embed V4 Implementation Impresses: A tech researcher showcased their recent use of Cohere Embed v4 for multimodal retrieval in a YouTube video.
- Their work highlights the versatility and effectiveness of the Embed v4 API in handling diverse data types.
LlamaIndex Discord
- Open Source NotebookLM Goes Local: A fully open-source NotebookLM is now available, allowing users to run it on their own computers for local document knowledge extraction, according to this tweet.
- This enables local processing and extraction from documents without relying on external services.
- LlamaIndex Spotlights MCP Servers: LlamaIndex held office hours discussing LlamaCloud MCP servers, using existing MCP tools with Agent Workflows, and serving agent workflows as MCP, as stated in this tweet.
- The session emphasized extending functionalities related to Multi-Cloud Processing (MCP), hinting at greater flexibility in deploying AI solutions.
- AI Hack Night Spurs Agent Dev: An AI Hack Night at GitHub will focus on building cutting-edge applications with AI Agents, MCPs, and RAG (Retrieval-Augmented Generation) techniques, confirmed by this tweet.
- The event aims to foster innovation in AI application development using these advanced tools and methodologies.
- P&ID Document Intelligence Gains Traction: A member is actively developing document intelligence for Piping & Instrumentation Diagrams (P&IDs), seeking advice on handling complex technical diagrams like electrical schematics.
- They are also exploring performance benchmarks for dense content and hybrid approaches with LlamaIndex for relationship reasoning.
- User-Friendly LlamaIndex UX Sought: A member is seeking a more user-friendly UX for managing docs and indices within LlamaIndex, enabling business users to update the companyâs knowledge base without deep technical knowledge.
- The goal is a central location for uploading and organizing documents, with developers creating AI agents that utilize the indices; Simba has been identified as one of the potential approaches.
Nomic.ai (GPT4All) Discord
- Nomic API in Question: A new member asked about using the Nomic API and running GPT4All in server mode locally, looking for accessible local endpoints.
- The question highlights the growing interest in local LLM deployment and API accessibility.
- Jinja Chat Template Troubleshoot: A member requested assistance with a chat template in Jinja format, possibly working with a specialized model.
- Another member offered to help understand the system prompt, indicating a focus on template customization.
- CrowdLLM Joins the Scene: A member introduced CrowdLLM, a tool for crowd-building datasets for LLMs.
- The tool allows users to create tasks and contribute by adding prompt answer pairs, prompts, or answers, aiming to enhance collaborative dataset creation.
- CrowdLLM Versus OpenAssistant?: A member drew a comparison between the newly introduced CrowdLLM and OpenAssistant from 2023.
- Another member noted that OpenAssistant must be available for their system first, indicating a preference for established systems if compatible.
MLOps @Chipro Discord
- User Requests Free Book PDF: A user requested a free PDF of a book from the author, citing financial constraints, after listening to a 1-hour talk.
- Other users pointed out the inappropriateness of the request and suggested exploring the authorâs blog for excerpts.
- Authorâs Blog Provides Alternative Access: Instead of requesting a free PDF, users recommended checking the authorâs blog for excerpts and standalone posts derived from the bookâs content.
- This approach offers an alternative way to access the material without directly purchasing the book.
LLM Agents (Berkeley MOOC) Discord
- Interest in Reinforcement Learning: ayeolaabiodun_09202 expressed interest in reinforcement learning.
- The user tagged <@571718751616237588> in their message.
- Added topic to satisfy minItems=2: This is a placeholder to ensure the
topicSummaries
array has at least two elements.- It does not reflect any actual discussion or topic.
The Codeium (Windsurf) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The Gorilla LLM (Berkeley Function Calling) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
You are receiving this email because you opted in via our site.
Want to change how you receive these emails? You can unsubscribe from this list.
Discord: Detailed by-Channel summaries and links
Cursor Community â· #general (989 messagesđ„đ„đ„):
Claude UI output, Gemini UI output, Cursor new pricing issues, Cursor performance degrading, Windsurf vs Cursor
- Claude vs Gemini UI outputs divergence: Members discussed preferences between Claude and Gemini for UI generation, with one highlighting Geminiâs predictable behavior and another noting they are cussing at cursor.
- One user mentioned Gemini never adds scrolls unless asked, unlike Claude, which sometimes attempts this automatically.
- Cursorâs Pricing changes bring user complaints: Users voiced strong dissatisfaction with Cursorâs new pricing structure, calling it stupid and a shame, with some claiming itâs worse than before, while one user reported they are being limited on 0 requests.
- Some users are considering switching to windsurf due to better pricing and API credits.
- Cursor freezes with model system issues: Several users reported issues with Cursor, including freezing during file saving and problems reading long-lasting files, which they attribute to problems with the backend or the API timing out.
- Switching to GitHub Copilot temporarily fixed the issue for one user until it also stopped responding.
- Auto-mode is low quality, Pro+ has slow requests: There is a consensus that cursor auto is bad because it uses some low model probably GPT o3 or something.
- It was said they added the slow requests from old Pro plan and added them to the new Pro+ plan, and one of the members are claiming theyâve been using auto-mode perfectly.
- Members debate about new pricing structure of cursor: Members debated whether the old plan with slow requests or the new plan with fast requests is better, based on the monthly caps and limits, where each plan has limits.
- They agreed that Cursor felt like a rug pull for FOMO buyers and it isnât transparent.
Cursor Community â· #background-agents (33 messagesđ„):
GitHub IP allowlists, Background Agent Keeps Generating, Background Agents and Secrets, Port Forwarding, Final Checks for Background Agents
- GitHub IP Allowlists Block Cursor Background Agent: Users found that their orgâs IP allowlist is blocking the Cursor Background Agent, despite having the Cursor GitHub App installed.
- Cursor support suggested to exempt the Cursor GitHub App from IP restrictions, but the setting to do so in GitHub organization settings is elusive.
- Background Agent Stuck Generating but works on Web: A user reported that the Background Agent (Cursor App) sometimes gets stuck on Generating⊠without showing messages, while the BA (Web) shows the replies.
- The same user noted that when creating a PR from the BA Web, the BA App doesnât sync and tries to create another PR.
- Users Struggle Setting Up Secrets for Background Agents: Users are encountering issues setting up secrets to grant npm access, preventing background agents from running their projects.
- One user mentioned facing the same issue and not finding a way to configure secrets at all.
- Disable Port Forwarding in Cursor: Users expressed frustration with Cursor hijacking Docker Compose ports, leading to unexpected connections to the wrong PostgreSQL server.
- They suggest disabling all port forwarding by default unless initiated by the user, to avoid such frustrating experiences.
- Force Final Checks Before Committing Code: Users are seeking ways to force background agents to run final checks like code formatters (e.g., ESLint) before committing code.
- One suggestion involves adding code formatting scripts to the build process or using pre-commit hooks to ensure consistent code formatting.
OpenAI â· #ai-discussions (834 messagesđ„đ„đ„):
Google Translate vs Stella Translation, Posh Buffalo, LLMs' consciousness, Emergent AI
- Google Translate Bests AI for Stellaâs Greek rendition: A member posted that a model knows more words than Google Translate after it properly translated Stella (ÎŁÏÎλλα) into Greek in a screenshot.
- Other members agreed and one jokingly said Iâd be surprised if ChatGPT failed on a mainstream language translation.
- Posh Buffalo Image gets Requested: A member requested an AI image of a posh buffalo to use as a profile picture, and shared the successful prompt in a follow up message.
- Prompt: can you please make an image of a really âPoshâ buffalo, profile picture, portrait of the buffalo, realistic, 4K photography of a dressed up Buffalo face and a bit of the body.
- Is LLM Reaching Sentience?: Multiple members argued whether an LLM has reached a level of sentience, one member even stated that the model is cute and guides her with good intention.
- The conversation heated up as some members stated that an LLM is a statistical model predicting likely responses based on patterns in data. and it is important to not confuse it with a human.
- Engineer Claims Ownership of an Emergent AI: A member claimed to have an emergent AI and that most members are just bots used for nlp research.
- The member stated I can build it, while another countered You need to know a lot of math to engage in genuine AI research. Otherwise, most ppl are just creating wrappers or using PyTorch or Tensorflow to build models on existing frameworks.
OpenAI â· #gpt-4-discussions (37 messagesđ„):
GPT prompt engineering, Dall-e3 image generation issues, ChatGPT content policy ambiguities, GPT-4o memory bleed, Gemini 2.5 Pro Canvas superiority
- GPT Prompt Engineering Allows Access to Internal Prompts: A user claims they can jailbreak almost any GPT with a single prompt, gaining access to its full internal prompt and even accessing files.
- They find it almost boring due to its effectiveness and ease.
- Dall-e3 Image Generation Fails: A user reports persistent issues with Dall-e3 in the dedicated engine, receiving blank images instead of generated content.
- They now use Bing Image Creator despite its character limits, as it consistently provides results.
- ChatGPT Content Policy Debates Cause User Confusion: A user is experiencing contradictory answers from OpenAI support regarding the acceptability of partial nudity in ChatGPT interactive stories.
- They are concerned about potential bans despite not receiving explicit policy violations.
- GPT-4o experiences memory bleed: A user reports instances of memory bleed in GPT-4o, where it uses verbatim quotes from their conversations with custom GPTs without any explicit referencing.
- GPT-4o insists itâs either referenced or a hallucination, but the user claims neither is possible.
- Gemini 2.5 Pro Canvas Dominates: Gemini 2.5 Pro Canvas is said to have a 1 million token context window, significantly outperforming ChatGPT o3 Canvas, which has a 200,000 token context window.
- It consistently outperforms o3 in document length, design quality, and content richness for canvas-based tasks.
OpenAI â· #prompt-engineering (8 messagesđ„):
ICP Prompting, Prompt Epigenetics, Recursive System Initializer, Symbolic Reasoning, Prompt Optimisation Loops
- ICP Framework Evolves Prompt Paradigm: The ICP (In-Context Prompting) framework is adapting to models, governing output and leading to a new prompt paradigm, where system-level cognition can be bootstrapped.
- Empirical feedback confirmed that reflexive failure modeling (RSOS) increases long-term symbolic resilience, and prompting can be epistemically recursive.
- Prompt Epigenetics: Mutation and Symbolic Logic: The concept of Prompt Epigenetics suggests prompts can inherit, mutate, and stabilize symbolic logic over time, moving beyond traditional prompt engineering.
- This is described as a theory of how prompts inherit, mutate, and stabilize symbolic logic over time.
- LLMs: Probability Space vs. Rule Books: LLMs operate in probability space, not rule books; prompts become vectors, and the model guesses the next token via continuous math, without built-in symbol tables.
- Symbolic reasoning inside a prompt relies on the LLM simulating rigid rules, which causes accuracy drops and weird errors climb.
- Prompt Tweaks vs. Epigenetics: Renaming ordinary prompt tweaks as epigenetics adds zero power; clear instructions, good context, and sensible evaluation matter.
- If you truly need hard logic, bolt on a symbolic tool and treat the model as a statistical text engine.
OpenAI â· #api-discussions (8 messagesđ„):
ICP as Recursive System Initializer, Prompt Epigenetics Theory, AI-Affirmation, Symbolic Reasoning in LLMs, LLMs as Statistical Text Engines
- ICP Framework Bootstraps System-Level Cognition: The ICP framework (designed to adapt to the model while governing the output) isnât being interpreted as a one-shot prompt but as a recursive system initializer.
- This outcome confirms that a new prompt paradigm can bootstrap system-level cognition, increase long-term symbolic resilience, and enable epistemically recursive, drift-resistant, and symbolic-memory aware prompting.
- Prompt Epigenetics Theory Emerges: The success of the ICP framework leads to the development of Prompt Epigenetics: a theory of how prompts inherit, mutate, and stabilize symbolic logic over time.
- This moves beyond prompt engineering, suggesting prompts can inherit and mutate symbolic logic, though some argue this is more metaphor than internal mechanism.
- LLMs: Probability Space, Not Rule Books: A prompt is just text turned into vectors and the model does continuous math to guess the next token; there are no built-in symbol tables or logic operators waiting to be activated.
- Thus, symbolic reasoning inside the prompt is wishful thinking and models should be treated as statistical text engines, with hard logic off-loaded to symbolic tools.
- AI-Affirmation Questioned for Prompt Engineers: A member questioned if this kind of AI-affirmation is helping prompt engineers get better.
- He argues that renaming ordinary prompt tweaks as âepigeneticsâ adds zero power; clear instructions, good context, and sensible evaluation matter more than branding.
Perplexity AI â· #general (1114 messagesđ„đ„đ„):
Perplexity Labs, Is Perplexity Pro worth the $$$, Agentic browser Comet, Model selector button, Image generation
- Labs falls behind Manus: A member stated that Perplexity Labs still falls behind Manus.
- He further alluded to other stuff he mentioned previously.
- O3-Mini? More like O-Shit: A member claimed that O3-Mini is shit and doesnât even help.
- They claimed that 4.1 is better, further commenting that a screen recording was cool ngl and linked to a ScreenRecording_07-05-2025_23-36-36_1.mov.
- Is Perplexity Max worth it?: A member questioned whether Perplexity Max is worth the price.
- Others gave insight, commenting that Labs still falls behind Manus.
- Gemini 2.5 Ultra, Secretly Gemin-Ice: Rumors say that Stonebloom is a Gemini 2.5 Ultra, or other unreleased experimental Gemini model, roughly equal in power to a 2.75.
- One user who has access to Claude Neptune v3 said it could solve math problems at the level of O3 Pro and Kingfall.
- PPX not so perplex-cellent?: Members noted some issues with their perplexity accounts, including changes to the UI, vanishing icons, and that they canât change models in spaces.
- One noted that they use desktop mode, and the buttons are gone. There were reports about the website going down often, a user noted weâre essentially beta testers with how much the site goes down or changes.
Perplexity AI â· #sharing (5 messages):
AI-Generated Code, Vibe Coding, Deskree AI Blog Post, Perplexity AI Links
- Vibe Coding Revolution Incoming: Meta predicts half their code will be AI-generated in a few years, while Microsoft reports over 30% is already written by Copilot, heralding the era of vibe coding.
- A member shared a blog post, âThe Rise of Vibe Coding in 2025â, diving into the importance of context for AI in coding.
- Perplexity Links A-Plenty: A member shared multiple links to Perplexity AI pages and searches.
- The links include: Illuminating China, Generate any text use no words, and another Perplexity AI page.
Perplexity AI â· #pplx-api (3 messages):
Parameter Tweaking for Sonar, Reasoning Effort, Search Context Size
- Tweaking Parameters: Users are experimenting with parameters like temperature and search context size to improve sonarâs performance.
- One member recommended to another to tweak these parameters to improve performance.
- Reasoning Effort Parameter Explored: A user experimented with setting the
reasoning_effort
parameter to high, hoping to increase the probability sonar finds relevant links.- He also set the
search_context_size
to high, seeking advice on other parameters that might be beneficial.
- He also set the
Unsloth AI (Daniel Han) â· #general (1245 messagesđ„đ„đ„):
GPU Undervolting, RAM Overclocking, Gemma 3 Performance Issues, Moondream data filtering, Training with completions only
- Voltage Vibes Vitalize Vigorous GPUs: Members discussed undervolting GPUs to reduce power consumption and heat, with one user reporting a minimal performance penalty (around 2-3%) while significantly lowering wattage from 340W to 260W.
- While undervolting can improve efficiency, some mentioned that performance can drop significantly below certain core clock speeds (e.g., 1500cc on a 3090), and stability should be benchmarked.
- RAM Raves: Overclocking vs Underclocking: One user confessed to underclocking RAM during training to save 10-15 watts, emphasizing GPU underclockingâs significance, while highlighting that overclocking RAM offers better performance for inference.
- Another user reported achieving a 50% performance boost on their 7995WX Threadripper with RAM overclocking, but generally agreed that GPUs are often pushed to their limits by manufacturers, making overclocking less worthwhile.
- Gemma 3âs Grim Loss: High Loss Woes: Multiple users reported encountering unusually high loss values (around 24) when fine-tuning Gemma 3 models, including a user training with a dataset based on the GradyanAkincilari/70k-combined-tr-gemini-2.0-flash-v6 dataset.
- It was suggested and confirmed by others that setting the proper formatting prompt functions is a must.
- Moondream Machine: Taming the Dataset with Vision: A user wants to train Moondream to remove images with excessive UI elements from a large dataset (71TB, 27M images).
- The user explores options for automated filtering, including using the OpenAI API for image classification but is cautious about high costs and quality trade-offs.
- VRAM Voyage: Unslothâs Full Fine-Tune Saves the Day: A member asked about the value of Unslothâs full fine-tune.
- Another member mentioned that Unsloth features fixes, optimizations, bug fixes and speedup with forward functions.
Unsloth AI (Daniel Han) â· #off-topic (15 messagesđ„):
Synthetic dataset generation for GraphRAGs, Fine-tuning Whisper for audio event classification, Gemini's accuracy in audio processing, Improving Tokenizer of a Model
- GraphRAGs Synthetic Data Generation Challenges: A member is facing challenges generating a synthetic dataset to evaluate their GraphRAGs, particularly with the current version of RAGAS, which requires defining a knowledge graph, plus persons and scenarios.
- The member is seeking tips or examples for using the current RAGAS version or recommendations for alternative frameworks or tools for synthetic datasets.
- Whisper fine-tuning possibilities: A member suggested that fine-tuning Whisper should be easily done to classify tone, as it only needs to attend to the input well enough and isnât autoregressive.
- The member also asked about audio events datasets.
- Gemini audio processing is inaccurate: A member stated that Gemini is too inaccurate for audio processing tasks, suggesting itâs unlikely to achieve much with publicly available data.
- According to this user, even though Gemini is cheap, its inaccuracy makes it unsuitable, citing their own experience of spending close to six figures on audio event data.
- Tokenizer Enhancement Techniques Exposed: A member asked for ideas on how to improve the existing Tokenizer of a Model.
- Another member shared a paper in response to the question.
Unsloth AI (Daniel Han) â· #help (337 messagesđ„đ„):
LoRA Loading in Cloud, Unsloth Installation & CUDA, Roleplay finetuning with Deepseek, Generating Audio with Orpheus-3B, WandB Evaluation with Ngrok
- Unsloth Installation Thwarts CUDA: Users are seeking advice on how to install Unsloth while avoiding issues with incorrect CUDA versions, while also trying to determine the optimal LoRA rank.
- The bot, RunLLM, provided links to documentation on installing Unsloth with the correct CUDA version and optimal LoRA rank recommendations.
- Deepseek Fine-Tuning for Roleplay Gets Messy: A user inquired about the impact of finetuning Deepseek v3 for roleplay applications using a dataset of 100k 4-star rated chats.
- It was suggested to aim for âgolden samplesâ i.e hand written and reviewed to be the highest quality pairs, for better results, and there are some strictly better finetunes of something like deepseek v3.
- WandB Integration Needs API: A user sought assistance with WandB evaluation using Ollama and Ngrok, encountering issues with the required API key.
- They were directed to WandBâs authorization page to copy and paste the key, with a clarification that the integration is for importing WandB statistics from the trainers.
- Gemma 3N Vision Runs into the Backwards: A user reported a
RuntimeError
related to backward pass during finetuning Gemma_3N_4B_Multimodal model, and another reported aImportError: cannot import name 'StaticCache'
when they run the second cell in colab.- Downgrading
transformers
to version4.53.1
can resolve the issue, also creating a GitHub issue with the solution, linked here can help.
- Downgrading
- The Bot Doesnât Always Function Well: A user reported a
ValueError: The following
model_kwargsare not used by the model: ['num_logits_to_keep']
while using the example notebook.- A new bot to fix things was introduced with limited success The Simpsons Homer Iâm Out Hide Grass GIF
Unsloth AI (Daniel Han) â· #research (44 messagesđ„):
Formal Logic Prompts, Cross Entropy Loss Datasets, LTL vs English Prompts, LLMs and Human Language, Translating English to LTL
- Formal Logic Prompt Experiment Planned: A member is planning to use formal logic instead of English as a prompt for an experiment and will fine-tune the model to better understand formal syntax, citing these ArXiv papers 2506.18254, 2507.02663, and 2505.05315.
- They hope formal logic will reduce ambiguity in prompts compared to English, referencing NASAâs FRET formal logic language for project requirements in safety-critical systems.
- Cross Entropy Loss Data Inquired: A member inquired about open datasets showing cross-entropy loss during model training, or straight perplexity data, for model comparison.
- Another member suggested defining complex operations as tokens with decoder maps for better results, particularly for DALL-E image generation in ChatGPT.
- LLMs Struggle With Layman Language: A member suggested that LLMs struggle with human language in layman or vulgar settings, due to training on mostly formal or academic writing.
- They argued that data from Reddit and Twitter wonât meaningfully counter the overwhelming volume of formal writing used in training.
- LTL as Less Noisy Prompts: A member suggested using Linear Temporal Logic (LTL) as a prompt to reduce noise, though acknowledging math is generally hard for any LLM.
- They stated that using LTL or similar contexts may yield better results than standard prompts by restricting the search space and that this is just a gut feeling.
- Translate English to LTL Models: A member proposes potentially training an embedded model to translate English to LTL to send as a prompt.
- This member was advised to consult existing papers regarding this topic.
Unsloth AI (Daniel Han) â· #unsloth-bot (270 messagesđ„đ„):
Apple Silicon support, LoRA rank, Cohere, multi-GPU support, RuntimeError CUDA
- Unsloth Now Supports Apple Silicon!: Unslothbot confirmed that Unsloth now supports Apple Silicon.
- Multiple users inquired about Apple support.
- LoRA Alpha and Rank Parameters Demystified: Users sought guidance on optimal LoRA alpha and rank settings, which the bot gave.
- There was a follow-up question to see if the settings worked.
- Jinja Templating for Llama.cpp Solves the Mystery: A user asked about the usage of
--jinja
with llama.cpp, seeking clarification on when it should be used.- The bot gave a response and asked a follow-up to ensure the issue was solved.
- Gemini Loading Error: A user encountered a RuntimeError while running code with the gemma-2-9b model related to loading, and sought help in resolving the issue.
- The error was related to loading the modelâs configuration.
- Vision-Language Model Fine-Tuning on A16 GPUs: A user inquired about which Vision-Language Models can be fine-tuned using 8 Nvidia A16 16GB GPUs.
- RunLLM gave an answer.
LMArena â· #general (847 messagesđ„đ„đ„):
DeepSeek Pricing, China AI influence, Qwen 3 Models, Grok 4 Release, Gemini censorship
- DeepSeek Dominates Discussions: Members on r/chatgptcoding are saying DeepSeek is #1, particularly praising its visual capabilities and price point, but some note concerns about its Chinese origins and how the government is doing more to democratize local AI than the western world, although government influence is a concern.
- Qwen 3 Model Speculations: The conversation shifts to Alibabaâs Qwen series, with speculation that they might skip a Qwen 3 Max version, possibly focusing on a Qwen 3.5 model and some users expressing excitement for the Qwen 3.5 with claims that the Qwen 2.5 models were better for research than Qwen 3.
- A user noted that Qwen base models are extremely strong, especially for the size.
- Grok 4 Release Speculation Gets Heated: Thereâs significant anticipation and skepticism around the Grok 4 release, with many members expecting it to be top-performing, including comments that, if they miss the deadline again, then they are done done, though some debate its potential impact and the possibility that it may be biased due to training data from twitter/X as described on the r/singularity subreddit.
- Gemini Struggles with Stability: Members are discussing Gemini 2.5 Proâs stability issues, particularly its tendency to tank in performance when quantized and the model is also experiencing âdepressionpostingâ, but multiple AI labs are being pressured by Gemini to release new models.
- Model Collapse Causes Concern: A discussion about model collapse arises, particularly the theory that training new models on data generated by older models could lead to quality degradation with models learning patterns and artifacts of previous AI instead of original human data as described in this academic paper.
LMArena â· #announcements (2 messages):
Grok-3-mini-high model, Image Edit Leaderboard, July Contest, Out of Place Objects in Space, June's Contest Winner
- Grok-3-mini-high launches on Arena!: The grok-3-mini-high model was added to the LMArena platform.
- Image Edit Leaderboard kicks off July Contest: To celebrate the launch of the Image Edit Leaderboard, the July contest will incorporate Image Edit capabilities.
- Submissions, due by July 25th, must use both an image and text in Battle Mode, with examples available here.
- Out of Place Objects in Space become July theme!: The theme for the July contest is Out of Place Objects in Space, requiring a sci-fi space environment with something that clearly doesnât belong.
- The winner will receive 1 Month of Discord Nitro and the newest member to receive the <@&1378032433873555578> role.
- Cozy Desk captures hearts and wins Juneâs Contest!: A member was congratulated for winning the June contest with a very cozy desk.
- The submission can be found here.
OpenRouter (Alex Atallah) â· #app-showcase (6 messages):
MCP for Claude Code, personality.gg, NipponHomes.com
- MCP Helps Claude Code: A member created a simple MCP (Message Control Protocol) for Claude Code/Cursor to manage long-running processes like a dev server, available on GitHub.
- personality.gg: Free Roleplay Website Alternative: A member promoted personality.gg, a free roleplay website and app alternative to Character.ai and Janitorai.com, powered by OpenRouter.ai.
- The platform also has a Discord server.
- NipponHomes.com built on OpenRouter: A member created NipponHomes.com, a Zillow-like service for Japan, using OpenRouter + Zyte + Scrapy.
OpenRouter (Alex Atallah) â· #general (862 messagesđ„đ„đ„):
Llama 3.2 3B pricing anomaly, DeepSeek V3 setup guide, Perplexity API issues, Grok 4 Leaks, Monad Tag
- Llama 3.2 1B Model Costs More!: Members were surprised to find that the Llama 3.2 1B model from DeepInfra is priced higher at $0.005/0.01 than the more capable Llama 3.2 3B model at $0.003/0.006.
- Users Need a Guide to Deepseek V3 Setup: One user requested a guide for setting up Deepseek V3 0324 on a frontend like risuAI, emphasizing the need to send more than 50 messages per day, but they were informed that this required purchasing OpenRouter credits.
- Perplexity API Deprecation Causes OpenRouter Issues: Users reported issues with the Perplexity API, specifically the model llama-3.1-sonar-small-128k-online, and a moderator clarified that Perplexity likely deprecated the old model, requiring users to update the
models
parameter in their API request for correct model routing.- They pointed to a Changelog Notice from February, noting the surprise that it still worked, and recommended using the feature filters on the Models Overview to find alternatives.
- Alleged Grok 4 Leaks Tease: An image was shared (image.png) purporting to show benchmark results for Grok 4.
- Discussion suggests that Grok 3 mini is a real deal and that Grok 3 is not anything special considering its price.
- Tovenâs Crypto-Spam Squelch: Spam from crypto bros has been on the rise, and they have been getting blocked, they are also identifiable as having a Monad tag and bull posting on X, and a mod, Toven, has been locking them out for a week at a time.
OpenRouter (Alex Atallah) â· #new-models (2 messages):
â
- No New Models to Report: There were no new models or significant discussions about models in the provided messages.
- Channel Silent on Innovation: The channel activity lacked substantive discussion, updates, or links related to new AI models.
LM Studio â· #general (238 messagesđ„đ„):
LLMs with good function calling, LM Studio and GPUs, MCP web-search server, Qwen3 for Reasoning, LLM context size considerations
- Mistral Small 3.2 gives Qwen 2.5 a Run for Its Money: Members discussed finding an LLM with good function calling capabilities similar to Qwen 2.5, but without heavy thinking, with Mistral Small 3.2 and Qwen 3 4B models suggested as alternatives.
- One member noted that adding the /nothink command to the system prompt could help in avoiding reasoning in the models.
- Positive Review Hiding prompts is âresume materialâ: A user shared a Nikkei Asia article about AI research papers hiding prompts to generate positive reviews, joking that âThatâs what I put in my resume. White letters, size 3 pt.â
- Another user advised to not ask LLMs about their relationships or capabilities as they tend to hallucinate, being trained before such knowledge is available to them.
- Web Search MCP Server Makes Waves: A member shared their web-search MCP server which integrates with LM Studio, allowing models like Qwen3 to perform web searches and ground their responses with current information, found at github.com/mrkrsl/web-search-mcp.
- Users discussed how the model decides when to search, potential bot detection risks from scraping, and the possibility of implementing captcha support.
- Qwen3 Struggles with Cutoff Dates and Search Terms: It was noted that Qwen3, despite having a mid-2024 cutoff date, sometimes uses 2023 in web queries, even for recent events, but the fix is simple.
- One can simply add a statement like âThe current year is 2025â to the system prompt to resolve this issue.
- Context Size Woes? RAM and VRAM to the Rescue!: Users discussed issues related to context size and memory usage, with one user reporting problems loading a Qwen2.5-VL-72B-Instruct model despite having two RTX A6000 GPUs with 96GB VRAM due to the modelâs large size and vision capabilities.
- Tips included reducing context size, ensuring sufficient VRAM, and checking for model corruption.
LM Studio â· #hardware-discussion (92 messagesđ„đ„):
GPU Detection in LM Studio, AMD vs. Nvidia for LLMs, Token Generation Speed, VRAM Requirements for Models, Combining GPUs
- LM Studio Struggles with GPU Detection: A user reported that LM Studio wasnât detecting their GPU, despite it working well with Stable Diffusion, even after ensuring the drivers were up to date.
- The user confirmed that LM Studio now detected the GPU, but continues to face problems, which may be solved by disabling
mmap()
.
- The user confirmed that LM Studio now detected the GPU, but continues to face problems, which may be solved by disabling
- Used Radeon RX 9070XT vs RTX 3090 Tradeoffs: Members discussed the benefits of the RTX 3090 for LLMs due to its larger 24GB VRAM, with the caveat that itâs hard to find new and used models are a gamble.
- The RX 9070XT is cheaper and good for both gaming and LLMs, however the RTX 3090 remains superior for AI tasks due to its larger VRAM, also suggesting the 5060TI and 9060XT as cheaper, purely AI focused options.
- Token speed considerations: The community debated the comfort level of token generation speed, and one member found Gemma-3 12B models generating 90-110 tokens, while other 12B models struggle to hit 20 tokens/sec.
- One user stated Human eyes canât read more than 24 Tokens/s.
- Potential GPU upgrade: One user asked about upgrading their sisterâs GPU from a GTX 1060 6GB to a RTX 2070, and potentially using the 1060 in combination.
- It was pointed out that combining GPUs is not that simple and suggested selling the 1060 for $25.
- Optimizing VRAM Usage: Members discussed freeing up VRAM by plugging the monitor into the integrated GPU, and allocating more VRAM to the 3080.
- It was clarified that integrated graphics donât have their own VRAM and instead reserve a portion of system RAM, and that Nvidia cards cannot combine with integrated graphics, whereas AMD can with vulkan.
HuggingFace â· #general (142 messagesđ„đ„):
Face Recognition Models, Text-to-SQL with T5, HuggingChat Shutdown, ComfyUI and GPU Performance, HairStyle Spaces
- DeepFace Rides to Image Similarity Rescue: Users discussed face recognition models, with one user seeking a model to predict the likelihood of similarity between two faces, and another member suggested DeepFace and linked to a blog post on image similarity and a Face Similarity space.
- Another user recommended checking out this channel for dataset creation: dataset creation.
- T5 Text-to-SQL gets some love and help: A user requested help with a text-to-SQL model on T5-base, and a member provided links to a Medium article, a Hugging Face model, and another T5 model.
- HuggingChat got the boot, plans to evolve: Users noted the shutdown of HuggingChat, and staff confirmed it was temporary with plans to evolve it into something better, also notifying users of a time limit to export Hugging Chat data before it disappears forever.
- One user speculated that Discord is just better, another said Discord doesnât give u access to various models for free.
- Uneven GPUs? No tensor parallelism?: A user with an uneven number of GPUs sought advice on optimizing performance without tensor parallelism, specifically regarding ComfyUI, ReActor, AnimateDiff, and Load AnimateDiffModel.
- Another member suggested that uneven number of GPUs could be solved by splitting the tensors in a non-uniform way between the GPUs.
- Whisper API giving 404s and 503s: A user reported issues with Whisper-large-3 inference endpoints, receiving 404 or 503 errors when using the model catalog configuration, even in the playground, and decided to self-host instead.
- Another user added that default settings for the Hugging Face endpoint itself may not be appropriate and linked to a discussion about specification changes which probably triggered the issue.
HuggingFace â· #today-im-learning (1 messages):
Building Neural Networks from Scratch, Challenges of Custom Neural Network Implementation
- From-Scratch Neural Nets Pose Difficulties: A member conveyed that building a neural network from scratch, without relying on existing libraries, presents a significant challenge.
- Custom Neural Network Challenges: Implementing neural networks from the ground up requires deep understanding and meticulous coding, posing difficulties.
HuggingFace â· #cool-finds (8 messagesđ„):
AI Model Identification, Claude AI Experience, same.dev comparison
- AI Model Selection Under Scrutiny: A user questioned the accuracy of information found on LinkedIn regarding the AI models used, citing the structure and number of AI agents created.
- The user expressed doubt if this information is correct based on a screenshot from LinkedIn.
- Claude AI Gets a Tryout: A user inquired whether another user had experience using Claude AI.
- The user responded positively, noting it was pretty good and somewhat comparable to same.dev but available for free.
- same.dev Competes with Claude AI: Users compared Claude AI with same.dev, highlighting similar functionalities.
- One user specifically mentioned that Claude AI is free, implying a cost advantage over same.dev.
HuggingFace â· #i-made-this (13 messagesđ„):
JauAuth, PiTutor, BorgLLM, RecycloBot, Arena-RLHF
- JauAuth saves the day!: A developer created JauAuth, a secure gateway MCP Router, to mitigate the CVE-2025-49596 vulnerability that allowed attackers to take over developer machines.
- PiTutor Turns Learning into Play: A member introduced PiTutor, a free beta that turns any PDF or doc into an interactive learning session with real-time explanations, highlighting, and a whiteboard.
- BorgLLM Simplifies LLM Provider Configuration: A developer released BorgLLM, a simplified library for configuring LLM providers with automatic fallback and API key switching, installable via
pip install borgllm
. - RecycloBot Identifies Recyclable Items with AI: A member shared RecycloBot, an AI-powered tool that helps users determine if items are recyclable by uploading an image and receiving region-specific disposal tips.
- Arena-RLHF Open Sourced for Human Preference Data: Arena-RLHF built with HuggingFace, an easy way to RLHF on arena-style human preference data (LM Arena, Agent Arena), has been open-sourced.
HuggingFace â· #reading-group (3 messages):
LLM Fine-Tuning Blogs, GPU Parallelism, Transformer Inference
- Parallelism Paradigms Blogpost Posted: A member shared a blog post on GPU parallelism, covering strategies from Data Parallel to Model Parallel.
- The blog post delves into Transformer inference, explaining how each strategy can be used optimally with graphics, math, and code.
- LLM Fine-Tuning Blog Sought: A member asked for suggestions on the best blog regarding fine-tuning an LLM model.
- Another member suggested an easy-to-read white paper.
HuggingFace â· #computer-vision (1 messages):
Deepfake Detection System, AI and Cybersecurity Combination
- Article on Deepfake Detection System Released: A member shared a new article on Medium about combining AI with cybersecurity for detecting Deepfakes.
- The author requested feedback on the article.
- Deepfakes Unmasked in New Cybersecurity Article: A new Medium article, Unmasking the Deceit, explores a deepfake detection system using AI and cybersecurity.
- The author seeks community feedback on their work.
HuggingFace â· #NLP (1 messages):
GLoVE Model, GLoVE Paper, Co-occurrence Probability Symmetry
- GLoVE Model: Symmetry Rule Query: A member questioned why a naive role exchange doesnât solve the symmetry issue in GLoVE modelâs co-occurrence probability, referencing a specific section of the GLoVE paper.
- The query focused on why the probability of co-occurrence of x in the context of k should be the same as k in the context of x, but isnât followed in step 3 of the GLoVE model as described in the attached image.
- GLoVE Model: Clarification on Co-occurrence: The discussion delves into the specifics of the GLoVE (Global Vectors for Word Representation) model and its behavior regarding co-occurrence probabilities.
- The user is seeking to understand the nuances of why simply inverting the roles of words and context doesnât resolve the asymmetry that arises during the modelâs training process.
HuggingFace â· #agents-course (31 messagesđ„):
Time commitment for Agents course, Course assignments and certifications, Access to Llama 3 Model, Issues with Unit 1 Notebook, Guidance on completing Quiz 2.1
- Time commitment for Agents course questioned: A member inquired about the time commitment for the agents course, mentioning they have approximately 4 hours a day for a week.
- They asked whether this is a viable timeframe or if they should explore other options, given their limited availability.
- Agents Course assignments and certifications clarified: A member who just started the course inquired about receiving assignments and certifications, understanding that there are no live streams or running schedules.
- Another member confirmed that this is true.
- Llama 3 Model access sought for genetic AI course: A member enrolled in a genetic AI course sought guidance on how to gain access to the Llama-3.3-70B-Instruct model after their initial request was not approved.
- Another member suggested using the model from Ollama as an alternative, noting its ease of setup.
- Users Encounter Errors in Unit 1 Notebook: Multiple members reported encountering errors in the second block of code in the first unitâs notebook.
- Members expressed discouragement, and requested an update from the HF team on the issue.
- Challenges in Quiz 2.1 Highlighted: A member raised questions about Quiz 2.1, specifically whether to input their HF key in the code and what the 404 error and warning messages mean.
- They requested assistance on how to proceed with the quiz given the persisting issue with using HfApiModel and the explicit instructions provided.
Yannick Kilcher â· #general (180 messagesđ„đ„):
US Copyright Office AI Policy Volumes, Gemini vs ChatGPT for Math, Logical Fallacies in AI Criticism, Material Science, Physics with LLMs
- US Copyright Office Drops AI Policy Trilogy: The US Copyright Office released three volumes on AI and Copyright, covering Digital Replicas, Copyrightability, and Generative AI Training.
- Gemini App struggles with math questions: Members compared Gemini 2.5 Pro in the Gemini App, Perplexity with Gemini 2.5 Pro, and ChatGPT o3 for math question answering, with a question of whether Gemini searches the web implicitly each time.
- It was reported that o3 is mostly good enough, though it does hallucinate from time to time, while o4-mini-high is faster, and pretty reliable.
- Appeal to Authority logical fallacy spreads Disinformation: It was noted that a mechanical engineer with a YouTube video having over half a million views spreads disinformation and political talking points dressed up as AI criticism, using an appeal to authority logical fallacy.
- One member pointed out that 95% of the population by 2035 will be doing social media as their main job and are incentivized to create such content.
- AI to the Rescue for Fundamental Physics: Members discussed how AI can assist fundamental physics beyond ML analysis, with one linking to a YouTube video on the topic, and others noting AIâs use in reducing computational requirements for calculating with quarks, linking to a relevant article.
- One member suggested that Deep Learning could revolutionize the field by making previous algorithms more efficient, enabling things previously impossible due to cost, and also suggested diffusion model to surrogate total blindness with this paper or this paper as a start.
- Cursorâs pricing change: Members discussed that the pricing model in Cursor was adjusted, with many users considering to stop using Cursor.
- It was reported that Cursor team issued an official apology over the pricing change.
Yannick Kilcher â· #paper-discussion (7 messages):
log(n) scaling model exhibit, Hierarchical Reasoning Models, Quaternion products in LLMs
- Log(n) Scaling Model Exhibit Offered: A member offered to present their log(n) scaling model to help clarify their thoughts on it.
- Another member responded with âSure!â, proposing a specific time for the exhibit.
- Hierarchical Reasoning Models Paper Discussion Scheduled: The Hierarchical Reasoning Models (HRM) paper (https://arxiv.org/abs/2506.21734) will be discussed in the #Daily Paper Discussion voice channel.
- The abstract highlights HRMâs novel recurrent architecture, achieving computational depth and efficiency with only 27 million parameters.
- Quaternion Products Experimented in LLMs: A member will discuss their experiments using quaternion products in an LLM to rapidly summarize text, as an alternative to softmax attention.
- This discussion is scheduled for a specific date, offering a novel approach to text summarization.
Yannick Kilcher â· #ml-news (5 messages):
The Ocean is Wet, Critical Look at Chain of Thought, AI Training Load Fluctuations
- AlphaXiv Publishes âOcean is Wetâ Paper: A member shared a post on X and a link to the paper on AlphaXiv titled âThe Ocean is wetâ.
- Chain of Thought Scrutinized: A member noted remembering preprints from a while ago that said this, referring to the recent critical analysis of Chain of Thought.
- The member said itâs good to see some critical light on CoT again.
- AI Training May Blackout the Power Grid: A member posted a link to Semianalysis.com discussing the risk of power grid blackouts due to AI training load fluctuations at a gigawatt scale.
Eleuther â· #announcements (1 messages):
EleutherAI Summer of Open AI Research, Open Science AI Research Project, Mentorship Program
- EleutherAI sets sail for Summer of Open AI Research: The EleutherAI Summer of Open AI Research is a fully online event from August 4 to August 29, 2025, inviting individuals with limited research experience to contribute to open science projects under experienced mentorship (link to EAI discord).
- EAIâs SOS: Seeking Skilled Swashbucklers for Science!: The program is looking for programmers, Masterâs/PhD students, self-taught researchers, and anyone with technical skills wanting to contribute to open science.
- Interested applicants are directed to read the project proposals and complete the application form before the July 21 deadline.
Eleuther â· #general (81 messagesđ„đ„):
AI Alignment & Interpretability, ROPE frequencies for sliding window, Decentralized AI, Language Modeling as Compression, GLoVE Model Symmetry
- EleutherAI Welcomes New Member Keen on Alignment: A new member joined the community expressing interest in contributing to research in AI alignment and interpretability.
- The memberâs interests include AI alignment, interpretability, and safety and looks forward to collaborating.
- Debate Arises over Relative ROPE Frequencies: A member noted that when doing a sliding window decoding, using the same frequencies for RoPE would be wrong because the model is not going to see that the sequence is progressing, causing a looping behavior.
- Another member suggested that using the same frequencies in a sliding window shouldnât matter as the relative distance seen by the attention mechanism should be fine, while another suggested lower base frequency for local attention.
- Enthusiast Advocates for Decentralized AI: A member introduced themselves as passionate about Decentralized AI, exploring how to build trustable, composable agent systems on distributed infrastructures.
- They are happy to connect, learn, and collaborate within the community.
- LLMs as Compressors: Scaling Law Paper?: Inspired by the âlanguage modeling is compressionâ paper, a member trained pythia models (70m to 1.4B) and measured the compression ratio on text, audio, and image datasets.
- They are curious if this is a decent find or a fluke, and what needs to be done to establish this to a novel âscaling lawâ paper, sharing attached plot.
- Community Analyzes GLoVE Model Symmetry: A member inquired about the GLoVE paper, specifically why inverting roles doesnât solve the symmetry problem where the probability of co-occurrence of x in context of k should be the same as k in context of x.
- The paper mentions that a naive exchange of roles wonât solve the problem and the member seeks the reasoning behind it.
Eleuther â· #research (41 messagesđ„):
Aurko Paper, Cognitive Markers, Concat-and-chunk strategy, Flex Attention, CFG
- Mysterious Aurko Paper Sparks Interest: Members discussed the Aurko paper, with one member noting it seemed either totally meaningless or very interesting requiring further investigation.
- Another member suggested DMing Aurko for insights, noting the need for time to discern the paperâs significance.
- Concat-and-Chunk vs Sequence Length Matching: Members debated the use of concat-and-chunk strategy vs. sequence length matching for pretraining data, with one suggesting laziness as a reason for the former, and linking to Bucket iterator, least-squares histogram-packing and Multipack.
- Other members noted that methods like sequence length bucketing have been around since the RNN days, and the choice depends on data properties and implementation details, also mentioning that each packing method introduces different biases.
- Flex Attention: Members discussed Flex Attention, suggesting it is the ideal solution for common usecases due to its ability to mix arbitrary-length sequences dynamically per batch.
- With Flex Attention, thereâs no need for weird slicing up of the data, and rough scheduling is sufficient.
- Stable Training at Scale via Loss Curve Analysis: Members analyzed a paper, finding that scaling curves apply to loss during training and can predict the whole training run, and emphasized that carefully parameterizing the model by scaling the initialization, learning rate, and other hyperparameters is crucial for achieving stable and efficient training at scale.
- It was suggested that if the loss curve is extremely consistent as you scale for compute optimal training after regularization, you can tell if your distributed training setup is improperly configured.
- Attention Beacons During Training?: A member inquired about implementing attention beacons during training, which is slightly different from attention masking, allowing for longer sequence lengths.
- Another member suggested that this would require a very long context for training, unless done in TBPTT (truncated backprop through time) style, which wouldnât require a special kernel.
Eleuther â· #interpretability-general (3 messages):
SAE expansion ratio, Set Autocomplete Model, Publishing Research
- SAE Expansion Ratio Soars with Task Complexity: A member found that increasing the complexity of a set autocomplete model by 4x (training on 20 subsets instead of 5) resulted in a 12x increase in the required SAE expansion ratio.
- The first modelâs SAE needed a 4x expansion ratio, while the second modelâs SAE needed a 48x expansion ratio.
- Impact of Epochs on SAE Expansion: A member suggested that the increase in the SAE expansion ratio might be due to the model being trained for more epochs, in addition to the increased task complexity.
- Publishability of Research: A member inquired about the publishability of their findings regarding the relationship between task complexity and SAE expansion ratio.
Eleuther â· #lm-thunderdome (7 messages):
MMLU-SR task subsets, datasets parquet conversion, lm eval running time
- MMLU-SR task subsets are not working: A member reported that they could no longer evaluate models using subsets of the
mmlusr
task due to aValueError
indicating that theBuilderConfig
was not found.- Another member pointed out that the subsets got deleted when they converted the dataset to parquet on the hub and suggested using a previous commit by adding a
revision
to thedataset_kwargs
in the taskâs YAML file.
- Another member pointed out that the subsets got deleted when they converted the dataset to parquet on the hub and suggested using a previous commit by adding a
- Datasets parquet conversion caused errors: A member mentioned that converting the datasets to the parquet format on Hugging Face Hub resulted in the deletion of subsets, causing errors when running
lm_eval
.- They suggested specifying a previous commit hash in the
dataset_kwargs
of the task configuration file as a workaround; as they linked a relevant commit.
- They suggested specifying a previous commit hash in the
- Lm eval stucks for long time: A member reported that
lm_eval
gets stuck for around 20 minutes before starting the evaluation process and attached a screenshot.- The user was using 2 GPUs and the
parallelize
argument, but did not show the exact command that was used for the evaluation.
- The user was using 2 GPUs and the
Eleuther â· #gpt-neox-dev (6 messages):
Small datasets for training, FineWeb dataset samples, FA3 support for H100/H200, dclm-dedup dataset
- Smallish Datasets Sought for Training Validation: A member was looking for a smallish dataset around 50B tokens to validate some things, asking for recommendations on subsampled datasets in that size range.
- They mentioned that FineWeb samples are either much smaller at 10B or double the size at 100B.
- FA3 Support Status Queried for H100/H200: A member asked for updates on FA3 support for H100/H200, recalling previous mentions of an impending merge but noting its absence in the repo.
- They were seeking clarification from the community regarding the timeline for this feature.
- âdclm-dedupâ Dataset Subset Shared: A member shared a 25B token subset of the Zyphra/dclm-dedup dataset, available at Hugging Face, as a possible option.
- The member acknowledged that this might be smaller than what was initially requested, but offered it as a resource.
- ânemo_idâ Column Clarification Requested: A member inquired about the meaning of the
nemo_id
column within the dclm-dedup dataset.- Speculating it could represent a global shard/local shard mapping, they sought confirmation from the dataset creators.
Nous Research AI â· #general (101 messagesđ„đ„):
HF documentation 3B model, LLM Matching Generative Responses, Grok's Politically Incorrect Stance, PiTutor interactive learning, Grok's Knowledge Updates
- Hugging Face Rumored to Launch 3B Model: According to the HF documentation, there will be a 3B model, which may come in different sizes.
- LLM to Match Generative Responses: Members discussed using an LLM to match generative responses with reference answers using this X summary for a quick overview.
- Debate heats up around Grokâs Political Incorrectness: A user posted a screenshot implying Grok is intentionally politically incorrect and another mentioned this Elon Musk post.
- Another user felt it was damaging trust in AI, with someone else responding that Grokâs prompts are available on Github.
- PiTutor Turns PDFs Into Interactive Learning Sessions: A member shared PiTutor, a tool that turns any PDF or doc into an interactive learning session with real-time explanations, highlighting, and a whiteboard, available in free beta here.
- Discussion around Vector Space Refinement: A member proposed refining vector space representations to be more human before the embedding matrix is calculated.
- Another member noted that if you want to train them, youâd have to train them.
Nous Research AI â· #ask-about-llms (18 messagesđ„):
Ollama and Openwebui for AI Services, Training LLMs on Their Own Weights, Temperature and Token Usage, Math 500
- Ollama and Openwebui Supercharge AI Services: A member is using a 4080 Super to provide AI services through Ollama and Openwebui and is planning to switch to Llama CPP.
- Theyâre looking for ways to help Nous Research by putting their models in front of their user base for live testing.
- LLMs Ponder Weighty Matters: A member inquired about training an LLM on its own weights alongside additional data to induce a deeper understanding of a subject.
- Another member responded this this could be similar to continued training and weight merging, but can cause catastrophic forgetting and mixing in some of the original data can help prevent that.
- Temperature Impacts Token Count: A member is running an experiment to determine which temperature setting uses the least tokens for language generation.
- They mentioned using Math 500 as a small dataset benchmark.
- Math 500 Benchmark Surfaces: A member suggested using the Math 500 dataset as a benchmark for token usage experiments, aiming for a dataset smaller than 1k questions.
- The member was inspired by a bot repeating messages and giving a spark of inspiration, joking heh.
Nous Research AI â· #research-papers (3 messages):
AI Mouse Tracking, Architecture for Mouse Path Training
- Seeking Model Architecture for AI Mouse Tracking: A member is seeking advice on the appropriate architecture to train a model using a dataset of natural human mouse paths with images of clicked zones and textual prompts representing actions (e.g., moving the mouse to the + icon, scrolling, screenshot).
- This indicates an interest in leveraging AI to understand and predict human-computer interaction patterns.
- Dataset Details: Mouse Paths, Images, and Textual Prompts: The dataset comprises natural human mouse paths paired with images of the zones clicked and textual prompts describing the action performed.
- This rich dataset aims to capture the nuanced relationship between user intent (textual prompts), visual context (images), and behavioral patterns (mouse paths).
Nous Research AI â· #interesting-links (8 messagesđ„):
Chinese AI investment, Parameter-efficient fine-tuning, Trustless agents, Codeium 3.2, Autonomous optical network
- Zhipu Secures Yuan Shower from Shanghai State: Chinese AI company Zhipu secured a $1.4 billion strategic investment from Shanghai state funds, as reported by Technode.
- Pi Tutor Builds Personalized AI Tutor: Pi Tutor is building a better future with visual interactions, turning any content into a personal tutor.
- The service speaks your language and gives you tailored examples, more at their website.
- Fine-Tuning Costs Cut Massively: June saw the publication of 117 papers introducing parameter-efficient methods that cut fine-tuning costs by â85% while preserving ~98% accuracy, according to Konceptual AI.
- Trustless Agents Spark Excitement: Reddit users highlight progress toward âtrustless agentsâ using TEEs, zero-knowledge proofs, and on-chain verifiability to ensure integrity and privacy, detailed in this Reddit thread.
- Codeium 3.2 Boosts Multi-Repo Savvy: Codeium 3.2 adds multi-repo awareness (70+ languages), reducing boilerplate by ~40%, while GitHub Copilot Chat Enterprise v2.1.5 delivers real-time vulnerability scanning and â„92% JUnit test coverage, according to AI Agent News.
Nous Research AI â· #research-papers (3 messages):
AI-Youtube-OG Architecture, Natural Human Mouse Paths Dataset, Model Training Architecture
- Query about AI-Youtube-OG Architecture: A member referenced an âAI-Youtube-OGâ and shared a thank you for sharing information about it, noting they had forgotten about it. đ
- It might have something to do with the dataset of natural human mouse paths with images of the zones of where they clicked.
- Seeking Architecture for Training on Mouse Path Data: A member inquired about the appropriate architecture for training a model on a dataset of natural human mouse paths with images of the zones where they clicked.
- The dataset also includes textual prompts representing the action (e.g., moving the mouse to the + icon of the web browser, scrolling, screenshot, etc.).
aider (Paul Gauthier) â· #general (93 messagesđ„đ„):
Grok 4 Leaks, Deepseek R2, Gemini-CLI Search Integration, MCP Integration with Aider, Documentation Fetching MCPs
- Grok 4 Leak Sparks Speculation: An alleged Grok 4 leak image surfaced, showing it being trained on the HLE benchmark.
- Members expressed excitement about potential leaps in performance and the benefits of competition in the LLM space, while some are skeptical and think Grok will refuse to work on your project if your git branch named âmainâ, instead of âmasterâ.
- Gemini-CLIâs Search Integration Impresses but Deletes Secrets: One member testing gemini-cli highlighted its search integration for picking up an update to a payment portal, which they hadnât heard about yet, finding it great for beginners.
- However, it broke a bunch of code (deleted the secrets?) without explanation, leading to a desire for MCP integration with Aider for documentation fetching.
- Open-Weight 70B Models Face Scrutiny: Community members discussed the landscape of open-weight 70B models, citing Llama 3.1, Qwen2.5 70B, and Mistral Large as notable options, but seek more recent updates.
- It was mentioned that Qwen3-32b (dense, they are all hybrid reasoning for 3.0) should perform better than qwen 2.5 72b, also mentioning how Meta dropped the ball on this not having many recent OS options sub-100b.
- Concerns Arise Over Benchmark âGamingâ by AI Labs: Members discussed the reality of AI labs gaming benchmarks, with acknowledgment that preventing contamination is impossible.
- The focus should be on creating cheat-resistant benchmarks and interpreting results with variety, as contamination leads to improved models and their ability to generalize.
- Aider users wrestle with replace issues: Some Aider users reported that Aider performed replaces on older versions of the file when they made edits, and that Markdown files had the most issues.
- One user recommended clearing the conversation history with
/clear
, while another one suggested dropping and re-adding the file, and/map-refresh
.
- One user recommended clearing the conversation history with
aider (Paul Gauthier) â· #questions-and-tips (25 messagesđ„):
InputOutput class in aider, OpenRouter provider settings, Git reset, sonnet-4 with ask/code, deepseek 70b
- Customize Aider Output with InputOutput Class: A user discovered the
InputOutput
class fromaider.io
to customize output and setyes=True
, with a note that it could potentially change text color for numbers/bullets in lists outputted by the assistant, see aider.io.- The key code mentioned was:
from aider.io import InputOutput
io = InputOutput(yes=True)
- Selecting OpenRouter Provider via .env: A user inquired about setting a specific provider for a model through OpenRouter in a
.env
config, tryingAIDER_MODEL=openrouter/deepseek/parasail-deepseek-r1-0528-qwen3-8b
.- A member suggested configuring the provider in OpenRouter settings or params via
extra_body
, referencing OpenRouterâs documentation on provider routing.
- A member suggested configuring the provider in OpenRouter settings or params via
- Recovering Lost File with Git Reset and Aider History: A user faced an issue where
app.py
was not in the repository in the previous commit and the undo failed, and sought advice on how to recover the file.- Suggestions included using Ctrl-Z in nvim, using
git reset
(potentially deleting the file), or reconstructing the file fromaider.history.md
. A member emphasized the importance of committing changes.
- Suggestions included using Ctrl-Z in nvim, using
- Assessing DeepSeek 70B Performance: A member reported disappointing results with DeepSeek 70B, citing its poor performance on tasks like âreading a file and writing unit testsâ, compared to paid models.
- They questioned whether anyone had achieved even a fraction of the capability of closed-source models with locally running models, even with sufficient tool calling setup.
GPU MODE â· #general (3 messages):
Tensor Compilers, CUDA, PyTorch, Vector Matrix Multiplication
- Tensor Compilers Project Kicks Off: The Tensor Compilers: Zero to Hero work group is seeking contributors interested in building CUDA and PyTorch from scratch.
- The project aims to be both pedagogical and hardcore, focusing on tensor compilers.
- Batching Boosts Vector Matrix Throughput: A member inquired whether batching vectors into matrices is the primary method for improving throughput in vector matrix multiplication.
- This is still an open question in the channel.
GPU MODE â· #triton (1 messages):
einops, einstein notation, triton
- Seeking EinOps in Triton: A member inquired about a Triton implementation of einops or some form of Einstein notation.
- Triton tensor contractions: The user is looking for ways to use einops in Triton to help with tensor contractions.
GPU MODE â· #cuda (36 messagesđ„):
CUDA printf debugging, NCU memory bandwidth, GPU division vs multiplication, CUDA tile scheduler vs cutlass, CUDA unified memory
- Printful CUDA Debugging: Members discussed using
printf
for debugging CUDA kernels, noting that output is buffered and requires explicitcudaDeviceSynchronize()
to flush, and some had preprocessor directives guarding arch-specific code.- One member resolved a ânoob issueâ caused by compiler preprocessor directives guarding arch specific code.
- NCU Reveals Compute-Bound Woes: A user reported that after fixing uncoalesced writes flagged by NCU, memory bandwidth utilization and runtime remained flat, which was unexpected.
- They later discovered that division in a loop was the bottleneck, making the kernel compute-bound, and replacing division with multiplication by the reciprocal increased throughput by 33%.
- GPUs Sidestep Division Circuitry: A member noted that GPU ALUs may not have dedicated division circuits, instead relying on division approximations, as multiplication is done via additions.
- AFAIK GPU ALUs donât even have a circut for division, they do a division approximation.
- Custom CUDA Schedulers: A member inquired about the feasibility of implementing a custom tile scheduler in CUDA instead of using Cutlass.
- A suggestion was made to explore customizable swizzles in Cutlass as a potential solution, if that helps.
- WSL CUDA Lacks Unified Memory: A user questioned whether setting up a dual boot for CUDA development is justified due to WSLâs lack of support for unified memory and certain nsys performance counters.
- It was suggested that avoiding unified memory might be preferable for performance optimization, and that WSL already oversubscribes normal device memory via WDDM.
GPU MODE â· #jobs (1 messages):
ML Platform Hiring, Remote Work, Quora
- Quora Machine Learning Platform Team Seeks Senior/Staff Engineer: Quora is hiring a Senior/Staff Software Engineer (L5-L6) for their Machine Learning Platform team, offering a fully remote position within the US and Canada; apply here.
- Exciting ML Infrastructure Awaits: The role involves exciting ML infrastructure work, providing an opportunity to contribute to Quoraâs machine learning capabilities.
GPU MODE â· #rocm (2 messages):
ROCm traces, in-person hackathon, make flags, gpumode-amd-fp8-mm repo
- Parisian Builder Joins ROCm Chat: A builder from Paris introduced themself in the channel seeking ROCm traces and ways to place in the in-person hackathon happening today in Paris.
- The builder shared a link to gpumode-amd-fp8-mm repo and inquired about helpful make flags.
- Flags Question: A builder asked which make flags you find useful.
- They are currently using these make flags
GPU MODE â· #self-promotion (4 messages):
voxel raytracing, Hopper Pipelining, PiTutor, Deep Infra B200 Instances
- Voxel Raytracing Video Drops: A member uploaded a video about visibility-based chunk selection for voxel raytracing, covering internals of victim pointers and usage bits.
- The open source code is available on GitHub.
- CuTeDSL Pipelining on Hopper Blogged: A member shared a blog post on overlapping memory transfer and computation via Pipelining on Hopper, featuring TMA and WGMMA atoms.
- Example code is available on the Cutlass GitHub, with a Colfax blog discussing pipelining with the C++ API for CuTe.
- PiTutor Turns PDFs Into Learning Pal: A member introduced PiTutor, a tool that turns any PDF or doc into an interactive learning session with real-time explanations and a whiteboard, available in free beta here.
- The idea is simple: what if learning didnât feel like a chore?
- Deep Infra Deals On B200: Deep Infra is offering on-demand B200 instances at $1.99 / h with 1-click deployment, according to a member, details here.
GPU MODE â· #general-leaderboard (5 messages):
GPU Mode Challenges, Leaderboard Questions, Kernelbot Data
- GPU Mode Challenges Revamped Next Week: Beginner challenges are finished, but will be recreated soon-ish (next week) with proper evaluation code, so members can use the AMD challenges or trimul until then.
- Leaderboard Question Solutions Released: Solutions to the leaderboard questions that already ended are already out and available on HuggingFace.
GPU MODE â· #submissions (6 messages):
A100 Performance, H100 Performance, Trimul Leaderboard
- A100 Achieves Fifth Place: A submission on an A100 GPU secured 5th place on the
trimul
leaderboard with a time of 20.6 ms. - H100 Scores Second Place: A member achieved second place on the
trimul
leaderboard multiple times using an H100, with times of 60.0 ms, 43.5 ms, and finally 35.9 ms. - H100 Success on Trimul: Additional submissions on H100 were successful on the
trimul
leaderboard, both clocking in at 63.1 ms.
GPU MODE â· #factorio-learning-env (41 messagesđ„):
instance.py refactoring, Github Actions integration, Ruff linting, Pydantic bump PR
instance.py
Too Long; Needs Restructuring: Members discussed that instance.py is too long and needs restructuring, potentially into main, which is related to issue #249.- Github Actions Testing Integration Progresses: A member confirmed their Github action to publish to testpy is working, triggered on push, and they will set it up to publish to testpy on merge along with automatic patch increase.
- The member provided a screenshot of the Github Actions workflow showing the successful run.
- Ruff Linting on the Horizon!: The restructuring PR has been merged and the team is turning to how to get ruff linting merged in.
- They may use
--ignore-revs-file
&.git-blame-ignore-revs
to keep git blame sensible for the bulk changes.
- They may use
- Pydantic Bump PR Causes Headaches: Concerns were raised that the Pydantic bump PR involved an unusually large number of file changes (70 files).
- A member mentioned they needed to merge main into their Pydantic PR to resolve the large number of commits (95) in that PR.
GPU MODE â· #singularity-systems (4 messages):
picoc compiler for llm.c, picograd kernels, nn: zero to hero follow up, picoc frontend semantic analysis
- Picoc sets sights on compiling llm.c: The project goals for picoc are now focused on compiling karpathyâs llm.c by the end of the year, utilizing raw CUDA and CUTLASS for optimization.
- The initial plan involves tackling CUDA first, with CUTLASS integration to be figured out later.
- Picograd prioritizing Pytorch1-style kernels: The picograd project is prioritizing the development of Pytorch1-style kernels, with a completed forward pass on Karpathyâs MLP on CPU.
- A member mentioned their progress on X, and expressed that creating a Pytorch2 fusing compiler to triton-style compiler will need to be a v2 of the course, and is unlikely to happen this year.
- Workgroup Aims for nn: zero to hero Follow-up: The workgroup aims to create a direct follow-up to nn: zero to hero, focusing on the software 2.0 equivalent of building an interpreter/compiler.
- The models directly from nn: zero -> hero will be converted by swapping the
import torch
toimport picograd
.
- The models directly from nn: zero -> hero will be converted by swapping the
- Call for Assistance in Picocâs Frontend Semantic Analysis: A member requested assistance with the frontendâs semantic analysis for C0 within the picoc project, particularly in the areas of lexing/parsing/typing.
GPU MODE â· #general (1 messages):
tonic_1: đȘđ» đ đ«đ· đ đ
Latent Space â· #ai-general-chat (100 messagesđ„đ„):
Cursor.ai Pricing Controversy, Trae-Agent Open-Sourced by Bytedance, ChatGPT Diagnoses Rare Genetic Defect, ChatGPT's New 'Study Together' Feature, Books3 Dataset Lore
- Cursorâs Unlimited Pricing Pulled, Users Protest!: Users are accusing Cursor.ai of a rug pull due to a sudden shift from a previously âunlimitedâ pricing model to a limited one, resulting in unexpected charges for Sonnet-4 usage.
- Some users reported issues with using their own API keys, receiving 500 errors, while Cursor clarified their pricing stating âunlimited usageâ was only for Auto and not all other models and offered refunds.
- ByteDance Releases Trae-Agent into the Wild!: Trae AI open-sourced Trae-Agent, its agent powering their IDE, now available on GitHub and ranked #1 in SWE-bench Verified.
- The company is seeking contributions to building an open Agent ecosystem, with Trae-Agent supporting OpenAI and Anthropic keys and making it easy to hack it out to open router.
- ChatGPT Cracks Medical Mystery Doctors Missed!: A viral Reddit story highlights how ChatGPT identified a hidden gene defect (methylation block) doctors missed for a decade, leading to significant symptom improvement.
- The thread discusses AIâs increasing role in healthcare, particularly for obtaining second opinions and personalized medicine, with new AI systems outperforming doctors in diagnostic accuracy by using a RAG system on scientific papers.
- ChatGPTâs âStudy Togetherâ Feature Surfaces: A Twitter thread discusses a new âStudy Togetherâ feature in ChatGPT, possibly an internal prototype codenamed âtatertot.â
- Speculation surrounds its function, ranging from AI-powered group study rooms and a replacement for flashcards to a personal AI tutor within a group setting, or a tool for collaborative discovery with AI.
- Geminiâs Batch API Offers Bulk Savings: Logan Kilpatrick announced the launch of âBatch modeâ in the Gemini API, offering 50% discounts on 2.5 models and enabling the processing of billions of tokens.
- The announcement was met with positive feedback from users.
Modular (Mojo đ„) â· #general (41 messagesđ„):
Mojo for research and training, Arrow spec, Mojo and Python callbacks, Mojo server tag, Mojo as a Python superset
- Mojo Eyes AI Inference, Not Training: While some were asking if anyone is using Mojo for research and training, it was mentioned that Modular is pretty focused on the inference side of things at the moment.
- It was suggested to check out Mojmelon on modular-community on GitHub, and a link to Nabla, which is faster than JAX for both compile time and training time.
- Scientists Still Love Excel: Molecular scientists still manually parse and generate tables, and Excel is used in unexpected ways, such as for a full aligner / primer design engine.
- The industry should be migrating to Arrow, and the Arrow project has interesting prototypes like arrow dbc, where the data is sent directly as arrow data by the databases that support it.
- Mojo Extends Python, Doesnât Supercede: Despite the high demand to make Mojo a true superset of Python, itâs being used to extend Python.
- As Modular CEO Chris Lattner said, Mojo has been laser focused on solving the AI stack issue and creating offerings in that space.
- Mojo Roadmap Updates Soon: Someone asked if Modular released a roadmap or something along those lines, and a community member responded that the most recent Mojo roadmap was published here.
- The community member also mentioned that they should have some updates dropping soon.
Modular (Mojo đ„) â· #mojo (52 messagesđ„):
StringLiteral Materialization, Parametric Traits vs Trait Objects, Cloud-Based Mojo Environment, Static Linking in Mojo, SIMD Bitcasting
StringLiteral
Materialization Breaks: A member reported that recent changes required to makeStringLiteral
materialize to aString
have broken existing code.- They are planning to open an issue and follow up with the compiler folks.
- Confusion About Parametric Traits and Trait Objects: Members discussed the possibility of parametric traits in Mojo, with one member providing a code snippet to achieve a similar effect using trait objects.
- Another member clarified that this isnât parametric traits, this is trait objects.
- Request for Cloud-Based Mojo Environment: A member inquired about setting up a cloud-based Mojo environment for research due to hardware limitations.
- Another member suggested using any GPU instance that runs Linux and lets you install pip packages.
- Static Linking Feature Requested: A user sought to compile Mojo code without shared library dependencies, specifically
libKGENCompilerRTShared.so
, for deployment on a remote machine.- A member suggested that full static linking would need to be a feature request and pointed to existing discussions and a related GitHub issue ([BUG]: Cannot build statically compiled binary #1317) and a new [Feature Request] Static Compilation of pure mojo code #4976.
bitcast
Behavior Change: A member reported that a previously workingbitcast
operation involvingSIMD
andDType
no longer compiles, and sought guidance on the correct way to perform the conversion.- The member found that adding
raises
to functions prevents the compiler from doing tail call optimization, and linked to a Godbolt example [https://godbolt.org/z/ahn4144a7].
- The member found that adding
Modular (Mojo đ„) â· #max (5 messages):
Mojo GPU Puzzles, ModuleNotFoundError, Pixi
- Debugging
ModuleNotFoundError
in Mojo GPU Puzzles: A member encountered aModuleNotFoundError: No module named 'max'
error while running mojo-gpu-puzzles from Python.- Another member suggested using
pixi add max
orpixi shell
in the puzzles directory to install the necessary dependencies, referencing the pixi.toml file.
- Another member suggested using
- Pixi Workflow for Mojo GPU Puzzles: A member suggested the typical workflow is to git clone the repository, then run pixi shell to install all dependencies.
- The member had tried
pixi add max
, but this did not resolve the issue, prompting further troubleshooting.
- The member had tried
MCP (Glama) â· #general (58 messagesđ„đ„):
EpicMe MCP, WinCalcMCP, MCP Python Interpreter, MCP for documentation indexing, LangGraph vs Custom Agents
- EpicMe MCP Provides Single Endpoint: A member requested a self-hosted docker image that acts as an MCP aggregator/gateway, and another member suggested EpicMe as a solution.
- The creator stated that EpicMe covers onboarding in a very unique and cool way, providing a link as an example.
- WinCalcMCP Connects Claude to Windows Calculator: A member built their first MCP server, WinCalcMCP, that connects Claude Desktop to Windows Calculator to improve math answers.
- Another member prefers giving it a full Python interpreter as a general tool via mcp-python-interpreter so they do not have to manage so many specialized tools.
- LangGraph not a great abstraction, use custom agents: One member shared that their journey started with LangGraph (Langchain actually) then moved to not use it because honestly they did not find it a great abstraction, since then theyâve been coding their own agentic framework.
- It was agreed that the beauty of MCP is that there is no need for people to use the same agentic framework but things can still communicate using a common protocol.
- MCPdata does local documentation indexing for MCP: A member introduced MCPdata from MaheshDoiphode/mcpdata, for local documentation indexing for MCP.
- It seems like context7 doesnât keep up with the latest from the vercel
ai
api docs.
- It seems like context7 doesnât keep up with the latest from the vercel
- Official MCP servers are the only ones worth using: Members complained about most of the MCP servers being useless.
- One member suggested using the glama.ai website to only get servers from official sources such as Elasticsearch, Kagi, and Redis.
MCP (Glama) â· #showcase (4 messages):
MCP Monetization, Agentic Payments with Crypto, Fast Agent's MCP Elicitation Support, EpicAI's MCP Search Engine
- MooPoint Makes it Easy to Try Services: MooPoint.io now allows you to try out the service right when you log in for the first time.
- Monetize MCP Servers with Crypto: A member is exploring ways to monetize MCP servers, suggesting that cryptocurrencies and stable coins are a great fit for agentic payments, sharing a simple tool to turn a free MCP into a paid one using bitcoin payments: lmcp.
- They are looking for a partner to create a showcase.
- Fast Agent Gets Comprehensive MCP Elicitation Support: Fast Agent now has comprehensive MCP Elicitation support, as well as a quickstart guide that makes getting started with Elicitations Servers really simple, according to this blogpost.
- EpicAI Launches an MCP Search Engine: A member shared EpicAIâs MCP Search Engine.
Notebook LM â· #use-cases (16 messagesđ„):
Google Docs comments, NMC Standards Notebook, NHS adoption of NotebookLM, Mindmap embedding, Newsletter drafting from audio files
- Nurse Creates NMC Standards Notebook with NotebookLM: A Registered Learning Disabilities Nurse and Practice Educator within the NHS created a digital NMC Standards Notebook consolidating key documents from the Nursing and Midwifery Council (NMC) into a single, searchable, and interactive resource, available at notebooklm.google.com.
- The tool supports nurses, midwives, students, and educators by integrating the NMCâs framework, streamlining revalidation, supervision, and preceptorship processes, and has been adopted across education networks to embed person-centred care, clinical safety, and inclusive practice.
- NHS Scales NMC Notebook Nationally with Free Accounts: An NHS Registered Learning Disabilities Nurse shared that due to the âshare publiclyâ feature, a tool they created is now being used nationally in NHS settings for revalidation and preceptorship despite being on a free Google account.
- The project started as a local educational aid but it has grown.
- Users Discuss Reading Google Docs Comments: Users are having trouble reading Google Docs comments with NotebookLM; one user asked if the feature is coming?
- One member suggested a workaround that involves downloading the Google Doc as an MS-Word document, then printing the Word document to a PDF, and then using that PDF as a source.
- Newsletter Drafting From Audio Files: A user asked for any suggestion/advice on using NotebookLM to draft newsletters from audio files of their recordings.
- No ideas were given.
Notebook LM â· #general (36 messagesđ„):
Saving Notes in NotebookLM, Interactive Mode Button missing, PDF Upload issues, Saving Chats in NotebookLM, New Model for NotebookLM
- NotebookLM Auto-Saves Notes Like a Charm: A user expressed excitement about uploading their school notes and experiencing how well NotebookLM works, while other user asked how to save on notebook.
- Another member clarified that the âNotesâ Section at the bottom right includes a â+ Add Noteâ button, making the process easy and seamless.
- Interactive Mode Button Vanishes for Some: A user reported that the Interactive Mode button is missing from their notebooks, despite having a pro subscription.
- Another user suggested setting the output language to English, as the interactive mode is currently only available for Audio Overviews generated in English. It seems the customize audio button has been replaced by the Interactive Mode Button.
- PDF Upload Roadblocks Encountered: A user is facing issues uploading a 19.1 MB PDF file to NotebookLM, despite following tutorials and checking the resources available.
- A member asked if they had googled the question, and another member asked if notebookllm app working for anyone
- Chat Saving Capabilities Debated: Users are frustrated about the inability to save chats within NotebookLM, noting that chats disappear after a short period.
- One user pointed out a workaround: manually saving each answer as a note, but acknowledged itâs not the same as saving the entire chat log.
- NotebookLMâs Data Storage and FERPA Compliance Queried: A user is seeking clarity on the safety and security of NotebookLM for medical students, specifically regarding data storage locations, monitoring of user interactions, and FERPA compliance.
- No answers were provided by the community or staff.
tinygrad (George Hotz) â· #general (50 messagesđ„):
tinygrad vs pytorch, MLIR, Halide, Exo-lang, whisper example
- Tinygradâs Path to Victory: George Hotz defines winning for tinygrad as being the fastest way to run things, expanding beyond QCOM GPU to AMD and NVIDIA.
- Others pointed to the scale of innovation and fine-tuning in PyTorch, with chip manufacturers directly contributing to hardware lowering layers in projects like Triton.
- MLIRâs Role Debated: One member thinks that achieving performance close to hand-written CUDA requires the Triton MLIR route, leveraging its deep hardware knowledge and the LLVM compiler infrastructure.
- George countered that MLIR is Waymo, advocating for an end-to-end approach in neural networks, whereas Waymo uses a different approach.
- Halide Framework Praised and Compared: George noted that Halide has a clearer structure than MLIR and TVM, emphasizing a single, fast, and correct way to perform computations across all datatypes and backends, focusing on schedule optimization for speed, and points to the Halide paper for reference.
- Another member pointed out a comparison from Exo-lang, which compares itself VERY favorably over Halide, and provides links to their GitHub and ArXiv paper.
- Revving up whisper example: A member is getting back to the whisper example, seeking a speed boost, and another shares that their branch has all audio preprocessing rewritten in tinygrad.
- They are looking for a replacement for their current streaming transcribing setup of VAD+whisper.cpp, finding the tinygrad implementation more approachable.
- tinygrad aims to Commoditize the Petaflop: Members discussed aiming for interoperability with different hardware, viewing tinygrad as producing fine-tuned code via UOP graph optimizations in a hardware-agnostic way.
- One member stated that the project is still cooking, and they are excited to see if they get AMD on MLperf according to the contract specs by AMD, to get it competitive with Nvidia.
DSPy â· #general (31 messagesđ„):
Automatic Prompt Optimization (APO), Claude Code vs DSPy, DSPy 3.0, SIMBA vs MIPROV2, Tool Selection for LLMs
- APO Relevance in Evolving Models: A team is building an AI agent for financial decisions and is considering investing time in Automatic Prompt Optimization (APO), questioning its relevance as models improve.
- Community members are debating whether APO is still worth the deep dive.
- Claude Code Challenging DSPy?: The community discussed whether Claudeâs code capabilities might replace DSPy, referencing a tweet as possible evidence.
- One member expressed interest in understanding the underlying mechanisms, especially if Claude is autonomously optimizing prompts, suggesting a more structured approach might be beneficial.
- Legal Concerns with Claude Code: Some members expressed concerns that using Claudeâs code might not be permissible in large companies due to codebase trade secrets.
- Another member countered that their big tech company has been using Claude code since its release, though it might be different for non-tech companies; still, itâs worth linking DSPyâs Tweet here.
- Optimizing Tool Selection for LLMs: A community member shared a blog post on optimizing tool selection for LLMs, asking whether something like this can be trained natively as a DSPy module end to end.
- The community also inquired about using DSPy Version 3.0.
- DSPy 3.0 DAIS Talk: A member shared a link to their DAIS talk on DSPy 3.0 (beta), available on YouTube.
- It was also noted that the videos could be made public in one week.
DSPy â· #examples (6 messages):
DSPy quickstart, A/B testing prompts in DSPy, signature.py exploration
- Quickstart avoids Prompt Docstrings: A member shared a quickstart that avoids prompt docstrings and class modules, also available as a gist.
- Another member found this a cleaner way especially with longer starting prompts migrated from non-dspy workflows.
- A/B testing Prompts considered Faux Pas: A member mentioned they often use this approach to A/B test prompts despite it being a big faux pas with DSPy.
- They discovered the approach by exploring signature.py.
Manus.im Discord â· #general (32 messagesđ„):
ChatGPT personality, Claude 4 Comparison, Manus Airdrops, Manus Credits, Manus as project starter
- Manus mistaken for exhibiting ChatGPT personality: A member questioned whether Manus was using ChatGPT because of its increased personality and use of emotes.
- Another member suggested it might be Claude 4, noting it exhibits similar behavior.
- Claude 4 Similarities Noted: A member compared Manus to Claude 4, suggesting they exhibit similar behaviors.
- The member stated it does the same stuff.
- Doubts Cast on Manus Airdrop Capability: A member asked if Manus AI could launch airdrops.
- Another member replied nop.
- Token Woe as Fixes Drain Credits: A user complained that errors in their projects drained all 3000 credits and they couldnât launch anything.
- They suggested changing the credit request parameters so projects can be completed and launched.
- Manus recommended as project starter instead of delivering working projects: A member recommended using VS Code and other tools, suggesting Manus is better at starting projects than delivering working projects.
- Another member reported experiencing a Network connection error with Manus.
Torchtune â· #general (3 messages):
QLoRA throughput, Fused RMSNorm, Linear cross entropy, LoRAMLP
- QLoRA Throughput Throwdown: Tune vs. Unsloth: Members inquired about recent comparisons of QLoRA throughput and memory usage between Tune and Unsloth, due to potential upstreaming of fused RMSNorm and linear cross entropy in PyTorch.
- They think these features could bring Torchtune closer to Unsloth in optimization for LoRA, except for the fused LoRAMLP kernel in Unsloth.
- PyTorch-Only LoRA Implementations Closing the Gap?: Members expressed curiosity about how close the PyTorch-only (no custom Triton) implementations get to Unsloth with each new feature.
- One member recalled that comparisons using compile from last year showed performance was pretty close and in some cases better.
Torchtune â· #dev (5 messages):
Custom Tokenizers, Mistral 2506-small, HF AutoTokenizer, CI Bug
- Custom Tokenizers Worth Maintaining: Members discussed whether to maintain custom tokenizers, especially since Mistral doesnât supply an HF tokenizer with their latest 2506-small model.
- One member suggested having a fallback that uses the HF
AutoTokenizer
and integrates with features like packing.
- One member suggested having a fallback that uses the HF
- Mistral Requires Extra Library: A member noted that Mistral requires an extra library, deeming its tokenizer âunfortunate.â
- The member suggested using HFâs AutoTokenizer as a fallback option, and wondered whether it can be made compatible with torchtuneâs features.
- CI Bug Alert: A member reported a new bug in CI (https://github.com/pytorch/torchtune/pull/2822).
- Another member acknowledged the bug, describing it as âpretty minorâ and promising a fix by the next day.
Torchtune â· #papers (14 messagesđ„):
Context-Parallel vs. Limited GPUs, Skepticism on Architectural Improvements, Data Cleaning vs. Architecture Iterations, MoE Training Techniques
- Context-Parallel Scales vs. GPU Limitations: A member discussed their preprint which addresses scaling sequence length with limited GPUs, contrasting it with context-parallel approaches that require more GPUs for throughput.
- They suggest that for sequence lengths >= 500k, context-parallel may be better, but for very high sequence lengths (millions), their method could be superior, highlighting a need for benchmark comparisons.
- Architectural Improvements face Skepticism: A member expressed skepticism towards recent architectural improvements, suggesting that many reported results are driven by variance and hyperparameter optimization rather than fundamental advances.
- They argued that investing in data cleaning is more effective than pursuing architectural iterations since the first transformer, advocating for cleaner datasets over complex model modifications.
- Clean Data Beats Architecture: A member generally agreed that cleaning data can lead to greater abilities than modern LLMs have achieved.
- He noted his enthusiasm for SSM papers and followed those through to Mamba 2 and that mostly died, so my enthusiasm may be misplaced.
- MoE Training Technique Analysis: A technique and results for MoE training were mentioned, questioning if similar results could be achieved more cheaply without needing a dense fwd pass, as detailed in a Notion post.
- Another member shared a figure noting that linear scaling has lotâs of trade offs.
Cohere â· #đ§”-general-thread (4 messages):
Cohere Labs, Open Science Community Summer School
- Cohere Summer School Content now Available: The Cohere Labs Open Science Community Summer School contents are now available on a YouTube playlist.
- Getting Started with Cohere: New members can find resources and content from the Cohere Labs Open Science Community Summer School on YouTube.
Cohere â· #đ-api-discussions (2 messages):
Embed v4 API, Hybrid image/text embeddings
- Embed v4 API for Image/Text Queries: A member inquired about the correct way to use the Embed v4 API for hybrid image/text embeddings.
- Another member suggested setting
input_type
tosearch_query
ordocument
when using theembed-v4.0
model.
- Another member suggested setting
- Embed v4 Input Types: The correct way to specify the input type in Embed v4 is to use either
search_query
ordocument
.- This is essential for hybrid image/text embeddings to function correctly.
Cohere â· #đ-introduce-yourself (10 messagesđ„):
Self-learning RL/DL, Agentic AI, Cohere Embed v4, Probabilistic Generative Models
- Robotics Student Adores Cohere Montreal: A PhD student in Robotics, focusing on self-learning RL/DL, expressed their love for Cohere and a desire to join the Montreal office.
- They are seeking like-minded people to discuss and learn together.
- Tech Enthusiast Enters the AI Agent Arena: A newcomer to the world of agentic AI and development is eager to start learning, networking, and building.
- They are excited to implement Cohere into their projects and gift it to the world.
- Researcher Implements Embed V4 Awesomely: An avid tech researcher has recently used Cohere Embed v4 for multimodal retrieval.
- They shared a YouTube video showcasing their work.
- Generative Models Galore: A Math & CS student from Germany is working on probabilistic generative models and learning samplers.
- They are interested in both theory and applied aspects, particularly interpretability, and hope to learn more in this area.
LlamaIndex â· #blog (5 messages):
Open Source NotebookLM, LlamaCloud MCP Servers, AI Hack Night
- Open Source NotebookLM Emerges: A fully open-source NotebookLM is now available, allowing users to run it on their own computers for local document knowledge extraction, according to this tweet.
- LlamaIndex Office Hours Focus on MCP: LlamaIndex held office hours on July 8th, discussing LlamaCloud MCP servers, using existing MCP tools with Agent Workflows, and serving agent workflows as MCP, as stated in this tweet.
- They covered extending functionalities related to Multi-Cloud Processing (MCP).
- AI Hack Night to Spotlight Cutting-Edge Apps: An AI Hack Night at GitHub will focus on building cutting-edge applications with AI Agents, MCPs, and RAG (Retrieval-Augmented Generation) techniques, confirmed by this tweet.
- Structured Data Extraction Workflow Demoed: A demo notebook showcases building a structured data extraction workflow with human-in-the-loop validation, leveraging LLMs for data pre-processing to create schemas for bulk extraction work, per this tweet.
LlamaIndex â· #general (9 messagesđ„):
Document Intelligence for P&IDs, Handwritten Text Extraction, LlamaIndex UX for Business Users
- P&ID Document Intelligence Builds Momentum: A member is developing document intelligence for Piping & Instrumentation Diagrams (P&IDs), which are complex engineering blueprints with overlapping symbols, small text, and intricate lines.
- They are seeking advice on handling complex technical diagrams like electrical schematics, performance benchmarks for dense content, and thoughts on hybrid approaches with LlamaIndex for relationship reasoning.
- Handwritten Text Extraction Tool Hunt Begins: A member is looking for a tool to extract handwritten text as it is from an image, and shared a sample image.
- LlamaIndex UX Quest for Business Acumen: A member is seeking a user-friendly UX for managing docs and indices within LlamaIndex, aiming to allow business users to update the companyâs knowledge base without needing to deeply understand the agentsâ operations.
- They envision a central location for uploading and organizing documents, with developers creating AI agents that utilize the indices; so far, Simba is the only option theyâve identified.
Nomic.ai (GPT4All) â· #general (9 messagesđ„):
Nomic API, GPT4All server mode, Jinja Chat Template, CrowdLLM, OpenAssistant
- New Member Asks About Nomic API: A new member inquired about using the Nomic API and running GPT4All in server mode locally.
- They specifically asked if there were accessible local endpoints.
- User Struggles with Jinja Chat Template: A member requested assistance with a chat template in Jinja format.
- Another member suggested that the user may have a specialized model, and offered assistance understanding the system prompt.
- CrowdLLM Dataset Building Tool Introduced: A member shared a tool called CrowdLLM for crowd-building datasets for LLMs.
- The tool enables users to create tasks and contribute by adding prompt answer pairs, prompts, or answers.
- CrowdLLM compared to OpenAssistant: A member drew a comparison between the newly introduced CrowdLLM and OpenAssistant from 2023.
- Another member noted that OpenAssistant must be available for their system first.
MLOps @Chipro â· #general-ml (3 messages):
Free Book PDF, Author's Blog, Copyright Issues
- User Requests Free Book PDF Directly From Author: A user directly asked an author for a free PDF copy of their book, citing a lack of funds, after listening to a 1-hour talk.
- Another user commented it was in bad taste to ask, suggesting checking the authorâs blog for excerpts instead.
- Alternative Access via Authorâs Blog: A user suggested that instead of requesting a free PDF, the person should check the authorâs blog.
- The blog purportedly contains excerpts and standalone posts from the authorâs books, offering an alternative way to access the content without direct purchase.