a quiet day.

AI News for 3/19/2026-3/20/2026. We checked 12 subreddits, 544 Twitters and no further Discords. AINews’ website lets you search all past issues. As a reminder, AINews is now a section of Latent Space. You can opt in/out of email frequencies!


AI Twitter Recap

Coding Agents, Model Attribution, and the Cursor/Kimi Composer 2 Controversy

  • Cursor’s Composer 2 is built on Kimi K2.5, and the attribution gap became the story: The day’s most consequential engineering/product discourse centered on Cursor’s new coding model. Initial speculation tied Composer 2 to Kimi via tokenizer/model URL signals, with critics questioning why the base model was not disclosed up front and whether licensing terms were being followed (@Yuchenj_UW, @eliebakouch, @ClementDelangue). Cursor then clarified that Composer 2 started from Kimi K2.5, that only about 1/4 of final-model compute came from the base and the rest from continued pretraining plus high-compute RL, and that usage is covered via Fireworks-hosted commercial partnership terms (@leerob, @leerob, @amanrsanger). Kimi later publicly confirmed the partnership and framed Cursor’s work as an example of the open-model ecosystem: K2.5 provided the foundation, Cursor added continued pretraining and RL, and Fireworks supplied hosted RL/inference infrastructure (Kimi Moonshot).
  • Why this mattered technically and strategically: The episode sharpened an industry fault line: increasingly, high-performing products may be post-trained derivatives of strong OSS bases, especially Chinese open-weight models, rather than from-scratch pretraining efforts. Several practitioners argued this is exactly what “foundation models” are for—provided attribution and license obligations are handled cleanly (code_star, @Yuchenj_UW, @Teknium). Others pushed for stronger norms around naming bases, comparing against the base on evals, and giving OSS labs visible credit (@kimmonismus, @cloneofsimo). The net result is less a scandal than a signal: product differentiation is shifting toward domain-specific CPT/RL, evals, and UX, while base-model provenance remains strategically sensitive but increasingly important to disclose.

Open Coding Tooling: Claude Code, T3 Code, Deep Agents, Fleet, and Hermes

  • Claude Code’s ecosystem is spreading into third-party products and channels: Theo dropped Claude integration in T3 Code, effectively allowing users with local Claude Code CLI installed to use it inside T3 Code, immediately joking about possible legal exposure and community notes (Theo, Theo, Theo). Separately, Anthropic appears to be extending Claude Code beyond the terminal into Channels like Telegram and Discord (kimmonismus), while open-source maintainers described meaningful OSS productivity gains from Claude support programs for tasks like Diffusers integrations, profiling, and hardware-aware pipeline optimization (RisingSayak).
  • LangChain is broadening from orchestration to agent products: Multiple tweets highlighted Deep Agents/Open SWE as an open-source Claude Code alternative and LangSmith Fleet as a multi-agent/workforce-style product layer (KSimback, BraceSproul, EvanRimer, hwchase17). LangChain also shipped production-oriented pieces: a Building Reliable Agents course, owners-only prompt promotion in LangSmith Prompt Hub, React Suspense support in @langchain/react, and more messaging around observability for non-deterministic agents in production (LangChain, LangChain, LangChain, LangChain_JS).
  • Hermes/OpenClaw/local-agent workflows continue to mature fast: HermesWorkspace v0.2.0 added one-command startup, UI-based provider/model config, live model catalogs, and new config/model endpoints (outsource_). Hermes also gained parallel web search/page extraction, a workflow recorder/replay system, and stronger prompt-injection defenses via Camel Guard v0.4 (p0, 0xbyt4, WeXBT). Community comparisons between Hermes and OpenClaw emphasized Hermes’ compact, retrieval-heavy memory design versus OpenClaw’s larger replayed history, with concrete latency implications for interactive use (witcheer). A recurring theme across these posts: agent UX is becoming less about one-shot model IQ and more about memory architecture, tool reliability, and loop latency.

Model Releases and Benchmarks: Nemotron-Cascade 2, Mistral Small 4, V-JEPA 2.1, MiMo, and Design/Coding Rankings

  • NVIDIA’s Nemotron-Cascade 2 is the standout open-model release: NVIDIA released Nemotron-Cascade 2, an open 30B MoE with 3B active parameters, positioned as a high-density reasoning/agentic model. The claims are ambitious: gold-medal-level performance on IMO 2025, IOI 2025, and ICPC World Finals 2025, best-in-class math/code/alignment/instruction-following, and superiority over recent Qwen3.5-35B-A3B and Qwen3.5-122B-A10B variants, powered by Cascade RL plus multi-domain on-policy distillation (_weiping, HuggingPapers, ollama). The release is notable not just for weights but for making a compact, high-activation-efficiency reasoning model a first-class OSS option.
  • Mistral Small 4 adds hybrid reasoning + multimodality, but trails peers on intelligence: Artificial Analysis summarized Mistral Small 4 as a 119B MoE with 6.5B active parameters, Apache 2.0 licensed, supporting both reasoning and non-reasoning modes plus image input. It posts 27 on the AA Intelligence Index in reasoning mode, above prior Mistral smalls and matching Magistral Medium 1.2, but still behind peers like gpt-oss-120B (33), Nemotron 3 Super 120B A12B (36), and Qwen3.5 122B A10B (42). It does, however, appear relatively token-efficient and less hallucinatory than some of those peers (Artificial Analysis).
  • V-JEPA 2.1 is an important vision SSL update, especially for dense understanding: FAIR’s new V-JEPA 2.1 shifts from masked-only supervision to learning on both masked and visible tokens, adds deep self-supervision across intermediate layers, and uses modality-specific tokenizers under a shared encoder (TheTuringPost, murloren, massiviola01). Reported gains include +20% robot grasping success over V-JEPA 2 in zero-shot real-world manipulation and new SOTA numbers on Ego4D and EPIC-KITCHENS dense anticipation tasks (TheTuringPost).
  • Benchmark movement elsewhere: DesignArena reported Anthropic Opus 4.6 now leads a broad swath of design-centric coding tasks—web, mobile, 3D design, game dev, and data viz—with Gemini 3.1 taking SVG design (Designarena). Xiaomi’s MiMo V2 Pro/Omni appeared in Arena rankings and independent reviews as a serious but uneven entrant—good in instruction following and some long-task handling, weaker in coding consistency and hallucination resistance (arena, ZhihuFrontier). Alibaba also teased Qwen 3.5 Max Preview placements: #3 math, top-10 Arena Expert, top-15 overall (AlibabaGroup).

Training, RL, Retrieval, and Systems Efficiency

  • Pretraining with targeted data keeps showing leverage over finetuning: A notable research thread from the Stanford/Marin ecosystem argued for data-efficient pretraining via synthetic “megadocs”, reporting roughly 1.8× data efficiency gains and emphasizing that mixing small domain datasets during pretraining resists overfitting much better than repeated finetuning or replay (konwookim, percyliang, leavittron). This aligns with broader discussion that “midtraining = RL prior” and that model adaptation is becoming a key applied capability (cooperleong22, code_star).
  • RL is diversifying beyond math/chat into retrieval and forecasting: CMU/Meta introduced an RL recipe for code search models using only a bash terminal as the exploration interface, avoiding special tools while still yielding strong results (gneubig). Tinker and Mantic reported that RL on gpt-oss-120b for judgmental forecasting outperformed frontier models on event prediction, pushing toward “automated superforecasting” (tinkerapi, johnschulman2).
  • Infra and kernels remain a bottleneck—and a moat: High-engagement posts on kernel engineering underscored that writing custom kernels may now be among the highest-ROI skills for systems-focused engineers (jxmnop, clattner_llvm). ThunderKittens was cited as an example of research-to-production inference transfer where a few hundred milliseconds per generation materially compound over 50+ tool calls in coding agents (boyuan_chen). On the serving side, vLLM was called the de facto standard with roughly half of text-only endpoints in RunPod’s production dataset running vLLM variants (vllm_project).

Applied Tools, Product Launches, and Agent Infrastructure

  • Document parsing as an agent primitive is becoming commoditized: LlamaIndex launched LiteParse, a free local parser that plugs into 40–46+ agents via a one-line npx skills add ... --skill liteparse install path; it’s positioned as both a task-solving utility and a way to feed documents into coding agents as context (jerryjliu0, llama_index, Saboo_Shubham_). LlamaParse also shipped an official agent skill for more complex document understanding across formats/tables/charts/images (llama_index).
  • Local/offline deep research and local-agent stacks are getting credible: Several posts highlighted Local Deep Researcher, an MIT-licensed local research loop that writes its own search queries, scrapes, identifies gaps, and iterates to a cited markdown report using Ollama-compatible models (ihtesham2005, RoundtableSpace). Community demos also showed fully local agent stacks on Apple Silicon and older GPUs using combinations of Hermes/OpenClaw, Qwen, Nemotron, Ollama, and hybrid runners (agenticmate, elldeeone).
  • Perplexity, Devin, and enterprise-agent control surfaces continue expanding: Perplexity Computer added access to Pitchbook, Statista, and CB Insights data, pushing further into analyst/VC workflows (AravSrinivas). Devin added self-scheduling recurring tasks, turning one-off sessions into periodic workflows (cognition). Okta’s “AI agents as governed non-human identities” blueprint and Factory’s enterprise settings hierarchy both point toward a clearer pattern for agent governance in production (dl_weekly, FactoryAI).

Top tweets (by engagement)

  • Model provenance and product positioning: The biggest technical story by engagement was Kimi’s confirmation that Cursor Composer 2 uses Kimi-k2.5 via an authorized Fireworks partnership, effectively closing the loop on the attribution debate (Kimi Moonshot).
  • Open coding product rollout: Theo’s T3 Code now supports Claude post drove major attention and became a flashpoint for platform/TOS questions around coding-agent integrations (Theo).
  • Agent tooling for students and builders: OpenAI launched Codex for Students, offering $100 in credits for college students in the U.S. and Canada (OpenAIDevs).
  • Research release with broad systems implications: NVIDIA’s Nemotron-Cascade 2 announcement stood out among model launches for combining strong reasoning claims with unusually compact active parameter count (_weiping).
  • Vision self-supervision: FAIR’s V-JEPA 2.1 attracted strong attention as a denser, more scalable route to video-based visual understanding (TheTuringPost).

AI Reddit Recap

/r/LocalLlama + /r/localLLM Recap

1. Local AI on Classic Hardware

  • Running TinyLlama 1.1B locally on a PowerBook G4 from 2002. Mac OS 9, no internet, installed from a CD. (Activity: 282): The image depicts a PowerBook G4 from 2002 running Mac OS 9 with the ‘MacinAI Local’ software interface, highlighting a significant achievement in retro computing. This project, unlike previous retro AI endeavors, introduces a custom C89 inference engine designed specifically for classic Macintosh hardware, supporting multiple models like GPT-2 and TinyLlama. It features a 100M parameter transformer trained on Macintosh-specific text, optimized with AltiVec SIMD for a 7.3x speedup, and includes disk paging to handle large models on limited RAM. The setup is entirely offline, installed via CD-R discs, showcasing a unique blend of vintage hardware and modern AI capabilities. Commenters express admiration for the project’s novelty and technical achievement, with one noting the humor in the inference time for the TinyLlama model and another excited about integrating it with Hypercard stack XCMD.

    • The implementation of TinyLlama on a PowerBook G4 is notable for its use of AltiVec optimization, which achieved a 7.3x speedup. This optimization is crucial given the hardware constraints of a 2002-era machine. Additionally, the use of a disk paging system to manage layers that exceed the available RAM is a clever workaround, allowing the model to function effectively despite limited resources.
    • The project showcases the potential of running a language model on legacy hardware by leveraging agentic AppleScript control, which enhances its utility. This approach not only demonstrates the feasibility of running LLMs on older systems but also contributes to the retro computing community by providing a practical application of AI on vintage machines.
    • The discussion highlights the impressive nature of running any LLM on a G4, emphasizing the technical challenges overcome, such as optimizing for AltiVec and managing memory constraints through disk paging. These solutions make the project a significant achievement in the realm of retro computing and AI integration.

2. Qwen Model Performance and Optimization

  • Qwen3.5 is a working dog. (Activity: 623): The post discusses the Qwen3.5 model, emphasizing its need for extensive context to function effectively, particularly highlighting that the 27B model requires at least 3K tokens to be useful. The author suggests that these models are designed to be ‘agentic-first’, meaning they perform better when given clear objectives and environmental context, rather than minimal prompts. The post also critiques the 35B MoE model as underperforming. The Qwen models are described as ‘working dogs’ that excel when given specific tasks, aligning with Alibaba’s design intentions for open weights models. Commenters generally agree with the need for explicit instructions for the Qwen models, with one noting that the 27B model requires clear directives to avoid unnecessary actions. Another commenter shares a positive experience with the 122B model, noting its effectiveness with a 600tk system prompt limit, suggesting that it benefits from a high-level, open-world tools environment.

    • The user ‘abnormal_human’ discusses their experience with the 122B model, emphasizing the effectiveness of a strict 600 token limit on the system prompt. They note that this approach enhances the model’s performance by focusing on prompting behavior rather than pattern matching, likening it to a Claude code environment. This method reportedly prevents overthinking and improves task execution.
    • User ‘zasad84’ shares insights from experimenting with various models, including 35b-a3b, 27b, and 9b. They highlight the surprising efficacy of the 9b model for specific tasks when provided with a large, direct system prompt. Utilizing an ‘unsloth quant’ allows them to leverage the full context window on a 24GB card, which is crucial for tasks where context size is a limiting factor. They also mention using a SOTA model like Gemini 3.1 pro to craft effective system prompts.
    • ‘ggonavyy’ notes that when working with the 27B model, explicit instructions are necessary to prevent the model from attempting all possible solutions, even in planning mode. This suggests that without clear directives, the model may overextend its problem-solving efforts, highlighting the importance of precise prompt engineering to guide the model’s behavior effectively.

3. New Model and Hardware Considerations

  • 128gb M5 Max for local agentic ai? (Activity: 112): The post discusses the suitability of a 128GB M5 Max MacBook for running local large language models (LLMs) and personal AI agents, focusing on privacy and avoiding cloud-based solutions. The user currently uses an RTX 4070 with 16GB RAM but finds it limiting for local models. They are considering the M5 Max for its ability to handle large models like gpt-oss-120b and nemotron-3-super-120b-a12b with Q4/Q5 quantization, which allows for efficient local processing without relying on external APIs like Claude or ChatGPT. The MacBook’s flexibility and performance, even with large models, are highlighted, though it may produce significant fan noise under heavy load. Commenters generally agree that the 128GB M5 Max MacBook is capable of running large models efficiently, with one noting it can handle qwen 3.5 390b with some tweaks, achieving around 12 tokens/second. The MacBook is praised for its flexibility compared to a dedicated GPU setup, though some express hesitation due to its high cost.

    • JuliaMakesIt highlights the capability of the M5 Max 128GB MacBook Pro to handle large AI models like gpt-oss-120b, nemotron-3-super-120b-a12b, and qwen3.5-122b-a10b with ease, especially when using Q4/Q5 quantization. This setup allows for significant local processing power, reducing reliance on external APIs like Claude or ChatGPT, and offers flexibility beyond a dedicated GPU array.
    • Consistent-Cold4505 mentions that the M5 Max can run qwen 3.5 390b efficiently with some optimizations, achieving approximately 12 tokens per second. This performance is attributed to Apple’s design, which is particularly suited for such tasks. A comparison is made with a 48GB Mac Mini M3 Max, which achieves nearly 5 tokens per second at 4-bit precision, highlighting the M5 Max’s superior capabilities.
    • TimLikesAI compares the M4 Max 128GB to the M5 Max, noting that while the M4 has slower prefill speeds, it still supports running large models effectively. This suggests that even previous generation hardware can be quite capable, though the M5 Max offers improved performance for demanding AI workloads.
  • Ooh, new drama just dropped 👀 (Activity: 1482): The image is a meme that humorously depicts a controversy involving Cursor’s new model, Composer 2, which is allegedly built on top of Kimi K2.5 without proper attribution. This has sparked discussions about licensing and attribution, particularly focusing on the modified MIT License used by Kimi K2.5, which requires attribution if the software is used in commercial products with significant user or revenue metrics. The meme suggests a ‘reveal’ of this underlying issue, aligning with the title’s indication of new drama. View Image. Some commenters argue that Cursor’s approach is typical for companies that quickly capitalize on market gaps, but they face limitations due to reliance on existing products and API costs. Others criticize the trend of creating ‘wrapper’ products that are perceived as lacking substance and driven by hype.

    • wOvAN discusses a modified MIT License used by Moonshot AI, which includes a unique clause requiring companies with over 100 million monthly active users or $20 million in monthly revenue to display ‘Kimi K2’ on their user interface. This modification aims to ensure visibility and credit for the software’s use in large-scale commercial applications.
    • Everlier provides insight into Cursor’s business model, highlighting that they capitalize on market gaps by quickly offering solutions without a strong foundational base. They rely on existing products, which limits their innovation potential. Everlier also notes that Cursor’s use of Kimi 2.5 aligns with the licensing terms, which permit such adaptations by corporations.
    • Technical-Earth-3254 questions the limitations of Cursor’s pricing strategy, specifically whether the model is unlimited in their plan. This reflects a broader uncertainty about how Cursor structures its offerings and the potential constraints on model usage within their pricing tiers.

Less Technical AI Subreddit Recap

/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo, /r/aivideo

1. AI Model and Benchmark Launches

  • Cursor’s ‘Composer 2’ model is apparently just Kimi K2.5 with RL fine-tuning. Moonshot AI says they never paid or got permission (Activity: 739): The image reveals that Cursor’s ‘Composer 2’ model is essentially the Kimi K2.5 model with additional reinforcement learning fine-tuning. This discovery has raised concerns about Cursor AI’s practices, as they allegedly used the Kimi tokenizer without obtaining permission or paying the necessary fees. The terminal screenshot in the image shows a debug session that identifies the model as ‘kimi-k2p5-rl-0317-s515-fast,’ suggesting that the model’s core is derived from Kimi K2.5. Commenters express skepticism about Cursor’s future, noting a preference for open models and criticizing Cursor for not being transparent about using Kimi’s model. There is also criticism of Cursor’s approach to using models developed by others without proper acknowledgment or licensing.

    • The discussion highlights concerns about the transparency and originality of Cursor’s ‘Composer 2’ model, which is reportedly a fine-tuned version of the open-source Kimi K2.5 model. This raises questions about the authenticity of proprietary models and the ethics of rebranding open-source models without proper acknowledgment or compensation to the original creators, such as Moonshot AI in this case.
    • There is a critique of Cursor’s business model, particularly the decision to charge $20/month for what is essentially a rebranded version of an existing open-source model. This situation underscores the importance of open models and the potential for companies to exploit them without contributing back to the community or providing clear attribution, which can mislead users about the true nature of the advancements being offered.
    • The comments also touch on the licensing aspect, noting that while Cursor may have technically adhered to the modified MIT license terms, the ethical implications of not prominently crediting the original Kimi K2.5 model in their commercial offerings are questioned. This situation exemplifies the broader issue of how open-source licenses are interpreted and enforced in commercial contexts.
  • Cursor’s new Composer 2 just beat Claude Opus at coding and it’s 10x cheaper (Activity: 195): Cursor has released Composer 2, a coding model that achieves 61.7% on Terminal-Bench 2.0, surpassing Claude Opus 4.6 at 58.0%. It is priced at $0.50 per million tokens, significantly cheaper than Opus’s $5.00. Although it lags behind GPT-5.4’s 75.1%, it offers a cost advantage at 1/5th the price. The model is trained exclusively on code and features “self-summarization” to compress long sessions without losing context. Meanwhile, OpenAI has acquired Astral to enhance Codex, indicating intensifying competition in AI coding models. Some commenters suggest that Composer 2 might be based on GLM 5 and criticize Cursor for not developing from scratch, implying they are fine-tuning existing open-source models like Kimi-K2.5 or GLM 4.7/5.

    • The underlying model for Cursor’s Composer 2 is speculated to be GLM 5, as mentioned by users who suggest that the developers are not pretraining from scratch but rather fine-tuning existing open-source models. This approach might involve models like Kimi-K2.5 or GLM 4.7/5, indicating a reliance on established architectures rather than novel developments.
    • A user compared Cursor’s Composer 2 to Claude Opus, noting that while Composer 2 is marketed as being able to read the entire codebase unlike GitHub Copilot, its performance is reportedly inferior. The user experienced issues with code readability and integration, requiring frequent manual interventions, which contrasts with their experience using Claude Opus where such interventions were rare.
  • MacBook M5 Pro + Qwen3.5 = Fully Local AI Security System — 93.8% Accuracy, 25 tok/s, No Cloud Needed (96-Test Benchmark vs GPT-5.4) (Activity: 235): The post discusses the performance of the Qwen3.5 model running locally on an Apple M5 Pro as part of a home security system benchmark called HomeSec-Bench. The Qwen3.5-9B model achieved a 93.8% accuracy rate, closely trailing GPT-5.4 by 4.1 points, while operating at 25 tokens per second and using 13.8 GB of memory. The Qwen3.5-35B MoE variant demonstrated a faster first-token time (435ms TTFT) and higher throughput (42 tok/s) than any tested OpenAI cloud endpoint. The benchmark evaluates models on tasks like context preprocessing, event deduplication, and security classification, emphasizing the feasibility of running advanced AI models locally with full data privacy and no API costs. Full results and methodology are available on GitHub. A commenter suggests exploring the jang_q models, which reportedly outperform other models like mlx 4bit minimax on benchmarks such as MMLU, achieving near 80% accuracy with a 60GB model. Another user questions the choice of using a quantized version of Qwen3.5-9B instead of the full model.

    • HealthyCommunicat highlights the performance of the JANG_Q models, noting that the 2-bit JANG_Q equivalent (60GB) achieves nearly 80% on MMLU, outperforming the 4-bit MLX minimax (120GB) which scores below 30%. They emphasize the efficiency of JANG_Q models, particularly the 180GB version of Qwen 3.5 397b, which scores 93% on MMLU, suggesting it as a superior alternative for local AI systems.
    • just one Question inquires about the choice of the quantized version Qwen3.5-9B (Q4_K_M) over the full version. This question implies a trade-off consideration between model size and performance, where quantized models like Q4_K_M are often used to reduce computational load and memory usage while maintaining reasonable accuracy, making them suitable for devices with limited resources.
    • pascon asks about the compatibility of QWEN models with a Mac Mini having 24MB of unified memory. This question is crucial for understanding the hardware requirements and limitations when deploying AI models locally, as it directly impacts the feasibility of running large models on consumer-grade hardware.

2. Claude Code and Development Tools

  • I built a Claude skill that writes accurate prompts for any AI tool. To stop burning credits on bad prompts. We just hit 600 stars on GitHub‼️ (Activity: 1339): The prompt-master is a Claude skill designed to optimize prompt generation for various AI tools, achieving over 600 stars on GitHub. It intelligently detects the target AI tool and applies specific strategies, such as extracting 9 dimensions from user input and identifying 35 common prompt issues, to enhance prompt accuracy and efficiency. The tool supports a wide range of platforms including Claude, ChatGPT, Midjourney, and Eleven Labs, and features 12 auto-selecting prompt templates tailored to different tasks. The project is open-source and continuously improved based on community feedback, with the latest version v1.4 recently released. Commenters highlight the tool’s ability to route prompts specifically for different AI models, such as Midjourney and Claude Code, as a key differentiator from generic prompt tools. There is also interest in its compatibility with open-source models, as noted by a user running it locally with comfyui on a 5090 GPU.

    • dovyp highlights the importance of tool-specific routing in the Claude skill, noting that it differentiates between the structures needed for different AI tools like Midjourney and Claude Code. This specificity is crucial as most general prompt tools fail to address the unique requirements of each tool, making this skill particularly valuable.
    • JMdesigner inquires about the compatibility of the Claude skill with open-source models, mentioning their setup with ComfyUI on a 5090 GPU. This suggests interest in leveraging the skill’s capabilities beyond proprietary models, potentially expanding its utility in diverse AI environments.
    • dogazine4570 reflects on their experience with a similar tool for Claude Code, noting that while it was somewhat helpful, manual tweaking of prompts was still necessary. They express interest in the skill’s ability to handle tool-specific quirks, such as differences between Cursor and Claude Code, which could enhance its practical utility.
  • Pretty sure I’m not using Claude to its full potential - what plugins/connectors are worth it? (Activity: 863): The Reddit post discusses optimizing the use of Claude, an AI tool, by integrating it with various plugins and connectors. A notable suggestion is using the /insights command to generate usage reports and improvement advice. Another advanced setup involves creating a single MCP server that routes tool calls through a Chrome extension, leveraging existing web app sessions (e.g., Slack, Linear, Datadog, Google Sheets) to streamline workflows without managing separate API keys. This setup facilitates seamless task automation across 100+ web apps, with an open-source implementation available on GitHub. Additionally, the Superpowers plugin is recommended for developers, enhancing Claude’s capabilities and available via the official plugin list. The comments highlight a preference for integrating Claude with existing web apps to streamline workflows, emphasizing the efficiency of a single MCP server setup over managing multiple API keys. The Superpowers plugin is particularly valued by developers for its impact on productivity.

    • opentabs-dev describes a streamlined approach to integrating Claude with various web apps like Slack, Linear, and Datadog. Instead of managing separate MCP servers and API keys, they use a single MCP server that routes tool calls through a Chrome extension, leveraging active browser sessions. This setup allows Claude to perform tasks like slack_send_message or linear_create_issue without managing credentials, significantly enhancing workflow efficiency by automating ‘glue work’ across 100+ web apps. The open-source project is available on GitHub.
    • Judecale recommends the ‘superpowers’ plugin for developers using Claude, highlighting its transformative impact on professional development workflows. This plugin is part of the official plugin list and can be easily added via the CLI, suggesting it offers significant enhancements for coding tasks.
    • eo37 mentions using ‘context7’ and ‘superpowers’ as primary MCPs, emphasizing their utility in managing tasks. They also built a custom MCP to access the latest LLM models and track API costs, which aids in determining long-term running and production expenses. This approach underscores the importance of cost management and staying updated with the latest model advancements.

3. AI in Creative and Personal Projects

  • An Australian ML researcher, used ChatGPT+AlphaFold to shrink 75% of his life-threatened dog’s MCT cancerous tumor, developing a personalized mRNA vaccine in just two months - after sequencing his dog’s DNA for $2,000 (Activity: 768): An Australian machine learning researcher, Paul Conyngham, utilized ChatGPT and AlphaFold to develop a personalized mRNA vaccine for his dog, Rosie, who had a life-threatening mast cell tumor (MCT). By sequencing the tumor’s DNA for approximately $2,000, Conyngham identified neoantigens using ChatGPT and predicted protein structures with AlphaFold. Collaborating with Martin Smith from UNSW for genome sequencing and Pall Thordarson for mRNA synthesis, Conyngham successfully shrank the tumor by 75% within two months, despite having no formal background in biology or medicine. This case highlights the potential of AI in personalized medicine and rapid vaccine development (source). Commenters are debating the implications of this case, questioning whether it represents a significant shift in healthcare democratization or if it’s merely hype. Some suggest that regulatory barriers are hindering medical advancements, as demonstrated by the rapid progress made in this unregulated scenario.

  • Chatgtp is legit helping me cook insane (Activity: 721): The post discusses how ChatGPT is being used to enhance cooking skills by generating step-by-step recipes tailored to available ingredients and desired presentation. Users report that ChatGPT can dynamically adjust recipes based on ingredient substitutions and provide meal plans from grocery receipts. Additionally, it can calculate nutritional information such as calories and macros. A tool called Chef Genius Generator is mentioned, which creates prompts for ChatGPT based on available kitchen equipment, spices, and food preferences. Commenters highlight the flexibility of ChatGPT in adapting recipes on-the-fly and its utility in winning cooking contests. The ability to generate personalized meal plans and nutritional information is also praised.

    • ewbankpj highlights the dynamic adaptability of ChatGPT in cooking, noting its ability to adjust recipe ratios on-the-fly when substituting ingredients. This feature is particularly useful for those who need to make quick changes based on available ingredients or dietary needs.
    • DueCommunication9248 mentions using ChatGPT for nutritional analysis, emphasizing its capability to calculate calories and macros from ingredients or images. This functionality can be particularly beneficial for those tracking their dietary intake or participating in cooking contests.
    • bbum shares a tool called the ‘chef-genius-generator’ that creates prompts for ChatGPT based on available kitchen equipment, spices, and food preferences. This tool enhances the personalization of recipe generation, allowing users to optimize recipes according to their specific kitchen setup and dietary preferences.

AI Discords

Unfortunately, Discord shut down our access today. We will not bring it back in this form but we will be shipping the new AINews soon. Thanks for reading to here, it was a good run.