a quiet day.
AI News for 4/26/2026-4/27/2026. We checked 12 subreddits, 544 Twitters and no further Discords. AINewsâ website lets you search all past issues. As a reminder, AINews is now a section of Latent Space. You can opt in/out of email frequencies!
AI Twitter Recap
OpenAI Distribution Shift, GPT-5.5 Benchmarks, and Codex/Copilot Pricing Signals
- OpenAI loosens Azure exclusivity: @sama said OpenAI updated its Microsoft partnership so Microsoft remains the primary cloud, but OpenAI can now make products available across all clouds, with product/model commitments extending to 2032 and revenue share through 2030. The implication was quickly drawn by @scaling01 and @kimmonismus: OpenAI can now distribute via Google TPU / AWS Trainium / Bedrock, and Microsoftâs license to OpenAI IP becomes non-exclusive. @ajassy confirmed OpenAI models are coming to AWS Bedrock in the coming weeks. @simonw noted the new language likely means the old AGI clause is effectively gone.
- GPT-5.5 is a broad upgrade, but not uniformly dominant: Community evals from @htihle put GPT-5.5 no-thinking at 67.1% on WeirdML, up from 57.4% for GPT-5.4, but still behind Opus 4.7 no-thinking at 76.4% while using fewer tokens. LMSYS Arena results from @arena placed GPT-5.5 at #9 in Code Arena, #6 Document, #7 Text, #3 Math, #2 Search, #5 Vision, with Expert Arena #5. Arena also clarified current evaluation covers medium/high reasoning, with xHigh still pending (1, 2). Practitioner feedback was positive for hard coding tasks such as GPU kernels from @gdb, but there were also reports of âcompressed CoT leakageâ / malformed outputs in no-thinking mode from @htihle.
- Developer economics are becoming more explicit: GitHub announced Copilot moves to usage-based billing on June 1, a notable shift as agentic workflows consume much more runtime. Parallel to that, @Hangsiin documented Codex usage multipliers: GPT-5.4 fast = 2x, GPT-5.5 fast = 2.5x, with 5.4-mini and GPT-5.3-Codex materially cheaper. @sama argued Codex at $20 remains a strong value. OpenAI also open-sourced Symphony, an orchestration layer connecting issue trackers to Codex agents for âopen issue â agent â PR â human review,â via @OpenAIDevs.
Xiaomi MiMo-V2.5, Kimi K2.6, and Chinaâs Agent-Oriented Open-Weights Push
- MiMo-V2.5 is one of the dayâs biggest open releases: @XiaomiMiMo open-sourced MiMoâV2.5-Pro and MiMoâV2.5 under MIT, both with 1M-token context. The Pro model is framed as a complex agent/coding model and the smaller model as a native omni-modal agent. Community summaries from @eliebakouch add useful technical details: MiMoâV2.5-Pro is roughly 1T total / 42B active, trained on 27T tokens in FP8, while MiMoâV2.5 is about 310B total / 15B active, trained on 48T tokens, with aggressive interleaved SWA/global attention and no shared expert. Xiaomi also announced a 100T token grant for builders via @_LuoFuli. Day-0 inference support landed quickly in vLLM and SGLang/vLLM.
- Kimi K2.6 continues to lead in mindshare and deployment: @Kimi_Moonshot said Kimi K2.6 is now #1 on OpenRouterâs weekly leaderboard. Secondary reporting described it as a model for coding and long-horizon agents, including scaling to 300 concurrent sub-agents across 4,000 coordinated steps (dl_weekly). Practitioners remain split on speed/quality tradeoffs: @teortaxesTex found Kimi in Hermes much slower than DeepSeek V4 but sometimes capable of fixing bugs V4 could not.
- Broader China-model trend: Multiple posts framed Chinese labs as pushing aggressively on open-ish, agent-oriented, long-context systems: Qwen 3.6 Flash, DeepSeek V4/Flash, GLM-5.1 promotions (triple usage extension), and Xiaomiâs MIT release. A recurring theme was that smaller / cheaper variants are often outperforming their larger siblings on practical agent benchmarks.
Agent Runtimes, Orchestration, and Local-First Tooling
- Sakanaâs Conductor is a notable multi-agent result: @SakanaAILabs introduced a 7B Conductor trained with RL to orchestrate a pool of frontier models in natural language rather than solving tasks directly. It dynamically decides which agent to call, what subtask to assign, and which context to expose, and reportedly reached 83.9% on LiveCodeBench and 87.5% on GPQA-Diamond, beating any single worker in its pool. @hardmaru highlighted âAI managing AIâ and recursive self-selection as a new axis of test-time scaling.
- Local and hybrid agents keep getting better: Several posts showed coding/assistant stacks running locally. @patloeber and @_philschmid documented running Pi agent + Gemma 4 26B A4B locally via LM Studio/Ollama/llama.cpp. @googlegemma demoed a fully local browser agent using Gemma 4 + WebGPU, with native tool calling for browsing history, tab management, and page summarization. @cognition shipped Devin for Terminal, a local shell agent that can later hand off to the cloud.
- Agent ergonomics and framework evolution: Hermes had a strong day: @Teknium noted Hermes Agentâs repo surpassed Claude Code, while native vision became the default when supported. The broader ecosystem kept filling in missing pieces: Cline Kanban now supports different agents/models per task card; Future AGI open-sourced an eval/optimization stack for self-improving agents; and @_philschmid argued MCP works best either through explicit @mention loading or subagent-scoped tool assignment, not indiscriminate server attachment.
Inference Infrastructure, Attention/KV Engineering, and Systems Work
- Googleâs TPU split is a meaningful architecture signal: Several posts dissected Googleâs Cloud Next announcement that TPU v8 is split into 8t for training and 8i for inference, with claims of roughly 2.8x faster training and 80% better inference performance/$ than prior generation. @kimmonismus emphasized this is the first time Google split custom silicon by workload and that OpenAI, Anthropic, and Meta are reportedly buying TPU capacity.
- DeepSeek V4 support is maturing quickly in infra stacks: @vllm_project said support for DeepSeek V4 base models is coming, requiring an
expert_dtypeconfig field to distinguish FP4 instruct vs FP8 base. In the vLLM 0.20.0 release, highlights included DeepSeek V4 support, FA4 as default MLA prefill, TurboQuant 2-bit KV, and a DeepSeek-specific MegaMoE path on Blackwell. - KV cache optimization remains a hot battleground: There was dense discussion around long-context bottlenecks and KV strategies. @cHHillee summarized three main levers for long contexts: local/sliding attention, interleaved local-global attention, and smaller KV per global layer via GQA/MLA/KV tying/quantization. On the implementation side, @vllm_project and Red Hat/AWS published an FP8 KV-cache deep dive where a fix to FA3 two-level accumulation improved 128k needle-in-a-haystack from 13% to 89% while retaining FP8 decode speedups. Community critics also questioned DeepSeek V4âs specific KV tradeoffs relative to offloading-heavy approaches such as HiSparse (discussion).
Benchmarks, Evals, and Open Research Directions
- Open-world evaluation is gaining momentum: @sarahookr argued that most agentic benchmarks are overfit to automatically verifiable tasks, while the important frontier is open-world, uncertain, non-fully-verifiable work. Related threads connected this to continual learning, memory stores, and adaptive data systems (1, 2).
- Cost-aware agent evaluation is becoming first-class: @dair_ai highlighted a new study on coding-agent spend over SWE-bench Verified: agentic coding can consume ~1000x more tokens than chat/code reasoning, usage can vary 30x across runs on identical tasks, and more spending does not monotonically improve accuracy. This lines up with pricing-model changes from Copilot and growing concern over uncontrolled agent runtime economics.
- New benchmarks and domain-specific evals: ParseBench from LlamaIndex adds 2k verified enterprise document pages for parsing agents. AgentIR reframes retrieval for research agents by embedding the reasoning trace alongside the query, with AgentIR-4B hitting 68% on BrowseComp-Plus vs 52% for larger conventional embedding models. There were also several benchmark snapshots for frontier modelsâe.g. Opus 4.7 leading GSO at 42.2% and WeirdML / ALE-Bench / PencilPuzzleBench chatterâbut the stronger signal was methodological: more people are measuring runtime cost, retrieval quality, and open-world behavior, not just final answer accuracy.
Top tweets (by engagement)
- OpenAIâMicrosoft partnership reset: @sama on cross-cloud availability and continued Microsoft partnership.
- OpenAI on AWS: @ajassy confirming OpenAI models are coming to Bedrock.
- GitHub Copilot pricing change: @github announcing usage-based billing starting June 1.
- Xiaomi MiMo-V2.5 open-source release: @XiaomiMiMo with MIT license and 1M context.
- Open-source orchestration for Codex: @OpenAIDevs launching Symphony.
- Gemma local browser agent: @googlegemma showing a 100% local browser-resident agent with WebGPU.
AI Reddit Recap
/r/LocalLlama + /r/localLLM Recap
1. Qwen3.6 Model Performance and Optimization
-
Luce DFlash: Qwen3.6-27B at up to 2x throughput on a single RTX 3090 (Activity: 743): Luce DFlash is a new implementation of speculative decoding for the Qwen3.6-27B model, optimized to run on a single RTX 3090 GPU using a standalone C++/CUDA stack built on top of ggml. This setup achieves up to
1.98xthroughput compared to autoregressive decoding across benchmarks like HumanEval, GSM8K, and Math500, without requiring retraining. The system uses a compressed KV cache and sliding-window flash attention to efficiently handle large contexts, and it can serve over an OpenAI-compatible HTTP endpoint. The implementation is constrained to CUDA environments and does not support multi-GPU setups. Commenters are enthusiastic about the innovation, with one noting the potential for local AI inference and another expressing interest in the setup due to its speed advantages over existing configurations.- Tiny_Arugula_5648 raises a critical point about the impact of quantization on model accuracy. They emphasize that while the increased throughput is impressive, the heavy quantization involved can significantly affect the modelâs performance in certain applications, such as coding or tool calling, where precision is crucial. This highlights the importance of understanding the trade-offs between speed and accuracy when deploying such models.
- drrck82 expresses interest in the setup, particularly as a dual RTX 3090 owner. They mention currently using the Q6_K_XL model for enhanced intelligence but find the potential for doubling speed with the Qwen3.6-27B model very appealing. This suggests a focus on balancing computational efficiency with model sophistication.
- DeepV inquires about the possibility of dockerizing the setup, indicating a demand for containerized solutions that can simplify deployment and scaling. Dockerization could facilitate easier integration into existing workflows and enhance reproducibility across different environments.
-
To 16GB VRAM users, plug in your old GPU (Activity: 666): The post discusses leveraging an old GPU with at least
6GB VRAMalongside a primary16GB VRAMGPU to run dense models like30bmodels more efficiently. The author uses a5070Ti 16GBand an old2060 6GB, achieving a combined22GB VRAM, which approaches the capacity of a24GBclass card. The setup involves usingllama-serverwith specific configurations to optimize GPU usage, such asdev=Vulkan1,Vulkan2to enable both GPUs, andno-mmapto keep the model off RAM. Benchmark results show significant performance improvements, with186 tokens/sfor prompt processing and19 tokens/sfor generation at128kmax context, compared to4 tokens/son a single card. The post also provides detailedllama-benchresults using CUDA, highlighting the importance of fitting models on GPU VRAM for speed, especially at long contexts. Commenters debate the use of Vulkan versus CUDA, with some suggesting CUDA for better performance. Others note that while additional VRAM from a weaker GPU can help, it may bottleneck a stronger GPU, as seen with a3090 Tiand2070setup, where splitting across GPUs reduced speed compared to using the3090 Tialone.- Mysterious_Role_8852 discusses the performance bottleneck when using a 3090 Ti and a 2070 together. They note that the 2070 bottlenecks the 3090, resulting in a decrease in performance from
30t/sto20t/swhen splitting tasks between the GPUs. This highlights the importance of matching GPU capabilities to avoid performance degradation, especially when handling large models like Qwen 3.6 27b Q6 Quant. - mac1e2 provides a detailed account of running Qwen3.6-35B-A3B on a constrained system with a GTX 1650 4GB and 62GB RAM. They emphasize the importance of understanding hardware limitations and optimizing configurations, such as using
--cpu-moe,--mlock, and specific cache settings. The post underscores the value of disciplined resource management, contrasting it with modern practices that often rely on abundant hardware resources without optimizing for efficiency. - jacek2023 mentions using a 3060 as an additional GPU alongside three 3090s, but only for the largest models. This suggests a strategic approach to leveraging available hardware, where additional GPUs are utilized selectively to maximize performance for specific tasks, rather than a blanket approach of using all available resources indiscriminately.
- Mysterious_Role_8852 discusses the performance bottleneck when using a 3090 Ti and a 2070 together. They note that the 2070 bottlenecks the 3090, resulting in a decrease in performance from
-
Switched from Qwen3.6 35b-a3b to Qwen3.6 27b mid coding and itâs noticeably better! (Activity: 437): The image is a screenshot from a tower defense game developed using Qwen3.6 models, specifically highlighting the transition from the 35b-a3b to the 27b version. The user experienced improved performance with the 27b model despite using a more compressed IQ3_M version due to VRAM limitations. This suggests that dense models like Qwen3.6-27B-i1-GGUF handle compression better than MoE models, as evidenced by the modelâs ability to identify a difficult bug that the larger model couldnât. The user reports token processing speeds of 40 tokens per second with the 27b model, which maintained consistent speed compared to the 35b-a3b modelâs fluctuating performance. Commenters generally agree that dense models like Qwen3.6-27B are more efficient and reliable for users with limited VRAM, and they appreciate the availability of such models for local use without cloud dependency. Some users have successfully used these models for legitimate work, noting their satisfactory speed and context window.
- The user âridablellamaâ highlights the practical utility of the Qwen3.6 27b model, noting that while it may not match the performance of Claude Code, it is still capable of handling legitimate work tasks effectively. They emphasize the modelâs speed and context window as satisfactory, suggesting that with further fine-tuning, its performance could improve significantly. This underscores the modelâs potential as a reliable baseline for users with 16-24 GB VRAM, offering a cost-effective alternative to cloud-based solutions.
- âYairHairNowâ provides a detailed comparison of different Qwen3.6 model configurations, focusing on token generation speed and context length. The 27B IQ4_XS model is highlighted for its long-context mode with a maximum context of 196K and a token generation speed of 48 tokens per second. In contrast, the 35B-A3B Q3_K_S configuration offers a faster token generation speed of up to 149 tokens per second but with a shorter context length of around 65K. This comparison is valuable for users deciding between speed and context capabilities.
- âKillerX629â mentions a noticeable slowdown when switching models, which affects their workflow due to reduced tokens per second. This highlights a common trade-off in model selection between performance speed and other factors such as model size or context capabilities. The comment suggests that users need to balance these aspects based on their specific requirements and hardware capabilities.
-
Qwen3.6-27B-INT4 clocking 100 tps with 256k context length on 1x RTX 5090 via vllm 0.19 (Activity: 426): The post discusses the performance of the Qwen3.6-27B-INT4 model, achieving
105-108 tokens per second (tps)with a256k context lengthon a single RTX 5090 using vllm 0.19. The model, available on Hugging Face, benefits from MTP support and a smaller size, allowing full native context window utilization. The setup usesauto_roundquantization andfp8_e4m3for KV cache dtype, with a focus on interactivity and speculative decoding. The configuration includesflashinferattention backend andmtpspeculative configuration with3 speculative tokens. A notable comment highlights the use of Turboquant 3-bit NC KV Cache for compressing KV state, enabling a125K context windowwithin24GB VRAM. The MTP n=3 Speculative Decoding is praised for its throughput multiplier, and Cudagraph PIECEWISE Mode is noted for eliminating repetition loops. The setupâs efficiency is further enhanced by Chunked Prefill and Prefix Caching, stabilizing request times after initial cudagraph compilation.- The Qwen3.6-27B model achieves impressive performance on a single RTX 5090, with sustained throughput of 120-124 tokens per second (TPS) for narrative tasks and 156-159 TPS for code tasks. This is facilitated by the INT4 AutoRound quantization and BF16 MTP head preservation, running on vLLM 0.19.2rc1 with Genesis v7.0 patches. The setup utilizes a 258,048 token context window, close to the architectural maximum of 262,144 tokens, and maintains high GPU utilization at 93% with a power draw of 400-426W.
- The implementation of speculative decoding with MTP n=3 significantly enhances throughput, achieving a mean acceptance length of 2.65-3.46 and an acceptance rate between 55-82%. This method involves using three auxiliary heads to draft tokens per forward pass, which are then verified against the main head, providing a throughput multiplier of approximately 3x compared to non-speculative baselines. This technique is crucial for maintaining high performance in local inference scenarios.
- Turboquantâs 3-bit NC KV Cache is a notable feature, allowing the compression of the KV state to 3-bit non-uniform quantization. This enables a 125K context window within 24GB VRAM without running out of memory (OOM). Additionally, the use of Cudagraph PIECEWISE mode, which captures only attention-op boundaries, helps eliminate degenerate repetition loops caused by stale MTP state in FULL_AND_PIECEWISE mode on multi-GPU hosts.
-
Running Qwen 3.6 in Claude Code (Activity: 194): The user is attempting to run large local models like Qwen 3.6 27B and Gemma 4 26B on a system with an
RTX 4070GPU with8GB VRAMand32GB RAM, but faces performance issues such as slow execution and infinite loops. Smaller models like Qwen 2.5-coder 7B and Gemma 4 e4b are insufficient for coding tasks. A suggestion is made to use Qwen3.6-35B-A3B, a mixture of experts (MoE) model, which allows offloading part of the model to the GPU and the rest to RAM, potentially improving speed by2-3xwhile maintaining strong performance. The user also considers using Roo Code in VSCode as an alternative to Claude Code, where Qwen failed to complete tasks. Commenters note that local MoE models can struggle with complex tasks, especially when involving multiple tools, due to large initial prompts and slow processing speeds (20tps). This can lead to timeouts and incomplete tasks, suggesting that substantial VRAM (48GB+) is necessary for effective local model execution.- ghgi_ suggests using the Qwen3.6-35B-A3B model, which is a Mixture of Experts (MoE) model. This allows for offloading the 3B part to a GPU while keeping the rest of the weights in RAM, potentially offering a 2-3x speed boost over the Qwen 3.6 27B model. Despite being slightly less performant than the 27B model, it remains competitive across most metrics.
- OneSlash137 discusses the challenges of using local MoE models, particularly with complex tasks involving multiple tools. They note that while the model can handle simple interactions, it struggles with large prompts (10-20k tokens), leading to long processing times (3-5 minutes) and frequent timeouts. The issue is exacerbated by the modelâs tendency to resend entire prompts instead of incremental updates, causing significant delays and inefficiencies.
- Plane-Pause-469 inquires about running local models in Claude code and seeks advice on the best models to use with 128 GB of RAM and 16GB VRAM. This highlights a common concern among users about optimizing hardware resources to effectively run large language models locally.
2. New Model and Benchmark Releases
-
Microsoft Presents âTRELLIS.2â: An Open-Source, 4b-Parameter, Image-To-3D Model Producing Up To 1536Âł PBR Textured Assets, Built On Native 3D VAES With 16Ă Spatial Compression, Delivering Efficient, Scalable, High-Fidelity Asset Generation. (Activity: 376): Microsoft has introduced âTRELLIS.2â, a cutting-edge 4-billion parameter model for generating high-fidelity 3D assets from images. This model utilizes a novel âO-Voxelâ structure, enabling the creation of complex 3D topologies with sharp features and full PBR materials, achieving up to
1536Âłresolution with16Ăspatial compression. The model is open-source, with code available on GitHub, and a live demo hosted on Hugging Face. There is a technical debate regarding the support for ROCm, as the current documentation primarily mentions CUDA. A user reported issues running the model on an AMD 7800XT GPU, experiencing segmentation faults likely due to dependency conflicts and ROCm overrides.- DeedleDumbDee discusses the challenges of running TRELLIS.2 on AMD hardware using ROCm, noting that the model documentation primarily supports CUDA. They mention encountering segmentation faults when attempting to run the model on a 7800XT GPU, which has only been tested on NVIDIA GPUs with 24GB VRAM. The issues are likely due to dependency conflicts and using ROCm overrides.
-
AMD Hipfire - a new inference engine optimized for AMD GPUâs (Activity: 426): AMD Hipfire is a new inference engine optimized for AMD GPUs, utilizing a unique
mq4 quantizationmethod. It is not officially affiliated with AMD but shows significant performance improvements, particularly on RDNA3 architecture. Benchmarks on Localmaxxing indicate substantial speedups, with one user reporting a2.86Ăspeedup on an RX 7900 XTX compared to baseline. The engine is actively being tested and models are available on Hugging Face. Some users express a preference for industry-standard quantization formats like GGUF, suggesting it would simplify adoption. Benchmarks indicate that while Hipfire excels in AR decoding, it lags in prefill compared to llama.cpp, with performance highly dependent on workload type, particularly benefiting structured/code generation tasks.- User alphatrad reports a significant performance boost using AMD Hipfire on an RX 7900 XTX, achieving a
2.86Ăspeedup with coherent output compared to the AR baseline. However, they note that real-world application might differ from speed tests, especially in coding tasks. - Own_Suspect5343 provides a detailed benchmark comparison between Hipfire and llama.cpp on AMD hardware. Hipfire shows a
30%faster AR decode than llama.cpp, but llama.cpp excels in prefill tasks. DFlash, a speculative decoding method, offers substantial speedups in structured/code generation tasks, with a3.45xspeedup on a merge_sort prompt, highlighting its workload-dependent performance. - FullstackSensei expresses a preference for industry-wide adoption of GGUF for model quantization, suggesting it would simplify compatibility issues across different platforms and models. This reflects a broader desire for standardization in AI model deployment.
- User alphatrad reports a significant performance boost using AMD Hipfire on an RX 7900 XTX, achieving a
Less Technical AI Subreddit Recap
/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo, /r/aivideo
1. Claude and GPT Image 2 Innovations
-
The Comeback Chatgpt Did with Image 2 Is Insane (Activity: 1035): The post compares two AI-generated images of a Bugatti Chiron in Dhaka, Bangladesh, using the same prompt. The first image is generated by Nano Banana pro, while the second is by ChatGPT Image 2. The latter is noted for its realism, with one commenter stating it âlooks like a real imageâ. However, the AIâs limitations are evident in the text on signboards, which appear as a mix of Bangla and Hindi scripts, highlighting challenges in accurately rendering localized text. Commenters highlight the realism of the ChatGPT Image 2 output, noting its photographic quality. However, they also point out the AIâs failure to accurately depict local scripts, which detracts from the authenticity of the scene.
-
Stanford researchers fed a language model a DNA sequence and asked it to create a new virus. It wrote hundreds of them, and 16 worked. One used a protein that doesnât exist in any known organism on Earth. (Activity: 1080): The image is a research paper titled âGenerative design of novel bacteriophages with genome language models,â which details a study by Stanford researchers who used AI to design new bacteriophages. The AI-generated sequences resulted in 16 viable phages, with one utilizing a protein not found in any known organism on Earth. This research underscores the potential of AI in synthetic biology, particularly in designing bacteriophages that could target drug-resistant bacteria, highlighting both the promise and risks of AI in bioengineering. Commenters express concern about the dual-use nature of AI in bioinformatics, noting its potential for both beneficial applications, such as targeted therapies against drug-resistant bacteria, and harmful uses, given the relatively low barrier to entry for creating novel viruses.
- Saotik highlights the dual-use nature of AI in bioinformatics, noting that while AI-generated bacteriophages could lead to targeted therapies against drug-resistant bacteria, the same technology could potentially be used to create harmful viruses that affect humans. This underscores the need for careful regulation and ethical considerations in the deployment of such technologies.
- AI-powered bioinformatics is compared to nuclear technology in terms of its potential impact. The commenter emphasizes that once the knowledge of creating novel viruses from AI-generated sequences is widespread, the barrier to entry becomes low, raising concerns about misuse. This points to the critical need for stringent controls and oversight in the dissemination and application of this technology.
- yamankara clarifies that the term âlanguage modelâ in this context refers to a âgenome language modelâ rather than a general-purpose large language model (LLM). This distinction is crucial for understanding the specific capabilities and applications of the model used in generating novel viral sequences.
-
Claude 4.7 named a journalist from 125 words of unpublished writing (Activity: 800): Kelsey Piper conducted an experiment with Claude 4.7, where she input 125 words of unpublished writing and the model identified her name. She ensured anonymity by logging out, using the API, and testing on a friendâs laptop, ruling out account, browser, and IP identification. The experiment suggests that Claude 4.7 can identify a writerâs unique âvoiceâ from a small text sample, a capability not matched by ChatGPT or Gemini. This indicates a potential for models to measure and recognize writing style as a distinct fingerprint, highlighting Claude 4.7âs superior reading ability, though it may be less flexible in generating prose, possibly due to its deep encoding of prose patterns. Read more. Some commenters question the technical rigor of Piperâs methodology, noting potential oversights like account sign-in requirements for API use. Others draw parallels to past instances of linguistic analysis, such as identifying authorship through unique word choices, suggesting that writing style can be as distinctive as a voice print.
2. DeepSeek and Qwen Model Discussions
-
deepseek reduces prices again, the price for input token cache hits is reduced to 1/10 of the current level (Activity: 356): DeepSeek has announced a significant price reduction for input token cache hits, decreasing the cost to
1/10of the previous level, from$0.145to$0.0145. This reduction is permanent, contrasting with a temporary discount offered the previous day. This move is expected to make DeepSeekâs services more cost-effective, particularly for applications requiring1Mcontext length, enhancing its competitive edge in the market. Commenters express surprise and appreciation for the price reduction, with some attributing it to strategic generosity from Chinese companies, while others highlight the potential impact on large context applications.- DeepSeekâs recent price reduction for input token cache hits to $0.0145 from $0.145 marks a significant cost efficiency, especially for handling 1M context lengths. This move could make large-scale model usage more accessible and affordable for developers and researchers.
- DeepSeek Flash is noted for its performance, being on par with state-of-the-art (SOTA) models across various tasks. This suggests that DeepSeek is not only competitive in pricing but also in technical capabilities, making it a viable option for high-performance applications.
- The drastic price reduction by DeepSeek is seen as a strategic move to make their services nearly free, potentially disrupting the market by offering high-quality models at a fraction of the cost, thus encouraging wider adoption and experimentation.
-
Is DeepSeek V4 Pro expensive (without discounts)? (Activity: 183): The image is a bar chart that compares the costs of running various AI models, highlighting that DeepSeek V4 Pro has a total cost of
$1071, which is moderate compared to other models like Claude Sonnet 4.6 at$3959and GPT-5.4 at$2851. The costs are broken down into input, reasoning, and output costs, with DeepSeek V4 Pro having$614for input and$420for reasoning. This suggests that while DeepSeek V4 Pro is not the most expensive option, it is still significant in cost compared to subscription models like Claude and GPT, which can be around50%cheaper. The discussion also touches on the unpredictability of subscription models and the potential benefits of local models. One commenter notes that DeepSeek V4 Pro is approximately2.4xmore expensive than Mimo v2.5 Pro, with Mimo leading in intelligence benchmarks and having a cost advantage due to lower verbosity. Another commenter expresses skepticism about the claim that subscriptions cost only10%of API usage, suggesting that local models might be a more stable investment.- zoser69 provides a detailed cost comparison between DeepSeek v4 Pro and Mimo v2.5 Pro, noting that DeepSeek is approximately 2.4 times more expensive. In terms of performance, Mimo leads in Intelligence by two points, while DeepSeek has a slight edge in Coding. Mimoâs lower verbosity score contributes to its cost-effectiveness, and both models have similar response speeds on OpenRouter, making Mimo more time-efficient.
- Old_Stretch_3045 highlights the unpredictability of subscription models, suggesting that reliance on cloud-based models might not be sustainable due to potential changes in service limits and costs. They advocate for exploring local models and upgrading hardware as a more stable alternative.
- zoser69 lists projected costs for various models based on a hypothetical 10% pricing criterion for frontier models. They estimate DeepSeek v4 Pro at $661.2, while other models like GPT 5.5 and Mimo v2.5 Pro are priced at $225 and $276, respectively. MiniMax m2.7 is noted as the most affordable model with above-average intelligence at $104.4, even without developer discounts.
3. AI Tools and Frameworks for Developers
-
LTX2.3 in Ostris Ai toolkit on a 5090 Training done in 7 hours ⌠I went Thanos way and I said fine ⌠Iâll do it myself (Activity: 616): The post details a custom training process for the LTX2.3 model in the Ostris AI toolkit using an NVIDIA 5090 GPU, achieving completion in 7 hours. Key adjustments include setting the
lora rankto 48 initially, using600 stepsfor the first phase, and employing agradient accumulationof 2. The process involves multiple phases with specific settings fordifferential guidance,learning rate, and dataset configurations, such as512x512resolution and25 framesper clip. The author emphasizes the importance of accurate prompts and trigger words for effective training. Commenters are intrigued by the ability to capture likeness with1-secondclips and request examples of prompts and captions used in training. There is also a discussion on the feasibility and effectiveness of using short clips for training Loras.- DateOk9511 outlines a detailed multi-phase training process for LTX2.3 using a 5090 GPU, emphasizing the importance of dataset composition and training parameters. The process involves using 25 video clips, each 1 second long at 25 frames per second, with specific settings like âlow VRAMâ, âlora rankâ, and âdifferential guidanceâ. The training is divided into four phases, each with distinct step counts and settings, such as adjusting âlora rankâ and âdifferential guidanceâ to optimize performance and accuracy.
- Disastrous-Agency675 shares a technique for optimizing VRAM usage during LTX2.3 training on a 3090 GPU by disabling sampling, which significantly speeds up the process. This approach allows for achieving 7000 steps in 6-7 hours by enabling VRAM offload and other low VRAM settings, suggesting a method to efficiently manage resources and improve training speed.
- Upper-Reflection7997 reports a failed attempt at LTX2.3 LORA training on a 5090 GPU with 64GB RAM, indicating potential challenges or misconfigurations in the setup. This highlights the variability in training outcomes based on hardware and configuration, and the need for precise tuning of parameters to achieve successful results.
AI Discords
Unfortunately, Discord shut down our access today. We will not bring it back in this form but we will be shipping the new AINews soon. Thanks for reading to here, it was a good run.