All tags
Topic: "bias-mitigation"
How To Scale Your Model, by DeepMind
qwen-0.5 google-deepmind deepseek hugging-face transformers inference high-performance-computing robotics sim2real mixture-of-experts reinforcement-learning bias-mitigation rust text-generation open-source omarsar0 drjimfan tairanhe99 guanyashi lioronai _philschmid awnihannun clementdelangue
Researchers at Google DeepMind (GDM) released a comprehensive "little textbook" titled "How To Scale Your Model" covering modern Transformer architectures, inference optimizations beyond O(N^2) attention, and high-performance computing concepts like rooflines. The resource includes practical problems and real-time comment engagement. On AI Twitter, several key updates include the open-sourced humanoid robotics model ASAP inspired by athletes like Cristiano Ronaldo, LeBron James, and Kobe Bryant; a new paper on Mixture-of-Agents proposing the Self-MoA method for improved LLM output aggregation; training of reasoning LLMs using the GRPO algorithm from DeepSeek demonstrated on Qwen 0.5; findings on bias in LLMs used as judges highlighting the need for multiple independent evaluations; and the release of mlx-rs, a Rust library for machine learning with examples including Mistral text generation. Additionally, Hugging Face launched an AI app store featuring over 400,000 apps with 2,000 new daily additions and 2.5 million weekly visits, enabling AI-powered app search and categorization.
One Year of Latent Space
gemini-1.5 gemma-7b mistral-next opus-v1 orca-2-13b nous-hermes-2-dpo-7b google-deepmind nous-research mistral-ai hugging-face nvidia langchain jetbrains ai-ethics bias-mitigation fine-tuning performance-optimization model-merging knowledge-transfer text-to-3d ai-hallucination hardware-optimization application-development vulnerability-research jim-keller richard-socher
Latent Space podcast celebrated its first anniversary, reaching #1 in AI Engineering podcasts and 1 million unique readers on Substack. The Gemini 1.5 image generator by Google DeepMind sparked controversy over bias and inaccurate representation, leading to community debates on AI ethics. Discussions in TheBloke and LM Studio Discords highlighted AI's growing role in creative industries, especially game development and text-to-3D tools. Fine-tuning and performance optimization of models like Gemma 7B and Mistral-next were explored in Nous Research AI and Mistral Discords, with shared solutions including learning rates and open-source tools. Emerging trends in AI hardware and application development were discussed in CUDA MODE and LangChain AI Discords, including critiques of Nvidia's CUDA by Jim Keller and advancements in reducing AI hallucinations hinted by Richard Socher.