**Honesty is all you need.**

AI News for 7/2/2024-7/3/2024. We checked 7 subreddits, 384 Twitters and 30 Discords (418 channels, and 2896 messages) for you. Estimated reading time saved (at 200wpm): 341 minutes. You can now tag @smol_ai for AINews discussions!

Arvind Narayanan et al published a paper about how Agent papers are mostly not reproducible and ignore cost, Meta published a text-to-3D assets model, Magic.dev and Poolside are code model companies seeking unicorn rounds, OpenDevin is now a company, Kyutai released a realtime Audio LLM that maybe doesn’t work as advertised, Peter Thiel backed some AGI Blockchain thing, The New Stack published one and two writeups of AIEWF.


{% if medium == ā€˜web’ %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

AI Model Releases and Updates

  • Meta 3D Gen: @AIatMeta introduced Meta 3D Gen, a new system for end-to-end generation of 3D assets from text in <1min, producing high-quality 3D assets with high-resolution textures and material maps. Details are available in the technical report.
  • Perplexity Pro Search Update: @perplexity_ai announced an updated version of Pro Search that can perform deeper research on more complex queries with multi-step reasoning, Wolfram|Alpha, and code execution.
  • Phi-3 Mini Update: @rohanpaul_ai shared that Microsoft updated Phi-3 mini with significant improvements in long-context understanding, instruction following, and structured output, all achieved by post-training improvements.
  • GPT4All 3.0: @andriy_mulyar announced GPT4All 3.0, supporting 1000’s of models and all major operating systems, with major UI/UX improvements and Local File Chat with LocalDocs.
  • Yi-Large Launch: @01AI_Yi celebrated one week since Yi-Large launched on the Fireworks AI Playground, asking for user feedback on the model.

Research Papers and Techniques

  • Reinforcement Learning from Human Feedback (RLHF): @cwolferesearch provided an overview of the evolution of RLHF research, tracing its roots to papers studying the use of human feedback for training summarization models. Key papers were linked.
  • Persona-Driven Data Synthesis: @rohanpaul_ai shared a paper proposing a persona-driven data synthesis methodology using Persona Hub, a collection of 1 billion diverse personas, to create scalable and diverse synthetic data for LLM training and evaluation.
  • Meta-tuning for Few-shot Generalization: @slashML shared a paper on ā€œUnleashing the Power of Meta-tuning for Few-shot Generalization Through Sparse Interpolated Expertsā€.
  • Steering Vectors: @sarahookr shared work on steering model behavior towards non-differentiable objectives by constraining the generation process to explicitly steer towards minimization or maximization of non-differentiable features.

Frameworks and Tools

  • LangSmith: @LangChainAI shared a case study on how @newcomputer used LangSmith to iterate quickly and improve memory retrieval, leading to 50% higher recall and 40% higher precision for their agentic memory system, Dot.
  • Qdrant Engine v1.10: @qdrant_engine released Qdrant engine v1.10 with new features like Universal query API, Multivector search, Inverse Document Frequency, and more.
  • Leap AI: @LeapAI_ introduced their platform for building custom AI workflows to automate content creation, lead generation, and more, integrating state-of-the-art AI models like GPT-4.

Discussions and Perspectives

  • Gain of Function Research with AI: @JvNixon expressed concern about ā€œgain of function researchā€ with AI, drawing parallels to bioweapons research and the potential dangers of creating teams trying to generate novel, dangerous outputs to prove whether models are safe or not.
  • Probability of Doom vs. Probability of Life: @JvNixon argued that framing AI risk in terms of p(doom) is a deep collective psychological mistake, forcing people to imagine abstract superintelligence. They prefer p(life) - the probability of you and your loved ones surviving into the far future - as it brings in more of life and progress, and forces a balance of risks against benefits.
  • Idle Compute in AI Labs: @far__el noted that many AI labs have lots of idle compute sitting around, as they need compute in bursts. This leads to things like heavily subsidized inference, redefining compute cost as a marketing expense.

AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Models & Techniques

AI Video & Animation

AI Ethics & Societal Impact

Miscellaneous


AI Discord Recap

A summary of Summaries of Summaries

  1. Real-Time AI Models Steal the Spotlight:

    • Kyutai Labs launched Moshi, a 7B multimodal model for real-time text and audio generation with 160ms response times, garnering excitement for its open-source availability and rapid interactions (albeit a bit robotic), showcasing during a demo session with plans to address minor bugs.
    • The Phi-3 Mini model received a major update akin to a 3.5 Mini, with upcoming Gemma 2 support, but users noted startup issues reflecting the integration challenges of cutting-edge AI tools.
  2. Optimizing AI Deployment and Memory Management:

    • Extensive discussions on Colab and Kaggle notebooks shared best practices for memory management with methods like gc.collect() and torch.cuda.empty_cache(). Scaling LoRA rank for models based on dataset size was debated, emphasizing optimization via efficient resource handling.
    • Gemma 2 support enhancements for tools like Unsloth and LM Studio improve finetuning speed significantly, with Unsloth achieving 2x faster finetuning and 63% less memory usage, while LM Studio’s 0.2.27 update solved compatibility issues on Mac, Windows, and Linux.
  3. Innovations in AI Model Training and Fine-Tuning:

    • QLoRA was highlighted for its efficient finetuning of quantized LLMs, enabling finetuning of 65B parameter models on 48GB GPUs with near 16-bit precision performance using 4-bit quantization, as detailed in the QLoRA paper.
    • Members delved into optimizing CUDA operations with tools like DeepSpeed and Inductor backend for Nvidia, focusing on autotuning GEMM backends and troubleshooting torch.cuda.OutOfMemoryError, reinforcing the importance of hardware-informed optimizations.
  4. Privacy, Security, and Ethical Considerations in AI:

    • Concerns over data policy enforcement led to critical discussions on OpenAI’s GPT-4 subscription pricing and sporadic model parameter adjustments affecting user experience. Issues like dataset removal due to minor policy breaches sparked debates on enforcement consistency vs. user needs.
    • Discussions on anti-AI art software like Glaze and Nightshade raised ethical questions about balancing copyright protection and technological progress, highlighting community frustrations over potential circumvention of protective tools.
  5. Community Tools, Tutorials, and Collaboration:

    • Users shared various open-source tools and tutorials, such as creating custom pipelines with Transformers and Gradio apps for role-play prompts, fostering collaborative learning and practical implementation.
    • Docker image development for AI tools like AI Town saw active community participation, focusing on simplifying setup processes and ensuring compatibility with various platforms via detailed PRs and documentation submissions on GitHub.

PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord

  • Phi-3 Mini’s Marvelous Makeover: The Phi-3 Mini model underwent a significant update, akin to a 3.5 Mini, with the Gemma 2 compatible quantized versions expected soon, as indicated in announcements on Unsloth AI
    • Feedback from users suggests both excitement and startup issues with the new Gemma 2 support in Unsloth, reflecting the teething problems of cutting-edge AI tool integration.
  • Moshi’s Melodic AI Mastery: Kyutai Labs launched ā€˜Moshi’, a 7B multimodal LM, generating high-quality text and audio in real-time, achieving 160ms response times and planned for open-source availability.
    • The AI community is abuzz over Moshi’s capabilities, with its RLHF fine-tuning, backend versatility, and the anticipation of upcoming updates.
  • Colab’s Capacity Climb: Newly shared Colab/Kaggle notebooks offer extensive dataset support and introduce improvements such as scaling LoRA rank based on model and dataset size, garner Unsloth community’s attention.
    • Members discussed best practices for memory management including gc.collect() and torch.cuda.empty_cache(), while acknowledging the need to pin resource-heavy notebooks for ease of use.
  • Secretly Secure Docker Deployments: Discussions ensued about secure secret management in Docker deployments, with community consensus settling on the use of --env-file flag for environmental variables as a best practice.
    • Suggestions circulated for efficient container handling and deployment, such as using local registries and Docker commands like docker save and ctr images import.
  • Tackling Unsloth’s Local Quirks: Users report configuration issues when utilizing Unsloth locally, with recommended fixes involving updates to the config object to reflect changes in API.
    • Although Gemma2’s anticipated update within 1-2 days stirred the community, ongoing discussions continue to highlight delays and the eager anticipation for improvements in PHI’s JAVA evaluations.

OpenAI Discord

  • GPT-4 Subscribers Grapple with Usage Caps: Users voiced concerns over GPT-4 subscriptions, facing issues like quickly reaching message limits and decreased performance after upgrading. The community exchanged alternative approaches and highlighted the constraints of the model.
    • Debates emerged on the subscription pricing, as some users paying up to $60 a month challenged OpenAI’s sporadic parameter adjustments, questioning the cost-effectiveness for professional tools.
  • AI21 Unleashes ā€˜Jamba’ with Hefty Hype: AI21 Labs introduced ā€˜Jamba’, boasting a hybrid of Mamba SSM technology and Transformer architecture, flaunting its 256K context window and competitive pricing, stirring animated conversations.
    • Discussions ensued on applying Jamba to coding tasks, with reports of mixed results compared to other AI models like GPT-4 and Claude, igniting dialogues on potential approaches for accuracy enhancements.
  • Open-Source AI Tools Enter the Fray: The release of ā€˜Moshi’, a tool for real-time AI-powered conversations that’s open-source, caught the interest of many, despite its early-stage limitations.
    • The community weighed the pros and cons of open-source AI tools against proprietary models, discussing how these developments could influence the incorporation of AI into everyday technology.
  • Prompt Engineering’s Depth Explored: Prompt engineering surfaced as a key topic, with members sharing advice on honing prompts for more precise task performance with AI, especially for nuanced tasks like creating PDFs with formatted product tags.
    • Users tackled the intricacies of DALL-E prompt engineering, offering recommendations like prompt simplification and specificity to mitigate issues related to undesired image elements.
  • Nested GPTs Spark Curiosity and Debate: In the realm of GPT development, a user’s query about the feasibility of a GPT calling other GPTs opened up a discussion on the technicalities and hypothetical depth of such nesting functionalities.
    • The community also expressed dissatisfaction with data policy enforcement, pointing out the removal of a dataset involving a 15-year-old entry and sparking a conversation on the need for nuanced compliance versus strict guidelines.

LM Studio Discord

  • LM Studio’s Latest Gemma 2 Enhancements: LM Studio 0.2.27 introduces improved support and compatibility for Gemma 2 models with enhanced performance on Mac, Windows, and Linux platforms. Users are encouraged to update to the new version for a seamless experience.
    • Community contributors like abetlen have been instrumental in updating Gemma 9B and 27B models, which can be redownloaded from Hugging Face, ensuring compatibility with current setups.
  • Sailing Smooth on ROCM Seas: A concerning error ā€˜unknown model architecture: gemma2’ has sparked conversation surrounding the new LM Studio 0.2.27 release, with proposed solutions including a clear cache or a complete re-install.
    • Community testing on the ROCM GPU compatibility performance suggests success on models like AMD Radeon RX 6900 XT, with prompts to assist in validating the latest Linux ROCm extension pack for the updated software version.
  • Resolving Power Puzzles: A deep dive into LM Studio’s energy usage revealed a high idle consumption, prompting discussions on power efficiency and comparisons with other tools like Blender that suggest a need for optimizations.
    • Contrasts between operating systems emerged, as Linux users noticed a gentler power draw from their GPUs when running models, compared to the power surges reported by Windows users amidst similar activity.
  • Scaling Battles and Interface Improvements: Feedback on LM Studio pointed out scaling issues on 1080p monitors, restricting workflow efficiency due to a cramped interface, and highlighting the importance of layout optimization in multi-display environments.
    • Users proposed adding metadata such as publication dates to model listings on LM Studio’s interface, a suggestion that garnered positive responses from the community.
  • Gradio App’s Role-Play Revolution: In pursuit of a richer role-playing experience, a user has pioneered a Gradio app with dynamic variables aimed at improving immersive character interactions, igniting a flame of innovation for AI-driven storytelling.
    • The application’s ability to offer tailored prompts places it at the forefront, receiving an invitation for community feedback to enhance its capability, viewable at this creative space.

HuggingFace Discord

  • Transformers 4.42: New Models on Parade: The Transformers 4.42 release debuts novel models like Gemma 2, improvements in tool usability, and fine-tuning capabilities, marking another stride in model progress.
  • Data Abundance: AWS Chronos Datasets Go Public: AWS releases comprehensive Chronos datasets on HF, complete with both pretraining and evaluation benchmarks, providing a rich resource for temporal analysis.
    • Researchers can dive into temporal patterns with the AWS datasets, potentially sparking data-driven insights and model innovations.
  • AI Expertise Development: Free Courses Emerge: Prominent institutions like Harvard University offer free ML courses, boasting quality content and a pathway to certification.
    • These courses are a gateway for those aiming to elevate their ML proficiency without financial barriers, though the repetitive nature of basics is a consideration for prospective learners.
  • Community Engagement: New Roles and Resources: HuggingFace’s Discord community strengthens with ongoing discussions on the capabilities of large context window models like Qwen2, indicating a heightened interest in nuanced text processing.
    • Comparisons draw between the efficiencies of HF models like Meta-Llama and proprietary giants, revealing a landscape where open models tackle the dominance of closed-source tools.
  • Diffusers vs. A1111: Model Quality Disputed: Running the same generation parameters, users report RealVisXL V4.0 Lightning falls short in quality when using diffusers compared to A1111, despite identical setup.
    • Discussion centers on the trade-offs in quality between different execution methods, critical for achieving desired model performance in photorealistic tasks.

Eleuther Discord

  • GPT-4’s Colossal Capacity: Nvidia’s Sneak Peek**: GPT-4’s speculated parameter range of 1.7 to 1.8 trillion raised eyebrows, dwarfing GPT-3’s 175 billion, with a discussion involving Nvidia, suggesting the company’s close ties due to hardware support, despite NDAs.
    • Practical applications of InstructGPT showcased efficiency leaps by 10X to 100X, credited to Reinforcement Learning from Human Feedback (RLHF), generating a buzz about its potential.
  • Scaling Law Skirmishes: Kaplan vs. Hoffmann Unraveled**: Community debates addressed the discrepancy in scaling laws posited by Kaplan et al. and Hoffmann et al., with new insights on last layer costs and warmup duration, detailed in an arXiv paper.
    • The conversation highlighted potential flaws in the PyTorch FLOP counter and the importance of accurate FLOPs calculation methods for model scaling.
  • Interpreting Interpretability: Sparse Circuits Come to Light**: The paper on EAP and integrated gradients inspired a probe into sparce feature circuits, an approach to dissect language model behaviors, aiming for a methodical interpretability pipeline outlined in this work.
    • The SHIFT method for classifier generalization stoked curiosity, suggesting fine-grained interpretability units could ablate extraneous features, drawing insights from human judgement.
  • Perplexity in Preprocessing: Navigating Long Documents**: Stellaathena’s config perplexity baffled others with its error in proof-pile, a stark contrast to the smooth operation with lambada_openai, sparking a conversation on ensuring efficiency and accuracy in model evaluations.
    • Technical chatter included the loglikelihood_rolling feature and its use in turning loglikelihood into loss values, as part of the forum’s continuous agility in model assessment.

Perplexity AI Discord

  • Trying Gemini 1.5 Pro: Users engaged in a discussion about Gemini 1.5 Pro, emphasizing its large context window and rapid response times. Recommended for its solid performance, the chatbot garnered positive feedback.
    • Concerns were also raised regarding Perplexity’s live internet access, with mixed experiences reported on its ability to pull real-time data, causing frustration among users.
  • Navigating GPT4o Access Troubles: Members highlighted challenges in accessing GPT4o freely, instead directing others to Bing chat and Claude 3.5 Sonnet as viable alternatives for free conversations, subject to usage restrictions.
    • The conversation included tips on Perplexity’s Pro subscription refund process, with advice tailored to various regions such as the EU, UK, and Turkey.
  • Mobile Mastery with Perplexity: Queries about Perplexity’s mobile app features were clarified with confirmation of the inclusion of Wolfram Alpha and code generation capabilities on iOS.
    • A discourse on the importance of mobile features indicated a keen interest from users in the accessibility of advanced tools on handheld devices.
  • Sonnet’s API Silence: Discussions revealed that Sonnet 3.5 is not supported by the Perplexity API, prompting users to consult the official model documentation for alternative options.
    • Further to API capabilities, inquiries surfaced regarding the potential to leverage Perplexity’s search engine through the API, with the community showing enthusiasm for access to these extended functionalities.
  • AI Blackbox Building Blocks: Instructions and principles for creating a blackbox system in AI were provided, offering guidance on constructing these complex systems.
    • Material on topics including the Lean Canvas and the founding of Perplexity AI were shared, contributing to a broader understanding of strategic planning and entrepreneurial beginnings in the tech field.

CUDA MODE Discord

  • CUDA Conclave Convenes: A CUDA-only hackathon hosted by Ash Vardanian features Chris Lattner and is scheduled for July 13th at the AGI House in San Francisco, offering hands-on experience with H100 accelerators. Details available here, courtesy of Nebius.ai.
    • In a separate event, Meta’s Hacker Cup 2024 gears up for a September 20th start, with Mark Saroufim urging developers to dive into the code generation challenge. Meanwhile, GPU enthusiasts are in a dilemma over the NVIDIA 3090’s $1,000 price tag, as shared by Mark Saroufim who snagged a 4090 for $1,200.
  • Matrix Multiplication Mastery: Mobicham surfaces a guide to achieving over 1 TFLOPS performance on matrix multiplication on a CPU platform, specifically tuned for the AMD Ryzen 7700, which surpasses NumPy’s offering. Tutorial can be found here.
    • The 3D V-Cache technology garners attention for its contribution to AMD Ryzen’s performance, sparking debates around its specialization beyond the augmented cache size, affecting clock speeds and silicon layering.
  • Integrator Ins and Outs: Conversations unfold about compiling functions in Pytorch using the Inductor backend for Nvidia, mentioning John Carmack’s commendations for the PyTorch team while delving into buffer loading and dequantization processes with torchao.
    • A hiccup in forcing Inductor to generate Triton kernels for all operations is discerned, where GEMM succeeds but Conv fails, as detailed in a GitHub issue seeking resolution.
  • Model Memory Marvel: Cutting-edge memory efficiency strategies put the limelight on this channel’s models which comfortably manage batch sizes that would see PyTorch balking, emphasizing on models’ memory savings.
    • A cited GitHub Pull Request #667 addresses decimal places in batch sizes during training which caused integer division errors, marking an incremental improvement.
  • Optimizer Odyssey: A wave of optimism is evident with Facebook Research’s schedule-free optimizers, claimed to demonstrate accelerated convergence across a spectrum of tasks, potentially reshaping optimization methodologies.
    • The community shares findings that suggest a significant uptick in the potential to fine-tune models without rigorous schedule adherence, teetering on the brink of what could be an optimization renaissance.

Stability.ai (Stable Diffusion) Discord

  • Artist’s Allies Artifice Abated: Community dialogue centered on the development of anti-AI art software like Glaze and Nightshade to protect artist’s copyrights, yet several members voiced concerns about the ease of bypassing such tools.
    • The conversation underscored the challenge of maintaining the balance between copyright protection and technological advancement in AI training.
  • Pixel Perfection Predicament: Inquiries regarding 16x16 pixel art led to recommendations for training at 512x512 resolution, despite Crystalwizard’s remarks about possible trial and error in search of efficiency.
    • Emphasis was placed on experimentation in training methods to hone image generation for this specific art style, underscoring the granularity of AI model training.
  • Discord’s Employment Depot Discussed: Threads emerged questioning if the server had a dedicated job-posting channel, highlighting a surge in demand for freelance and job opportunities within the community.
    • Separate discussions pondered the ethics and logistics of upwork account rentals among freelancers, reflecting on the gig economy landscape in tech.
  • Prompt Prowess & Performance Puzzle: Debates unfolded over various prompting techniques such as [A|B], C versus [A, B, C], evaluating their impacts on image outputs, particularly when using models like SD1.5 versus segmoe and MixofExperts.
    • Interest focused on refining techniques to achieve higher fidelity in text2img results, with discussions assessing the effectiveness of different syntactical approaches.
  • Model Melee: MixofExperts and segmoe: Community evaluations detailed segmoe model’s advancements in prompt understanding, showcased in applications like ComfyUI, and its perceived superiority over niche SD1.5 finetunes.
    • Comparative analyses by members illuminated the nuanced differences in performance and the quest for precision in natural language understanding among emerging models.

OpenRouter (Alex Atallah) Discord

  • Models Morphing on OpenRouter: OpenRouter announced changes including a significant update to the /models page, and adjustments in Google Token Sizes for Gemini and PaLM models—equating bigger tokens with GPT, and thus affecting pricing models.
    • A deprecation wave hits the OpenRouter: both the Default Model on settings page and custom auth headers for OpenAI API keys are set to be retired, steering towards newer practices and standards.
  • Claude 3.5’s Connection Conundrum: Users across the community have been experiencing 500 errors when working with Claude 3.5, prompting some to pivot temporarily to alternate versions, like Claude 3.0, for stability.
    • Discussions on the OpenRouter touched on privacy settings and logging policies with varied provider stances; NovitaAI and Infermatic stood out for their commitment to not retain data, as highlighted by Alex Atallah.
  • Discussing LLM Precision: AI Engineers speculated on the quantization of LLM models on OpenRouter, with debate centering around whether deployed models are using FP16 or remain in their original precision unless specifically altered by providers.
    • Alternative frontends for leveraging Claude models, like SillyTavern and LibreChat, were debated for their efficacy, with suggestions such as Typingmind and Pal Chat being proposed for enhanced engagement.

Latent Space Discord

  • Cash Infusion Without Code at Magic.dev: In a surprising financial leap, Magic.dev surges to a $1.5B valuation with a mere assemblage of 20 staff, void of any product or revenue trails.
    • Unprecedented capital raise earmarked to position the emerging company as a formidable contender in the AI domain, setting a new fundraising benchmark for startup ventures.
  • The Billion Persona Playbook Unveiled: Groundbreaking strides in synthetic data generation as Persona Hub integrates 1 billion personas, yielding impressive enhancements on benchmarks.
    • Aran Komatsuzaki heralds the methodology, spotlighting its potency in generating quality synthetic data and bolstering diversity.
  • Real-Time Audio LLM ā€˜Moshi’ Speaks Up: Moshi, heralded by Kyutai Labs, debuts as the inaugural real-time Audio LLM, demonstrating minimal latency yet slightly robotic articulation.
    • Despite its eagerness to reply causing occasional interruptions, the technology heralds a new frontier for user interactions with artificial intelligence.
  • All Hands on Tech: OpenDevin’s Fresh Initiative: The entrepreneurial minds behind OpenDevin forge All Hands AI, committing to democratize AI software development via open-source initiatives.
    • The platform’s foundation symbolizes a collaborative step towards universally accessible AI tools and a shared development ethos.
  • Sentient’s Seed Success: Funding the Open AGI Quest: Sentient announces an $85M seed influx, co-led by notables like Peter Thiel, to sculpt a community-driven AGI platform inviting global participation.
    • The ambitious funding is a clarion call for collective intelligence in creating an egalitarian AI ecosystem.

LAION Discord

  • Decentralized Transformers Gain Ground: jaan.li introduced their projects focusing on decentralized edge transformers at onefact.org and usb.club, sparking interest in their potential applications and contact for collaboration.
    • While san.tosh sought updates on open GPT-4o, the community’s anticipation remained, with ongoing discussions but no concrete news.
  • Terminator Model Scrutiny Rises: The community criticized the Terminator model’s insufficient ablation tests and urged for a substantial justification of its changes, with a strong call for presenting detailed studies.
  • Vision Transformers’ QKV Questioned: A debate emerged on the necessity of QKV within Vision Transformers, with hypotheses suggesting potential redundancies and a need for empirical evaluation.
    • Alternative theories were shared and craved a rigorous review to shed light on the full impact of attention mechanisms within such architectures.
  • FORA Forges Faster Diffusion Transformers: Introduction of FORA proposed to speed up Diffusion transformers by caching reusable computations, offering a solution to computational efficiency challenges.
    • The technique garnered attention for its potential to mesh with existing models deploying swift processing advancements as outlined in their repository.
  • HyperZā‹…Zā‹…W Paper Provokes Polarized Opinions: HyperZā‹…Zā‹…W paper was welcomed with mixed reviews, showcasing how a nascent submission can stir both acknowledgment and skepticism regarding new methods for SOTA achievements.
    • Despite criticism, there’s an aura of curiosity around the novel ideas and potential revisions flagged by the HyperZā‹…Zā‹…W paper, hinting at a growing discussion on QKV’s impact in ViT as per Schmidhuber’s survey.

tinygrad (George Hotz) Discord

  • Tinygrad’s UNMUL Caught in a RuntimeError: A RuntimeError was reported within tinygrad: ā€˜failed to render UOps.UNMUL’ with efforts led by George Hotz to assert the condition that should ā€˜never happen’.
    • Discussions unfolded about making loop collapse optional, hinted by flat_l4.realize(), to avoid user impact and highlighted by Chenyuy’s proposed workaround.
  • Fuzzy Frontend: Tinygrad’s Testing Takeover: Chenyuy floated the notion of a frontend fuzzer for tinygrad, geared to root out edge cases using an approach similar to porting torch code with LLM.
    • The community buzzed about creating minimal repro tests for certain dimensions to address heuristic boundary quirks, leaving PRs open for ongoing deep dives.
  • Debug Dash Before Tinygrad 1.0: The need for improved error messages in tinygrad crystallized with Yosifrost emphasizing pre-1.0 developer tool enhancements.
    • Community collaboration ensued to reproduce errors and devise test cases, setting the stage for more robust debugging mechanisms.
  • Gradient Gripes and Memory Mysteries: AI engineers exchanged experiences of gradient accumulation mishaps leading to CUDA out-of-memory errors, with tips like detaching loss circling the forums.
    • TinyJit’s shortcomings in optimization were highlighted, including TinyJit’s failure to use assert t.grad is not None effectively, provoking a swift community response.
  • Tinygrad vs PyTorch: Tensor Creation Quirks: The inconsistency of Tensor.randn/randint and Tensor.full between tinygrad and PyTorch sparked an analysis of tensor contiguity and proposals for alignment.
    • The behavior was chalked up as an idiosyncrasy unique to tinygrad, yet it didn’t stymie discussion on refining future iterations for better compatibility.

LlamaIndex Discord

  • Pinecone’s Predicament and Potential Pivots: A DocumentSummaryIndex creation snag hit users due to Pinecone limits, with a node’s oversized metadata and improper embed exclusion filters as culprits, detailed in this GitHub snippet.
    • Potential fixes include metadata limitation and seeking alternatives like qdrant or pg_vector, as one user suggested, showcasing the community’s problem-solving prowess.
  • RAG Revolution on Raspberry Rig: @pavan_mantha1 showcased a RAG pipeline functioning on a Raspberry Pi, utilizing Docker and Ollama, sparking intrigue on how compact setups can still deliver, specified in this community highlight.
    • This feat emphasizes the adaptability of AI systems to resource-constrained environments and captures the guild’s admiration for efficient computing.
  • Democratizing Documents with OpenContracts: OpenContracts emerged as an open-source marvel for document analytics, leveraging LLMs for annotations, enabled by Llama Index. The tool’s reveal is captured on Twitter.
    • GenAI native technology is at the forefront, with the project bidding to make AI-powered document handling widely accessible.
  • Weaving Wisdom with Webinar Wonders: Weights & Biases partners for a webinar aimed at enlightening on RAG pipeline construction, critically analyzing a year of development, as elaborated here.
    • The event is pivotal in addressing evaluation challenges, underscoring a commitment to growth and knowledge sharing in AI application.
  • Agentic RAG Rouses Readers: In the article Unleashing AI Potential, Agentic RAG couples with LlamaIndex and Claude-3.5 Sonnet over MongoDB, catalyzing conversations on avant-garde AI strategies.
    • Its imminent promotion signals a surge in interest for transformative approaches in AI infrastructures, ready to be explored by the keen minds of the guild.

Nous Research AI Discord

  • Tortoise-TTS Snaps to GGML: A community member has successfully migrated Tortoise-TTS to ggml, opening up opportunities for real-time text-to-speech operations. The repo is enhanced with CUDA and CPU support, giving developers a wider platform choice.
    • This move invites AI developers to dive into optimizing transformers and diffusion models to quicken the inference process, making this an engaging project for those keen on performance enhancements.
  • vLLM’s Tool Call Triumph in Hermes 2 Pro: The integration of tool calling in vLLM for Hermes 2 Pro has been executed successfully, bringing the project closer to the finish line. This development invites fresh conversations about the efficient handling of ā€˜content’ and ā€˜tool_calls’.
    • Discussions ensue around the incorporation of <scratch_pad> in Hermann 3 training, aiming at a more nuanced parsing methodology and aligning with standards akin to those seen in OpenAI’s framework.
  • Instructional Ingenuity from Genstruct 7B: The Genstruct 7B model, taking cues from Ada-Instruct, has made its mark by generating precise instructions from documents, thus facilitating the creation of tailored datasets for instruction finetuning.
    • Geared towards AI engineers, this technique brings to the forefront the fusion of raw text corpora into conversational datasets, providing an intelligent solution for dataset expansion without hefty investments.
  • CommandR Rises in Huggingface’s Hands: Huggingface raised a pull request for Cohere’s CommandR, introducing advancements that refine tool-use and retrieval-augmented generation (RAG) techniques.
    • Their creative input revamps the system prompt using a combination of a preamble and smart content organization, facilitated by Jinja templates, indicating a strong collaboration potential in RAG developments.
  • GraphRAG: Graph-based Genius by Microsoft: Microsoft has unveiled a novel retrieval-augmented generation framework known as GraphRAG, focusing on modular designs to uplift efficiency in information retrieval and content generation.
    • Accessible on GitHub, GraphRAG stands as a signature offering thorough customization options which are imperative for today’s dynamic AI research and development landscape.

Modular (Mojo šŸ”„) Discord

  • Mojo on Ubuntu: Installation Tango: Users faced hurdles with Mojo on Ubuntu 24.04/Python 3.12.3, encountering compatibility issues, particularly with max-engine. A step-by-step guide for a successful installation with Python 3.11 was shared.
    • Discussion centered around List[String] lacking the Stringable trait, impacting printability, with detailed references on GitHub. Users noted variable startup times in programs due to loop unrolling and its compilation time.
  • Strassen’s Splendid Speed? Not Quite Stable: The Strassen Algorithm was outperformed by a naive vectorized approach, hitting 70 GFlops over Strassen’s 50 GFlops on 1024x1024 matrices, as per discussions and benchmarks shared on GitHub.
    • Concerns were raised over its numerical stability, with potential instability leading to test failures when adjusted for different types and sizes of matrices.
  • SPIRAL: Spinning New High-Performance Code: The SPIRAL project aims to automate the development of DSP algorithms, at times surpassing the performance of MKL. It’s tailored for direct hardware tasks and could be key for optimizing an array of numerical operations.
    • Discussions highlighted the complexity of optimizing algorithms beyond parallel processing and vectorization, hinting at cache locality benefits from recursive approaches over iterative ones.

Interconnects (Nathan Lambert) Discord

  • Apple Cracks OpenAI’s Boardroom Door: Bloomberg reported that Apple will secure a board observer seat at OpenAI, with Phil Schiller set to take the position, signaling strategic moves in tech collaborations.
    • Community analysis suggests Apple’s partnership could yield greater perks than Microsoft’s investments, spotlighting benefits like exclusive app integrations and piquing debates on corporate strategies in AI advancements.
  • Moshi Masters Multimodal Mantra: Kyutai Labs stunned audiences with Moshi, a trailblazing real-time audio LLM boasting 150ms latency, as acclaimed during its presentation, where it demonstrated superior simultaneous translation abilities and was recognized for its speed and multimodal prowess.
    • Plans to publish open models for community innovation were commended, including Moshi’s core 7B multimodal LM and VQ-VAE codec, which are poised to redefine on-device interactivity and user experience.
  • Code’s Constitutional Conundrum: Debaters invoked the EFF’s perspective on SB 1047, examining the defense of model weights and code as speech, drawing parallels to freedom of expression and 3D gun design precedents.
    • Discussions surged around the essence of model weights as a form of expression, questioning if these algorithmic outputs should enjoy similar protections as language, emphasizing their integral role in modern communication and innovation.
  • Claude 3.5 Grows Its Fan Club: A surge of excitement razed through the community with the release of Claude 3.5, drawing enthusiastic responses and comparisons with previous iterations, with professionals noting leaps in performance and potential application areas.
    • Advocacy for Claude TM likened its market positioning to the successful strategies of well-known brands, with members urging a boost in promotional efforts to match its reputable counterparts and to emphasize its enhanced capabilities.

LangChain AI Discord

  • Azure Agonizes with 429 Aches: Switching to AzureAIDocumentIntelligenceLoader from PyPDFium2Loader led to a consistent 429 error (Too Many Requests), highlighting the rate limiting challenges faced.
    • Community debates included finding ways to circumvent Azure’s rate limiting without sacrificing efficiency or accuracy.
  • PDF Puzzles & Markdown Mysteries: Efforts to transform PDFs into markdown via marker stumbled when facing complex table formats, with merged cells causing major migration malaise.
    • The allure of an open-source tool persists despite Azure Document Intelligence offering superior parsing precision, prompting a search for a local solution.
  • LangSmith’s Lost Link: Reports surfaced of LangSmith unexpectedly halting call traces, sparking discussions on the robustness of LangChain’s introspective faculties.
    • Technical scrutiny ensued as users worked to detect defects in the tracing mechanism, hinting at hidden bugs in LangChain’s infrastructure.
  • CriticGPT Cornering Coding Errors: The AI community dissected OpenAI’s CriticGPT initiative, aimed at identifying and amending mistakes from GPT-4, with a digestible video explanation circulating among peers.
    • Enthusiastic dialogues unfolded around how CriticGPT marks advancement towards self-correcting AI systems, envisaging upgrades in automated code reliability.
  • Mac Meets Toolio: Open Source Serendipity: Mac enthusiasts rejoiced as Toolio broke onto the open-source scene, promising private LLM deployment on macOS, as heralded in its YouTube showcase.
    • The innovation empowers users with fast inference and JSON schema output, tuning into the demands for enhanced control and personalization.

Mozilla AI Discord

  • Beef Up Your llamafile Linux Rig: For optimal llamafile performance, engineers recommend GPUs like 3090/4090 for personal projects or A6000/RTX 6000 Ada for professional environments; and CPUs such as older EPYC for their superior core counts and PCIe support.
    • Discussions indicated a preference for GPUs with substantial VRAM, highlighting that 24GB VRAM is necessary to manage models around the size of 33B parameters.
  • VRAM: The Bigger, The Better: AI enthusiasts stressed the importance of excess VRAM to run sizeable models, with a cautionary note on employing FP16 mode as it ramps up VRAM usage compared to its minor quality gains.
    • Community exchanges underscored q4 configurations that smoothly handle 33B parameters with 24GB VRAM, setting a benchmark for large model management.
  • CPU Inference Wizardry with Syncthread: Creative uses of the syncthread trick for CPU inference were spotlighted, potentially changing the way we approach CPU-based learning.
    • Links to a YouTube talk detailed the technique, capturing the community’s attention.
  • Threadripper Tames llama3’s 70B Model: A sophisticated AI engineer reported on the successful operationalization of the llama3 70B model using a powerhouse Threadripper CPU, indicating potential leaps in CPU realistic applications.
    • This successful deployment signifies Threadripper’s ability to hold its own in an arena dominated by GPU prowess.
  • Navigating llamafile on RK3588 NPU Challenges: The integration of llamafile with Rockchip RK3588 NPU hardware sparked inquiries among practitioners, advising on software versions like v0.8.9 to circumvent compatibility issues.
    • This discussion points to broader challenges and considerations necessary when leveraging specific versions for optimal hardware performance.

Torchtune Discord

  • Weighing in on phi mini’s new payload: The phi mini has been updated with new weights, yet maintains consistency with its original repository, raising questions among users regarding the necessity for adjustments in the torchtune process.
    • Speculations persist on whether the legacy methods will hold up, but the consensus seems to lean toward a smooth transition without requiring significant changes.
  • Gradients & Epochs: Torchtune’s Training Twists: A vibrant discussion ensued on optimal training strategies, contrasting the use of gradients 8 vs 16 and whether batch size adjustments, along with epoch variation, might yield superior outcomes.
    • To assist in unraveling this conundrum, Wandb was employed to track and log performance metrics, with community members sharing insights to refine the training process.
  • Conversion Conundrums: HF Format Finesse: Queries have arisen about the nuances of model conversion, particularly why parameters like num_heads, num_kv_headers, and dim are requisite when transitioning between the multihead formats used by HF and torchtune.
    • The inherent complexity of format conversion was highlighted as members exchanged tips on effectively navigating this technical terrain.
  • Checkpoint Champion: Torchtune’s Savior: The introduction of FullModelHFCheckpointer into torchtune has sparked interest for its ability to seamlessly translate models into HF-friendly formats.
    • This tool has been lauded for bridging compatibility gaps between diverse machine learning infrastructures, ensuring broader accessibility and utility.

Cohere Discord

  • Checkmating Challenges with Stockfish & LLMs: Community members are exploring the combination of Stockfish game data with LLMs to enhance strategic reasoning capabilities, with a side notion of developing a swift chess engine.
    • Discussions unfolded around the technical hurdles of fine-tuning LLMs with chess data, debating over its practicality and the risk of overfitting. The theory of using existing tools like Stockfish within LLMs was met with promising interest.
  • Slack Bot Draws Cohere: A novel Cohere Slack bot was crafted, showcasing the ability to swiftly handle Slack’s 3-second request demand, a testament to Cohere’s API efficiency.
    • The creator’s offer to share their code and produce documentation has sparked enthusiasm within the community, with many looking forward to detailed guidance on integrating Cohere with communication platforms.

OpenInterpreter Discord

  • Sound of Speed: Kyutai Moshi’s Audio LLM: Kyutai Moshi released a real-time Audio LLM that operates with virtually no delay, though feedback highlights a somewhat robotic tone. It’s been heralded for fast interactions, sometimes eager to the point of interruption.
    • Insights from user Mikebirdtech underscore the system’s speed, expressing it’s almost too fast as it can interrupt users during natural conversational pauses.
  • See-Through Intelligence: OI Glasses Concept: In a speculative conversation, user johnlenflure sparked the idea of integrating OI into eyewear, envisioning a future with smart glasses bolstered by OpenInterpreter capabilities.
    • No further details or technical discussion followed, leaving the concept at a high level of abstractive interest among members.
  • Game On for Open Interpreter Mods: User Nonadjective.eth_55058 is seeking advice on integrating Open Interpreter into a game, aiming to develop a working proof of concept, even if initially clunky.
    • This reflects a growing interest within the community to explore and expand the modding potentials for Open Interpreter, indicating a trend towards customizable interactive experiences.
  • Project Compatibility with Open Interpreter: A list of projects, including Open interpreter, taxyai, clickolas cage, self-operating computer, pywinassistant, and GPT computer assistant, were highlighted as compatible with Open Interpreter.
    • Interest in exploring and possibly configuring these projects to work in tandem with Open Interpreter was evident, suggesting a dynamic and collaborative environment for developers.

OpenAccess AI Collective (axolotl) Discord

  • Quantization Quandaries: LoRA vs QLoRA: Members delved into quantization, discussing the diversity in its application between LoRA and QLoRA, highlighting that LoRA leverages 8-bit quantization and QLoRA pushes the envelope further with 4-bit, citing the comprehensive treatment in the QLoRA paper.
    • A dialogue clarified QLoRA’s positioning as superior in finetuning 65B parameter models on a single 48GB GPU with finesse, aligning performance closely with 16-bit finetuning, as revealed in the /QLoRA: Efficient Finetuning of Quantized LLMs/ paper.
  • VRAM Vexations and CUDA Calamities: Colab conundrums surfaced with a user struggling with torch.cuda.OutOfMemoryError, noting the attempt to allocate 172.00 MiB on Google Colab resulted in failure.
    • Contributors concurred on VRAM being the bottleneck and suggested an increase in VRAM to facilitate seamless operation, spotlighting the hardware’s vitality in the running of models like axolotl.

AI Stack Devs (Yoko Li) Discord

  • Docker Docking at AI Town: Community excited about a Docker image for AI Town, with a call for contributions to enhance the tool’s accessibility.
    • The Docker effort seeks to streamline the setup process, as enthusiasts recommend pushing a well-received Windows WSL setup guide as a pull request to the main repository.
  • API-port Drama in Dockertown: A savvy developer encountered API communication issues while porting AI Town to Docker, specifically with Ollama API, and is committed to sharing a fix soon.
    • The technical hurdle doesn’t deter progress as the port gains traction, and the community remains watchful for updates to ensure seamless connectivity.
  • Convex Catches a Docker: In an effort to simplify the AI Town experience, a member is tweaking Docker to automatically download Convex, anticipating a smoother ride for future users.
    • The automated Convex setup via Docker is expected to be operational by 8 p.m. UTC+4, indicating proactive community involvement aimed at user efficiency.
  • AI Town’s Docker Test Fest: A member’s initiative to run Docker integration tests on their Legion Go setup has led to confidence in the port’s performance, suggesting readiness for a pull request.
    • Volunteers were sought for Docker integration testing with the expectation to merge successful results, demonstrating the collaborative ethos of the AI Town developer community.

LLM Finetuning (Hamel + Dan) Discord

  • Gradio Gridlock Grapples Engineers: Members face hurdles deploying a RAG app using Gradio on Modal. A discussion sprouted about it working locally but not on Hugging Face Spaces.
    • Modal Slack was suggested as an emergency eject for the issue, hoping community support could provide a fix for the deployment dilemma.
  • DeepSpeed Dilemma Draws Debate: Configuring DeepSpeed stirs up a storm among members attempting to enable data sharding without opting into model sharding, as seen in their exchange.
    • Clarification and assistance with DeepSpeed settings became a pressing concern, highlighting a knowledge gap that needs bridging.
  • Hugging Face Handover Headache: Troubles were aired over the inability to share private code deployments on Hugging Face, with sharing=True unsupported in private spaces, discussed here.
    • Frustrations flared as attempts to operate on Modal also hit hitches, sparking a search for alternative methods for private code collaboration.

LLM Perf Enthusiasts AI Discord

  • Legal Eagles Eye LLM Precision: A new report from Screens analyzes LLM performance in contract reviews by equating it to a ML classification problem, boasting a 97.5% accuracy rate for their system.
    • The challenges of assessing long-response accuracy are addressed, suggesting a classification-based methodology could enhance LLM effectiveness in legal tasks such as negotiation and document summarization.
  • Prompt Tuning For The People: Evan_04487 is seeking a straightforward, hosted prompt-tuning tool that’s accessible to non-tech experts like designers and managers to run prompt variations and review outcomes.
    • The ideal solution would be a freemium service, easy enough for low-stakes, with the capacity to juggle about two dozen variables, contrasting the complex, self-managed infrastructure meant for critical tasks.

Datasette - LLM (@SimonW) Discord

  • Datasette Discovery in Data Journalism: Derek Willis shared an article about foreign gifts, sparking interest in Datasette’s utility for investigative journalism.
    • The discussion involved how Datasette can be leveraged as a powerful tool for sifting through public records and datasets, emphasizing its role in transparency and accountability in journalism.
  • Datasette’s Deep Dive into Data: Enthusiasts highlighted Datasette’s implications for deep data analysis, considering the tool’s capacity for handling complex queries.
    • Engineers discussed the potential for Datasette to transform data-driven stories, underscoring the importance of accessible and interpretable public data in the digital age.

PART 2: Detailed by-Channel summaries and links

{% if medium == ā€˜web’ %}

Unsloth AI (Daniel Han) ā–· #general (201 messagesšŸ”„šŸ”„):

  • SEQ_CLS support in unsloth
  • Using jellyfin as an alternative to Plex
  • Fine-tuning models on another language
  • Sharing Colab notebooks and account information
  • VRAM requirements for LORA fine-tuning
  • Phi-3 Mini gets a huge update: The Phi-3 Mini model received a significant update, likened to a 3.5 Mini, with new quantized versions expected to be available tomorrow including Gemma 2 support. It’s noteworthy that Phi-3 Mini’s latest enhancements will improve its performance, enabling quick and efficient processing.
  • Moshi’s real-time voice model excites AI community: Kyutai Labs introduced ā€˜Moshi’, a 7B multimodal LM that generates high-quality text and audio with low latency, achieving a 160ms response time. The model is fine-tuned with RLHF and is designed to be open source, supporting various backend configurations with plans for future updates.
  • Gemma 2 support added to unsloth: The unsloth team announced the release of Gemma 2 support, allowing users to fine-tune the model for advanced AI tasks with improved efficiency. Initial user feedback notes that Gemma 2 is already working well via the provided notebooks, although some encountered initial setup issues.
  • SEQ_CLS support and fine-tuning in unsloth: Users discussed the functionality of SEQ_CLS support in unsloth for fine-tuning tasks, recommending JSON outputs for multi-class classifications. Experience with the Phi-3 model suggests significant improvements and learning speed using this approach.
  • Discussion on integrating graph-based RAG into unsloth: There was interest in integrating Microsoft’s graph-based Retrieval-Augmented Generation (RAG) system into unsloth to enhance its capabilities. Users speculated on benefits, highlighting advancements in AI and optimized workflows.

Links mentioned:


Unsloth AI (Daniel Han) ā–· #announcements (1 messages):

  • Gemma 2 release
  • Phi 3 mini update
  • Finetuning improvements
  • Increased context lengths
  • Notebooks and 4-bit models
  • Gemma 2 Boosts Finetuning Speed: Unsloth now supports Gemma 2 with finetuning that’s 2x faster and uses 63% less memory. Check out the blog post.
  • Extended Context Length Achieved: You can finetune Gemma 2 (27B) with QLoRA to 9.7K context lengths with Unsloth in a 40GB GPU, while HF+FA2 only allows 3K lengths. Unsloth also supports 11K context lengths on a 24GB card for the 9B model.
  • Free Colab Notebooks Available: Free notebooks for Gemma 2 (9B) and 27B are available, including a Colab notebook for the 9B model.
  • Phi 3 Mini Gets an Update: Phi 3 mini also saw an update, with new Instruct model available on Hugging Face.
  • Experiment and Share Results: The community is encouraged to experiment, test, and discuss results of their models on the Unsloth platform. A specific call to share results was made @here.

Links mentioned:


Unsloth AI (Daniel Han) ā–· #off-topic (3 messages):

  • mahiatlinux's comment
  • theyruinedelise's reaction
  • response from mahiatlinux
  • Positive Engagement on Channel: A member expressed their amusement by saying ā€˜That’s good lol’, which sparked a short, positive exchange.
  • Surprise and Agreement: Another member agreed with the positive sentiment and mentioned being ā€˜surprised to see it so good’, which was further acknowledged.

Unsloth AI (Daniel Han) ā–· #help (63 messagesšŸ”„šŸ”„):

  • Llama.cpp quantization issues
  • Loading model speed on Colab
  • Unsloth compatibility with Gemma 2
  • Inference on CPU after fine-tuning with Unsloth
  • Training issues with Huggingfaces SFTTrainer and Unsloth
  • Llama.cpp quantization puzzles members: A user asked about unresolved quantization issues with llama.cpp, and another committed to prioritizing checking these issues today. Another user shared a related issue and linked to a relevant Discord message.
    • Still, get the same issue even after following the instructions.
  • Speed up model loading on T4 GPUs: Users discussed how to speed up model loading times on Colab T4 GPUs, specifically for a 7GB model. Suggestions included keeping the model in VRAM to avoid repetitive loading times, although some believed this was not possible on Colab.
    • ā€œEvery time when I am running cuda code to load model (7gb) from disk to gpu memory, it takes around 30 seconds. Can I make my model load faster or even better load model only once?ā€
  • JIT libraries with Jax: A user asked about using JIT libraries with Jax for training on a prompt-answer dataset requiring linear algebra knowledge for GPU optimization. They also queried whether RAG is a better option than traditional fine-tuning, given extensive training times.
    • Am I right in saying that? Not sure if someone has done that.
  • Unsloth adds Gemma 2 support: Users discussed various errors and issues related to Unsloth. Notably, it’s been mentioned that Unsloth has recently added support for Gemma 2.
    • ā€œJust now Unsloth added support of Gemma 2, you can update and try again!ā€
  • Issues with training using Huggingfaces SFTTrainer and Unsloth: A user shared an error while trying to start training using Huggingfaces SFTTrainer and Unsloth locally, which led to a permissions issue with their user. Another referenced a GitHub issue related to missing Python.h.
    • Is this in local? Best bet is just to google it.

Links mentioned:


Unsloth AI (Daniel Han) ā–· #community-collaboration (241 messagesšŸ”„šŸ”„):

  • Tutorial/Intermediate Colab/Kaggle notebook with more dataset support
  • Improvements and suggestions for the community notebook
  • Memory management and optimization techniques for notebooks
  • Text classification optimized notebook for Unsloth
  • Secret management in Docker and application deployment
  • Intermediate Colab Notebook by flail_: A member introduced a tutorial/intermediate colab/kaggle notebook featuring support for many datasets (including dolphin, ultrachat, capybara, slimorca) and auxiliary functions with the capability of scaling LoRA rank according to model and dataset size.
  • Text Classification Notebook by timotheeee1: A separate modified Unsloth notebook optimized for text classification was shared, enabling efficient evaluation using batched inference, and improving stability by fitting the classification head to specific token labels.
    • The notebook features input trimming for classification tasks, which saves VRAM and incorporates ignore_index to direct the model’s capacity to necessary predictions without wasting resources.
  • Memory Management Techniques Discussed: Conversations included incorporating gc.collect() and torch.cuda.empty_cache() in loops to handle out-of-memory issues, with various members sharing their methods of managing memory more efficiently.
    • Members debated the usefulness of sleep intervals and loop constructs for memory clearance, fueled by code snippets from Unsloth.
  • Docker Secret Management: A member sought advice on handling secret management when deploying Docker containers without uploading .env files directly, eventually resolving to use the --env-file flag for passing environment variables securely.
    • Different approaches, like leveraging a local registry and using docker save my-app > my-app.tar followed by ctr images import, were discussed to ensure secure and efficient deployment workflows.
  • Community Notebook Pins and Improvements: It was recommended to pin significant notebooks, such as those for dataset support and text classification, to avoid losing them in scrollback and improve accessibility for community members.
    • These resources are also planned to be added to the Unsloth GitHub page upon further review and refinement.

Links mentioned:


Unsloth AI (Daniel Han) ā–· #research (10 messagesšŸ”„):

  • Issues with using Unsloth in a local system
  • Gemma2 update release timeline
  • Support for the latest Gemma
  • Discussion about Gemma
  • Evaluation of Java with PHI
  • Unsloth config error in local setup: A user reported an error with config settings when using Unsloth on a local system: config.hidden_act is ignored, use config.hidden_activation instead.
  • Gemma2 update expected soon: A member mentioned that the update for Gemma2 is expected to be released in 1-2 days.
  • Latest Gemma not supported yet: The latest version of Gemma is currently not supported, necessitating a wait for the new update.
  • Gemma release issues: A member highlighted that Gemma should be released today, but there are delays due to blog issues.
  • Java Eval with PHI noted: A user remarked on PHI’s strong performance for Java, hitting 93 in evaluation.

OpenAI ā–· #ai-discussions (194 messagesšŸ”„šŸ”„):

  • Issues with OpenAI GPT-4 subscription and performance
  • AI21's Jamba model announcement and discussion
  • User experiences with AI for coding and programming
  • Live and open-source AI models debate
  • AI for real-time conversations: Moshi demo
  • Subscription woes plague GPT-4 users: A user expressed frustration over the GPT-4 subscription, citing difficulties with reaching the message limit and performance issues post-upgrade. The community suggested alternative problem-solving approaches and highlighted the model’s limitations.
  • AI21’s Jamba promises high-tech benchmarks: AI21 Labs announced ā€˜Jamba’ with a cutting-edge hybrid architecture, combining Mamba SSM technology and Transformer architecture, touting a 256K context window and attractive pricing.
  • Coding hurdles with AI models: Discussions revealed challenges with using various AI models for coding tasks—particularly in producing correct and complete code. Users shared mixed experiences with GPT-4, Claude Sonnet 3.5, and Jamba, reflecting varied performance in task accuracy.
  • Open-source real-time AI tools: The newly released ā€˜Moshi’ real-time AI conversation tool sparked interest for its open-source promise, despite some limitations in its current model’s intelligence.
  • AI race heats up with open-source competition: As discussions on Moshi and other open-source capabilities emerged, users debated the competitive edge these tools have over proprietary models like OpenAI’s offerings. The race to integrate advanced AI into daily tech highlights the evolving landscape.

Links mentioned:


OpenAI ā–· #gpt-4-discussions (11 messagesšŸ”„):

  • New TTS model voices availability
  • Data policy enforcement
  • Running ChatGPT in command prompt with Google search capability
  • Subscription pricing frustrations
  • Nested GPTs functionality
  • Data Policy Enforcement gets Critiqued: virtue signalling around data policy enforcement draws frustration as ChatGPT flagged and removed a dataset due to a single entry of a 15-year-old, despite the data having no inappropriate activity. Users are upset with the shift in message from ā€œWe might have made a mistakeā€ to an authoritarian ā€œWe are right you are wrong REMOVED!ā€
  • Run ChatGPT in Command Prompt with Google Search: A user successfully managed to run ChatGPT through the command prompt, enabling it to perform Google searches using Python coding.
  • Subscription Pricing Frustrates Users: Users express frustration over OpenAI’s ability to change model parameters sporadically despite some paying up to $60 a month for the service. There are mixed opinions on the value of professional tools costing $60, with some seeing it as justifiable for doubling productivity.
  • Query on Nested GPTs: A user inquired about the possibility of having a GPT call on other GPTs and wondered about the limitations on how deep this nesting could go.

OpenAI ā–· #prompt-engineering (117 messagesšŸ”„šŸ”„):

  • Issues with GPT models not answering questions as expected
  • Difficulties with creating PDF documents using GPT APIs
  • Improving prompt engineering for better task performance
  • Challenges with AI-driven image generation using DALL-E
  • Developing an employee recognition program using AI prompts
  • GPTs Troubles with Instruction Following: A user expressed frustration with GPT models not properly following detailed instructions, resulting in skipped steps and incomplete processes. Suggested solutions include updating prompts, using the regenerate option, or structuring instructions in smaller, sequential parts.
  • Manual PDF Creation from AI Output: A user faced issues generating a PDF with formatted product tags using AI, citing problems with adding logos and adjusting text size automatically. They opted to split the task into smaller manual edits after AI failed to meet their needs.
  • Improving DALL-E Prompt Engineering: A user sought assistance with improving prompts for generating vector icons using DALL-E, reporting issues like unwanted shadows and extraneous elements. Advice included simplifying and clarifying the prompt to avoid conflicts and ensure precise output.
  • Detailed Prompt for Employee Recognition Program: An elaborate prompt was shared for developing an employee recognition program, with clearly defined goals, recognition methods, criteria, an implementation plan, and a feedback mechanism. This structured approach aims to create a comprehensive and effective program.

OpenAI ā–· #api-discussions (117 messagesšŸ”„šŸ”„):

  • GPT performance issues
  • Improving prompt structure and attention control
  • Converting documents into PDF with product tags
  • Enhancing an AI icon generator
  • Developing an employee recognition program
  • Address GPT Performance Issues with Structured Prompts: Members discussed improving GPT performance by structuring prompts effectively, such as using conditional imperatives like IF...THEN... and including clear, strong attention control mechanisms in the prompts.
  • Tips on Converting Documents into PDF with Product Tags: The conversation covered challenges in converting documents into PDFs with product tags, specifically regarding issues with fitting text and logos into predefined document rectangles.
  • Enhancing AI Icon Generator Output: A user sought help for an AI icon generator, highlighting issues such as unwanted background elements, shadows, outlines, and random objects being generated despite clear prompt instructions.

LM Studio ā–· #šŸ’¬-general (166 messagesšŸ”„šŸ”„):

  • Discussing technical issues and updates related to LM Studio.
  • Comparing different AI models like Gemma 2, Llama 3, and Mistral.
  • Enhancements and bugs in the Gemma 2 model including tokenizer and attention mechanisms.
  • Difficulty and recommendations for fine-tuning LLMs using LM Studio.
  • Best models and configurations for different hardware setups.
  • Technical Issues and Updates in LM Studio: Users reported and discussed various technical issues related to LM Studio, including problems with CPU utilization, difficulties with different versions, and challenges with certain prompts and response generation.
  • Gemma 2 and Other AI Models: User Experiences: Members shared their experiences with the Gemma 2 model, noting significant improvements in writing style and performance, despite some lingering issues like ignoring system prompts. Comparisons were made with other models like Llama 3 and Mistral, with varied opinions on their effectiveness.
  • Enhancements and Bug Fixes in Gemma 2: Discussions highlighted the latest updates to the Gemma 2 model, including attention layer fixes and tokenizer improvements. Users debated the necessity of redownloading models to benefit from these updates.
  • Fine-Tuning and Model Recommendations: Users expressed interest in fine-tuning LLMs for specific tasks, though it was clarified that LM Studio currently does not support fine-tuning. Recommendations were given for using other methods and tools like RAG for specific customizations.
  • Best Models for Different Hardware Setups: Recommendations were given for using the best models based on hardware configurations, with suggestions to use models that allow full GPU offloading for laptops and other devices with limited VRAM. Specific suggestions included Gemma 9B for 7B-size requirements and the unreliability of current gaming laptops for running LLMs effectively.

Links mentioned:


LM Studio ā–· #šŸ¤–-models-discussion-chat (52 messagesšŸ”„):

  • dolphin-vision compatibility in LM Studio
  • Gemma 2 model performance and issues
  • System and hardware requirements for running large models
  • RP stress testing for AI models
  • Code generation capabilities of Gemma 2
  • Dolphin-Vision compatibility questioned: A user questioned whether dolphin-vision works with LM Studio, expressing concerns about its format and memory requirements.
  • Gemma 2’s erratic behavior: Users reported issues with the Gemma 2 model, including repeating symbols and failing at various context lengths. Even with different settings, users noted it sometimes does not function as expected.
  • RPing to stress test models: A user shared insights on stress testing models through role-playing (RP), emphasizing it as a method to reveal a model’s faults. They proposed measuring models by the percentage of edited output needed during an RP session.
  • Gemma 2 praised for introspectiveness: Despite some issues, users praised Gemma 2 for its introspective abilities and strong context understanding. It was compared favorably to other models in terms of hidden context detection and correctness.
  • Code generation improvements sought: Users discussed the challenge of models inserting placeholders in code output, even with explicit instructions not to do so. Suggestions included using detailed system prompts to guide the model’s code generation more effectively.

LM Studio ā–· #announcements (1 messages):

  • LM Studio 0.2.27 Release
  • Improved Gemma 2 Support
  • Bug Fixes in lmstudio.js
  • Advanced Information on lmstudio.js
  • LM Studio 0.2.27 Launch Celebrates Enhanced Gemma 2 Support!: LM Studio 0.2.27 is now available for Mac (M1/M2/M3), Windows (x86 and ARM64), and Linux (x86). Users can download it or restart their app to trigger the auto-update.
  • Gemma 2 Models Revamped: Performance improvements for Gemma 9B and Gemma 27B models are thanks to contributions from abetlen, ngxson, slaren, ggerganov and others. Download the updated models from the Hugging Face community page: Gemma 9B, Gemma 27B.
  • Bug Fixes in lmstudio.js: [lmstudio.js] team has fixed the ā€œinvalid creation parameterā€ bug (issue #45). Additional updates addressed messages about no GPU support.
  • Advanced Information for Power Users: The latest llama.cpp commit ID is d08c20eddedb24515a3212e2de66bdff41a26b8c and OpenCL backend is now bundled again for both Windows and Linux platforms. However, Gemma 2 is not supported for OpenCL.
  • Updating AMD ROCm Extension Pack on Windows: Windows users with AMD ROCm can follow the instructions to update their ROCm extension pack. The Linux ROCm extension pack update is still under development.

Links mentioned:


LM Studio ā–· #🧠-feedback (11 messagesšŸ”„):

  • Basic functionality issue with drive installation
  • User confusion on model insertion
  • Scaling issues on 1080p monitors
  • Unsupported architecture message in new models
  • User feedback on improving LM Studio interfaces
  • Drive Installation Gripes: Members expressed frustration that LM Studio cannot be installed on a chosen drive due to Squirrel limitations. Instead, users must change the ā€˜My Models’ folder to handle storage issues.
  • Model Insertion Confusion: A user struggled with inserting a model, citing error messages about model operation failures despite having adequate system resources, such as 15.90 GB RAM and an NVIDIA GeForce GTX 950M.
  • Scaling Issue on 1080p Monitors: A user mentioned that LM Studio does not scale well on a 1/4 section of a 1080p monitor, leading to missing settings buttons and poor layout. This impacts multi-monitor workflows.
  • Unsupported Architecture Alerts: Members observed that new models often trigger ā€˜unsupported arch’ alerts, similar to issues with DeepSeek Coder v2, due to their configuration files.
  • Call for More Metadata: A user suggested adding the publication date and additional metadata for models on LM Studio’s homepage to enhance usability. This was well-received as a potential improvement.

LM Studio ā–· #šŸ“-prompts-discussion-chat (3 messages):

  • Prompting Llama 3 70B to remove conversational lines
  • Issue with prompt results on Llama 3 compared to Qwen2 72B
  • Creation of a prompt tool with Gradio app for role-play and character immersion
  • Prompt Llama 3 70B without annoying conversational lines: A user inquired about prompting Llama 3 70B to skip the trite conversational lines at the beginning of its responses. Example phrases like ā€˜What a wonderful thing it is to have a drink!’ were highlighted as unnecessary.
  • Prompt success on Qwen2 72B vs Llama 3: One user shared their success in removing conversational lines from Qwen2 72B, contrasting with their struggles to achieve the same results on Llama 3.
    • They expressed frustration, noting the challenge in getting similar prompt performance from Llama 3 despite applying the same techniques.
  • New Gradio app for role-play prompts: A user introduced a new Gradio app designed to create role-play prompts for immersive character experiences. The app includes dynamic variables to define character roles and scenarios.
    • They shared a sample prompt and invited feedback for improvement, providing a link to the app.

Link mentioned: System Roleplay Generator - a Hugging Face Space by xtreme86: no description found


LM Studio ā–· #šŸŽ›-hardware-discussion (3 messages):

  • Energy consumption of LMS
  • Hardware issues on Linux vs Windows
  • GPU usage comparison of LMS with other software
  • User experiences with different GPU setups
  • Potential future fixes for LMS energy consumption
  • Energy consumption in LMS at idle: A member reported unreasonable energy consumption in LMS at idle, noting that the power usage should be half per GPU at about 10W. They highlighted Blender as a better comparison than web browsing for GPU usage.
  • Windows vs Linux: Differing power consumptions: A member experienced no significant hardware issues with LMS on Linux but acknowledged possible differences on Windows. They observed that running LMS with a model loaded only made a 33-watt difference compared to running Firefox or Chrome, with an additional 22-watt difference when no model was loaded.

LM Studio ā–· #amd-rocm-tech-preview (14 messagesšŸ”„):

  • Gemma 2 loading issue
  • ROCM GPU compatibility and performance
  • Linux ROCm extension pack testing
  • Gemma 2 fails to load in LM Studio 0.2.27: A user encountered an error loading Gemma 2 with message: ā€˜unknown model architecture: ā€˜gemma2ā€ despite clearing the cache and running necessary scripts. Suggested fix included a re-download or clean install.
  • ROCM support success on AMD GPUs: Discussion revealed successful use of ROCM on AMD Radeon RX 6900 XT and 7800 XT GPUs, with testimony of running Gemma 2 at 8k tokens without RAM issues. Another user confirmed that ROCM builds work fine with these models.
  • Call for Linux ROCm extension testing: A community member called for assistance in testing the latest Linux ROCm extension pack for version 0.2.27. Instructions provided included installing via script and confirming ROCm llama.cpp appears under settings.

LM Studio ā–· #model-announcements (1 messages):

  • Gemma 2 model update on Huggingface
  • Compatibility updates for Gemma 2 models
  • Gemma 2 Models Updated for Compatibility: The Gemma 2 models on the lmStudio community Huggingface have been updated to the latest changes and are safe to redownload and use with version 0.2.27.
  • Gemma 2 Models Ready for Use: The updated Gemma 2 models are now compatible with the latest version 0.2.27 and can be downloaded from Huggingface.

LM Studio ā–· #šŸ› -dev-chat (70 messagesšŸ”„šŸ”„):

  • gpuOffload value discussion
  • bot configuration issues with TypeScript and Discord.js
  • ****Fixing discord token issue: New bot fails due to invalid token: Aquora initially struggles with an invalid token error while configuring a discord bot. The issue is eventually traced back to disallowed MessageContent intents, which is resolved by enabling them in the Discord Developer Portal.
  • Bot loops & ā€˜thinking’ state fix: Temperature and predicted tokens adjustment solves bot hallucinations: Aquora experiences issues with the bot
    • DJ suggests adding message history and direct message handling in a future article to improve bot functionality. Aquora is eager to contribute to further improvements.

Links mentioned:


HuggingFace ā–· #announcements (1 messages):

  • New fine-tunes for Transformers models with KerasNLP
  • Experimental API for searching HF datasets by column names
  • Transformers 4.42 release with new features and models
  • Nearly 100k public models on HF Hub storing tensorboard logs
  • Local Gemma release
  • Transformers 4.42 release introduces new models and features: The new Transformers 4.42 release includes Gemma 2, RT-DETR, InstructBlip, LLaVa-NeXT-Video, tool usage and RAG support, GGUF fine-tuning, and quantized KV cache.
  • KerasNLP bridges fine-tuning for any Transformers model: Access tons of new fine-tunes for any Transformers model using a KerasNLP implementation.
  • AWS releases Chronos datasets on HF: AWS released all datasets used in the Chronos paper on Hugging Face, including both pretraining and evaluation datasets.
  • Local Gemma offers 100% private and secure generation: The new Local Gemma is 100% local, private and secure, and can run anytime with pip install local-gemma.
  • Vision language models introduction released: Intro to vision language models has been released showcasing image-text-to-text models.
    • This includes tasks like image captioning, optical character recognition, and more.

Links mentioned:

  • Tweet from Julien Chaumond (@julien_c): Keras šŸ¤ HF
  • Tweet from Daniel van Strien (@vanstriendaniel): Search @huggingface datasets by column names with a new experimental API! This API allows you to: - Search for question-answering datasets that include context - Find alpaca-style datasets - Locate D...
  • Tweet from Omar Sanseviero (@osanseviero): Transformers 4.42 is out, and it has lots of amazing features🄳 šŸ”„New models: Gemma 2, RT-DETR (obj detection), InstructBlip, and LLaVa-NeXT-Video šŸ”§Tool usage and RAG support šŸ‘€GGUF fine-tuning šŸ¤Qu...
  • Tweet from Wauplin (@Wauplin): Almost 100k public models uses the Hub to store ššššŽšš—ššœšš˜šš›šš‹šš˜ššŠšš›šš logs! Storing training logs alongside checkpoints let you keep track of everything in a single place using the Metrics tab' ļæ½...
  • Tweet from Vaibhav (VB) Srivastav (@reach_vb): Introducing Local Gemma! šŸ’Ž 100% local, private and secure - run anywhere, anytime! cuda, mps, cpu - with presets to go as fast as possible - built on the shoulders of transformers, we ensure 1:1 ge...
  • Tweet from Abdul Fatir (@solitarypenman): šŸš€šŸš€šŸš€ We just released all the datasets used in the Chronos paper on Hugging Face. This includes both pretraining and evaluation (in-domain and zero-shot) datasets. We also open-sourced a script to...
  • Tweet from merve (@mervenoyann): Real-time DEtection Transformer (RT-DETR) landed in @huggingface transformers 🤩 with Apache 2.0 license šŸ˜ do DETRs Beat YOLOs on Real-time Object Detection? keep reading šŸ‘€
  • Tweet from Xenova (@xenovacom): Florence-2, the new vision foundation model by Microsoft, can now run 100% locally in your browser on WebGPU, thanks to Transformers.js! šŸ¤—šŸ¤Æ It supports tasks like image captioning, optical characte...
  • Tweet from merve (@mervenoyann): Just shipped: intro to vision language models (aka image-text-to-text)
  • Tweet from Ben Burtenshaw (@ben_burtenshaw): šŸš€ Excited to launch our new series, Data Explorer by @argilla_io ! šŸŽ„ We dive deep into datasets and their impact on model performance. Our first episode explores the PRISM dataset by @hannahrosekir...
  • Tweet from Zach Mueller (@TheZachMueller): How do you make @PyTorch dataloaders work efficiently during distributed training? Here's a video tutorial I did with @huggingface accelerate's dataloaders showing how we do so https://www.yo...
  • Tweet from merve (@mervenoyann): New RAG with Gemma recipe using @elastic search, @huggingface šŸ§‘šŸ»ā€šŸ³šŸ“– Find it below ⇓

HuggingFace ā–· #general (236 messagesšŸ”„šŸ”„):

  • Joining Hugging Face Discord Community
  • Adept Strategy Shift & Co-Founders Joining Amazon
  • AI Models for Text and Image Processing
  • Performance and Accuracy of Hugging Face Models
  • Suggestions for ML Certifications
  • Invite to Join HuggingFace Discord: <a Discord Community link was shared, clarifying past verified members will receive invitations soon.: A user inquired about joining the HuggingFace Discord Community; invitations will be sent to all past verified members soon, emphasizing a shared collaborative space for live projects.
    • Another user expressed interest in moderating the community due to their engagement in reporting scams.
  • Promising AI Models for Text Processing: Users discussed various AI models like Qwen with 132k context windows. Accurate models like Qwen2 are highlighted for their impressive context lengths and usefulness in detailed text processing.: HuggingFace’s user community discussed AI models capable of handling extensive context windows, recommending Qwen2 with a 132k context window for detailed text processing tasks.
    • Experimentation with long contexts versus more concise model outputs was suggested, considering performance and quality in various contexts.
  • Comparing Open Source and Proprietary AI Models: Users compare HuggingFace models like Meta-Llama to OpenAI’s GPT. A Meta-Llama 3 70B model was recommended for superior performance in benchmarks.**: A comparison was made between HuggingFace models and proprietary models like OpenAI’s GPT, with the Meta-Llama 3 70B model being recommended for benchmark superiority.
    • Users noted the balancing act between speed and reliability with various models, indicating some use cases still prefer proprietary tools for their efficiency.
  • Learning Resources and Certifications in ML: Suggestions were made for free and efficient online courses for proving ML proficiency. Harvard and Coursera courses were recommended for their comprehensive content and certification credibility.**: Several users shared their experiences with free courses from Harvard and Coursera, noting their balance of quality and certification credibility, beneficial for those looking to prove ML proficiency.
    • One user asked about skipping repetitive basics in these courses, highlighting a preference for progressive learning models.

Links mentioned:


HuggingFace ā–· #today-im-learning (7 messages):

  • advanced resources for CNN topics like ViT and Unets
  • request for more tutorials on torch.distributed
  • Gradio app for role-play and character immersion prompt creation
  • TIL about '|' and '&' operators for Sets and Dicts in Python
  • question about Bayes' theorem in French
  • Role-Play Prompt Tool Takes Center Stage: A member shared a Gradio app they created for role-play and character immersion prompt creation, seeking feedback and improvement tips. They showcased some generated results and provided a link to the tool.
  • Python Set Operators Unveiled: A member shared a new GitHub resource about the | and & operators for Sets and Dicts in Python, stating it is not directly related to ML but still interesting.

Links mentioned:


HuggingFace ā–· #cool-finds (8 messagesšŸ”„):

  • attention mechanism
  • transformer architecture
  • compatibility in shells
  • sequence transduction models
  • demo video reactions
  • Attention in transformers explained visually: A user shared a YouTube video titled ā€˜Attention in transformers, visually explained | Chapter 6, Deep Learning,’ detailing the key mechanisms inside transformers and LLMs. The video is praised as the ā€˜best video about transformer architecture’ seen so far.
  • Compatibility First with Starship.rs: A link to Starship.rs was shared, emphasizing the shell’s compatibility with the most common operating systems. This tool promises usability across various environments.
  • The Transformer Paper: A user highlighted the Transformer architecture paper from arXiv, proposing a network based solely on attention mechanisms. It achieved a BLEU score of 28.4 for English-to-German and 41.8 for English-to-French translations on the WMT 2014 tasks.

Links mentioned:

  • Attention Is All You Need: The dominant sequence transduction models are based on complex recurrent or convolutional neural networks in an encoder-decoder configuration. The best performing models also connect the encoder and d...
  • Attention in transformers, visually explained | Chapter 6, Deep Learning: Demystifying attention, the key mechanism inside transformers and LLMs.Instead of sponsored ad reads, these lessons are funded directly by viewers: https://3...
  • Starship: Cross-Shell Prompt: Starship is the minimal, blazing fast, and extremely customizable prompt for any shell! Shows the information you need, while staying sleek and minimal. Quick installation available for Bash, Fish, ZS...

HuggingFace ā–· #i-made-this (7 messages):

  • OpenAI's CriticGPT Release
  • Stable Release of Embodied Agents Toolkit
  • Open Source OCR for Kazakh Language
  • Blog on Reinforcement Learning Specialization
  • Zero-Shot Generating Spatial Sound from Images
  • OpenAI Unveils CriticGPT: A user shared a YouTube video introducing CriticGPT, a new AI model by OpenAI that identifies errors in code generated by GPT-4. The release was lauded as a significant step towards improving code accuracy.
  • Embodied Agents Toolkit for Robotic Integration: Embodied Agents toolkit was recently released to integrate state-of-the-art multimodal transformers into robotics with minimal code. The toolkit includes Gradio interface support and HuggingFace dataset integration.
  • OCR Solution for Kazakh Language Released: An open source solution for OCR in Kazakh language has been released. The solution aims to fill a significant gap in OCR technologies for underrepresented languages.
  • Reinforcement Learning Specialization Blog: A user shared their blog series on Reinforcement Learning Specialization from Coursera. The detailed notes cover multiple weeks and courses regarding RL.
    • Check out the details and give feedback, as suggested by the author.
  • Vietnamese Vision-Language Model Launched: The Vi-VLM team launched a Vietnamese Vision-Language model based on LLaVA and Vistral LLM. The model, optimized for image description tasks, leverages a proprietary dataset for pretraining and supervised finetuning.

Links mentioned:


HuggingFace ā–· #reading-group (4 messages):

  • Highway Net vs ResNet performance
  • Gradient vanishing problem in LSTM
  • Multi-branch structure inspiration from LSTM
  • Pre-trained models and fine-tuning techniques
  • topicSummaries
  • Highway Net performs worse than ResNet: Time to rethink: A user questioned why Highway Net performs worse than ResNet, suggesting it might be time to reconsider the design choices. Does the gating scheme from LSTM really solve the gradient vanishing problem?
  • LSTM-inspired multi-branch structure: A user admitted that their idea of multi-branch structure comes from LSTM. This raises questions about the gradient vanishing problem in LSTM’s gating scheme.
  • Fine-tuning pre-trained models with cost-effective methods: A user shared a paper discussing techniques to fine-tune pre-trained models without updating all parameters, focusing on resource-efficient methods like low-rank adjustments.
    • These methods often neglect higher-dimensional parameter spaces like 4D, causing structural integrity issues.

Link mentioned: FLoRA: Low-Rank Core Space for N-dimension: Adapting pre-trained foundation models for various downstream tasks has been prevalent in artificial intelligence. Due to the vast number of tasks and high costs, adjusting all parameters becomes unfe…


HuggingFace ā–· #computer-vision (6 messages):

  • ADVANCED_CNN_RESOURCES
  • Neighborhood_Attention_Transformer_usage
  • Developer_Job_Openings
  • MaskFormer_training_issues
  • Lightweight_AI_for_programming
  • Books on Advanced CNN Techniques Needed: A user is seeking recommendations for books or resources on advanced CNN topics like ViTs and UNets, including video processing.
  • Neighborhood Attention Transformer Maintenance Issue: One user shared a link to the Neighborhood Attention Transformer documentation, highlighting that the model is in maintenance mode only and suggesting to re-install version 4.40.2 if issues occur with newer versions. They referenced the paper titled Neighborhood Attention Transformer for more context.
  • Developer Seeking Jobs: A user inquired if there are any companies or projects currently seeking a skilled developer.
  • Issues Training MaskFormer Model: A user is having trouble with training a MaskFormer model for instance segmentation, struggling with mask accuracy and training times. They are using the Hugging Face Trainer class and asked for personal assistance.
  • Looking for Lightweight AI for Programming: A user asked for recommendations on an AI model suitable for programming that is also lightweight.

Link mentioned: Neighborhood Attention Transformer: no description found


HuggingFace ā–· #NLP (9 messagesšŸ”„):

  • Custom pipeline creation
  • Text summarization models with high max input token length
  • Performance of open-source models vs. ChatGPT
  • Challenges in downloading and using Meta LLaMA
  • Inference freezing issues with Mistral model
  • Creating Custom Pipeline Guide by Andy Singal: Andy Singal shared a guide on creating custom pipelines using Transformers. The guide is part of the LLM course available on GitHub.
    • Exploring custom pipelines could be beneficial for those wanting tailored solutions in their NLP tasks.
  • Need for Text Summarization Model for Long Documents: A user asked for recommendations on text summarization models that handle extremely long documents, specifically those with high max input token length. This request highlights the challenge of summarizing lengthy texts effectively.
  • Open-source Models vs. ChatGPT: Discussion on whether Hugging Face models can match the performance/accuracy of ChatGPT 3.5 or 4. A member mentioned open models claiming better performance than ChatGPT 3.5.
    • Models tend to overfit on benchmarks, as emphasized in the discussion.
  • Meta LLaMA Download Challenges: A user experienced difficulties downloading Meta LLaMA and considered building an API call to the model. They expressed concern about potential failures due to temporary file storage limitations during the 20-minute download process.
  • Inference Freezing Issue in Mistral: Running an experiment with Mistral model froze at iteration 1800 when running 3000 inferences, taking a day to proceed. Inference choking might be due to some caching or other resource management issues between runs.

Link mentioned: llm-course/transformers/custom-pipeline.md at main Ā· andysingal/llm-course: Contribute to andysingal/llm-course development by creating an account on GitHub.


HuggingFace ā–· #diffusion-discussions (1 messages):

  • Discussion on running RealVisXL V4.0 Lightning model with diffusers.
  • Comparison of quality between A1111 and diffusers.
  • Support on Boosty.
  • Recommended negative prompt and generation parameters.
  • Issues with model performance during training phase.
  • Support RealVisXL V4.0 on Boosty: A member shared that you can support the development of RealVisXL V4.0 Lightning on Boosty.
  • RealVisXL V4.0 model training: The RealVisXL V4.0 Lightning model, aimed at achieving photorealism, is still in the training phase and may contain artifacts and perform poorly in some cases.
  • Running RealVisXL V4.0 with diffusers: A member reported that running RealVisXL V4.0 Lightning with diffusers results in much poorer quality compared to using A1111 despite using the same parameters (prompt, step, scheduler, etc.).

Link mentioned: SG161222/RealVisXL_V4.0_Lightning Ā· Hugging Face: no description found


Eleuther ā–· #general (93 messagesšŸ”„šŸ”„):

  • GPT-4 parameter discussion
  • Nvidia involvement in GPT-4 development and leaks
  • Mixture of Experts (MoE) models
  • InstructGPT efficiency
  • Discord server scraping and ToS violations
  • GPT-4 parameters **: How Big is It Really?: Discussion emerged around GPT-4’s parameter count, with numbers suspected to be around 1.7 trillion to 1.8 trillion according to various sources, including Nvidia. Interestingly, this figure shows a monumental leap from GPT-3’s 175 billion parameters, leaving members curious how MoE (Mixture of Experts) contributed to this scaling.
  • InstructGPT’s Real-World Boost: InstructGPT gains were highlighted, particularly in practical applications where it offers a 10X to 100X increase in efficiency. The community underscored the significant impact of RLHF (Reinforcement Learning from Human Feedback) as a key driver behind this improvement.
  • Nvidia in the Know: GPT-4 Secrets: Nvidia’s familiarity with the GPT-4 model size stirred debate, given their hardware support contracts. Despite being bound by NDAs, many believe Nvidia has insights into OpenAI’s models due to their integral role in hardware provisions.
  • Scraping Discord: Risky Business: Attempting to scrape Discord servers for data, even personal ones, violates Discord’s ToS and could lead to bans. Some tools like DiscordChatExporter avoid rate limiting but pose significant risks, as highlighted by several members.
  • MoE Model Efficiency Insights: Technical deep dives into MoE models revealed constraints and efficiency boosts. While MoE can significantly reduce computational loads by activating selective weights, it still faces memory bandwidth and VRAM challenges during inference.

Link mentioned: AI News: We summarize top AI discords + AI reddits + AI X/Twitters, and send you a roundup each day! See archive for examples. ā€œHighest-leverage 45 mins I spend everydayā€ - Soumith ā€œbest AI new…


Eleuther ā–· #research (76 messagesšŸ”„šŸ”„):

  • UL2 vs traditional training objectives
  • Starcoder2 and UL2 performance
  • PrefixLM and training implications
  • Scaling laws and learning rate schedules
  • FIM and UL2 comparisons
  • Industry slow to adopt UL2 training objectives: A member expressed surprise at the industry’s slow adoption of UL2 training objectives despite their theoretical and empirical benefits in fixing issues like short-term planning and the reversal curse. Despite tests from Starcoder2 and Mosaic showing underperformance compared to traditional methods, the member remains optimistic about future tweaks.
  • Scaling Laws Discrepancies: Researchers resolved discrepancies between Kaplan and Hoffmann scaling laws by identifying factors like last layer computational cost and optimizer tuning. Their findings debunked the necessity of careful learning rate decay for the validity of the Chinchilla scaling law.
  • PrefixLM and training efficiency concerns: Members debated the efficiency of PrefixLM in training, noting its slower training speeds and potential inefficiency with current attention algorithms. One member pointed out that models might adapt differently to bidirectional contexts versus causal ones, impacting performance.
  • FIM vs UL2 Objectives: Members discussed Fill-in-the-Middle (FIM) and how it compares to UL2/masked language objectives. They noted that FIM could be more efficient as it also predicts front and tail segments, potentially offering better outcomes than the UL2 approach.
  • Attention Mask Adaptation: A discussion emerged on the adaptation of models to different attention masks, with some noting the challenges in training efficiency and the importance of specific tokens. A related paper on causal to bidirectional attention adaptation was shared to illustrate these dynamics.

Links mentioned:


Eleuther ā–· #scaling-laws (22 messagesšŸ”„):

  • Discrepancy in compute optimal scaling laws between Kaplan et al. and Hoffmann et al.
  • Kaplan et al.'s last layer computational cost, warmup duration, and scale-dependent optimizer tuning
  • Attention flops and the 6ND approximation in scaling law computations
  • PyTorch flop counter utility and FLOPs calculation methodologies
  • Chinchilla paper scaling law and extrapolation issues
  • Resolving Scaling Law Discrepancies Between Kaplan and Hoffmann: Researchers explain the discrepancy between Kaplan et al. and Hoffmann et al.’s scaling laws by identifying issues such as last layer computational cost, warmup duration, and scale-dependent optimizer tuning.: Researchers corrected these factors and obtained excellent agreement with Hoffmann et al.’s scaling law, also known as the Chinchilla scaling law. They found the learning rate decay hypothesis by Hoffmann et al. to be non-essential and derived scaling laws for optimal learning rate and batch size.
  • Attention FLOPs Impact on Scaling Laws: Community members discuss the inadequacy of the 6ND approximation for small scale models.: They suggest incorporating attention flops using a different formula from Kaplan et al., specifically C = 6ND + 6 * n_layers * seq_len * d_model instead of 6ND.
  • Utility Flaws in PyTorch’s FLOP Counter: Discussion on PyTorch’s built-in flop counter utility.: Concerns were raised about the utility not hitting an error if it encounters an operation it doesn’t know the FLOPs for, defaulting to ignoring it.
  • Extrapolation Pitfalls in Scaling Law Fits: Community emphasizes the risks of using too small-scale experiments to fit scaling laws.: They highlight that chin*chilla numbers should not be cargo-culted** and emphasize the need for spending substantial compute to ensure accurate extrapolation.
  • Chinchilla Work on Rubik’s Cube Data: Reference to a GitHub project focusing on synthetic data scaling law fits.: A project on GitHub demonstrates these principles using Rubik’s cube generated data to validate scaling laws.

Links mentioned:


Eleuther ā–· #interpretability-general (2 messages):

  • EAP with integrated gradients
  • Methods for discovering and applying sparse feature circuits
  • Generalization improvement using SHIFT
  • Scalable interpretability pipeline for sparse feature circuits
  • Weekend Updates: A community member mentioned an event happening this weekend, but no specific details were provided.
  • EAP and Integrated Gradients Insight: Discussion on EAP with integrated gradients referencing the paper Methods for discovering and applying sparse feature circuits.
  • Sparse Feature Circuits Paper Explored: The paper introduces methods for discovering and applying sparse feature circuits that enable detailed understanding of language model behaviors. They are useful for downstream tasks and offer an unsupervised, scalable interpretability pipeline.
  • SHIFT for Classifier Generalization: The paper discusses the SHIFT method that improves classifier generalization by ablating task-irrelevant features as judged by humans. This method leverages fine-grained units for better interpretability.

Link mentioned: Sparse Feature Circuits: Discovering and Editing Interpretable Causal Graphs in Language Models: We introduce methods for discovering and applying sparse feature circuits. These are causally implicated subnetworks of human-interpretable features for explaining language model behaviors. Circuits i…


Eleuther ā–· #lm-thunderdome (26 messagesšŸ”„):

  • PR confirmation and lm-eval reference
  • Loglikelihood_rolling functionality and usage
  • Handling document length longer than model's context in perplexity evaluations
  • Errors in model evaluation with specific configurations
  • Preprocessing functions and pipeline consistency
  • PR confirmation for lm-eval reference: A member requested confirmation for a pull request because they want to reference it in their paper for running evaluations of their benchmark. Stellaathena inquired if the numbers in the paper were from the current code, and the member confirmed.
  • Understanding loglikelihood_rolling: A member asked about the purpose of loglikelihood_rolling and whether it means inputting a model to get a loglikelihood value that can be turned into a loss value. Hailey Schoelkopf explained that it gives the loglikelihood of producing a document from the empty string with reference to the documentation.
  • Handling long documents in perplexity evaluations: Stellaathena asked how to compute perplexity for documents longer than the model’s context window without errors. Hailey Schoelkopf noted that default perplexity tasks handle chunking based on model length automatically within the loglikelihood_rolling method.
  • Dataset-specific issues in evaluation configuration: Using a specific config file with proof-pile dataset throws an error for Stellaathena, but the same config works with lambada_openai. Hailey Schoelkopf mentioned it could be related to the metrics, suggesting a fix and indicating a potential silent failure with metric use.
  • Reusing preprocessing functions: A member inquired about preventing preprocessing functions from rerunning every time and if preprocessed data can be stored and reused in the pipeline. This concern aligns with ensuring efficiency in the evaluation process.

Links mentioned:


Perplexity AI ā–· #general (174 messagesšŸ”„šŸ”„):

  • Discussion About Trying Gemini 1.5 Pro
  • Access Issues with GPT4o
  • New Perplexity Features on Mobile
  • Refund Process for Pro Subscription
  • Concerns About Perplexity's Live Internet Access
  • Gemini 1.5 Pro chatbot recommendation: Members discussed the performance and features of Gemini 1.5 Pro, noting its large context window and fast performance. One user particularly recommended trying it due to its decent capabilities.
  • Access Issues with GPT4o and Alternatives: Several users reported issues finding free ChatGPT 4o options, suggesting alternatives like Bing chat in precise mode and Claude 3.5 Sonnet on claude.ai, which is also praised for its free use despite some usage limits.
  • New Perplexity features on mobile: A user inquired about the availability of the new Wolfram Alpha and code generation features in Perplexity’s mobile app. Another user confirmed that these features are available on iOS.
  • Refund process for Pro subscription: A user asked about the refund process for Perplexity Pro subscription. Another member provided detailed refund policies, specifying guidelines for EU, UK, Turkey, and all other customers including timelines and conditions.
  • Concerns about Perplexity’s live internet access: A member reported erratic behavior in Perplexity’s ability to access live internet information, especially for checking real-time data like sports scores and weather. Despite correcting itself at times, it often denied such capabilities, leading to consistent frustration.

Link mentioned: gateway/cookbook/integrations/Phidata_with_ Perplexity.ipynb at main Ā· Portkey-AI/gateway: A Blazing Fast AI Gateway. Route to 200+ LLMs with 1 fast & friendly API. - Portkey-AI/gateway


Perplexity AI ā–· #sharing (9 messagesšŸ”„):

  • Lean Canvas Guide
  • Starting Perplexity AI Story
  • Building a Blackbox
  • OpenSSH Query
  • Sober Living in Echo Park
  • Guide to Lean Canvas: Explore Lean Canvas solutions with a streamlined guide on Perplexity AI. Contains valuable insights and step-by-step instructions.
  • Perplexity AI Founding Story: Dive into the origins of Perplexity AI with this comprehensive narrative. The article shares the inspiration and challenges faced during the creation process.
  • Building a Blackbox: Learn about constructing a blackbox system in the AI domain through a detailed Perplexity AI search result. Discusses methodologies and potential applications.

Perplexity AI ā–· #pplx-api (7 messages):

  • Usage of Sonnet 3.5 with Perplexity API
  • Availability of Sonnet in Perplexity API
  • List of available models in Perplexity API
  • Search engine usage via Perplexity API
  • Issues with llama-3-sonar-large-32k-online model
  • Sonnet 3.5 not available via Perplexity API: Sonnet is not provided via the Perplexity API. The available models can be found in the documentation.
  • Interest in API for search engine usage: Multiple members expressed interest in using Perplexity’s search engine via the API. One mentioned emailing [email protected] for beta access.
  • Incorrect answers from llama-3-sonar-large-32k-online model: A member noted that llama-3-sonar-large-32k-online gave wrong answers for a simple query. Another suggested using the after: parameter to refine the search.

Link mentioned: Supported Models: no description found


CUDA MODE ā–· #general (19 messagesšŸ”„):

  • CUDA-only hackathon at the AGI House in San Francisco
  • Meta Hacker Cup 2024 schedule
  • Discussion about the price and purchase of NVIDIA GPUs (3090, 4090)
  • CUDA Hackathon Opens in San Francisco: Ash Vardanian is hosting a CUDA-only hackathon at the AGI House on July 13th, with Chris Lattner as a notable speaker. H100 access will be provided to all participants, sponsored by Nebius.ai.
  • Meta Hacker Cup Returns for 2024 Season: The Meta Hacker Cup will kick off on September 20th with its Practice Round, followed by a series of rounds culminating in the Finals on December 7th. Mark Saroufim, part of the organizing team, encourages participation, especially for those interested in code generation.
  • Debate Over NVIDIA 3090 Prices: Members discussed whether to purchase a 3090, noting that current prices are around $1,000. Mark Saroufim mentioned he bought a 4090 for $1,200 during a Meta layoffs scare.

Links mentioned:


CUDA MODE ā–· #torch (14 messagesšŸ”„):

  • Steps involved in compiling a function in Pytorch with Inductor backend for Nvidia device
  • Difference between triton IR and MLIR
  • John Carmack's positive feedback on PyTorch team and contributing to open source
  • Issue with forcing Inductor to generate Triton kernels for GEMM and Conv
  • Steps for Compiling Pytorch Function with Inductor: Discussion on the steps for compiling a function in Pytorch with Inductor backend for Nvidia, from PYTorch (python method) to PTX. Confusion about whether MLIR should be a separate step. (source)
  • MLIR is not a Separate IR: Clarification that MLIR is a toolkit for building your own IR, not a separate IR. Triton uses ttir, ttgir, llir, ptx, and cubin as steps but ā€˜the ttir => ttgir translation is the most important’.
  • John Carmack Praises PyTorch Team: John Carmack is impressed by the @PyTorch team’s response to bug reports, stating that while the project has a substantial learning curve, the setup documentation made him self-sufficient.
  • Inductor Triton Kernels Issue: Forcing Inductor to generate Triton kernels for everything works for GEMM but not for Conv, despite having kernel templates. Issue raised on GitHub, looking for a fix.

Links mentioned:


  • High-performance matrix multiplication on CPU
  • 3D V-Cache performance on AMD Ryzen
  • Difference between 3D and non-3D Ryzen chips
  • Discussion on specialization of 3D V-Cache chips
  • Simulation benchmarks for CPUs
  • High-performance matrix multiplication on CPU: Mobicham shared a tutorial on high-performance matrix multiplication on CPU with code available at matmul.c. The implementation, optimized for AMD Ryzen 7700, outperforms NumPy by achieving over 1 TFLOPS using 3 lines of OpenMP directives.
  • 3D V-Cache boosts AMD’s performance: Iron_bound wondered about the performance impact of 3D V-Cache, citing a review showing it has 96MB of L3 cache and comparing it across games and simulations.
  • 3D vs non-3D Ryzen chip differences: As_ai and iron_bound discussed the differences between 3D and non-3D Ryzen chips, noting that 3D versions have double L3 cache but operate at lower clocks to prevent damage to the extra cache silicon layer.
  • Specializations in 3D V-Cache chips: As_ai questioned if the 3D V-Cache had any specializations beyond the extra cache, and iron_bound confirmed the differences are primarily more cache and lower clocks.

Links mentioned:


CUDA MODE ā–· #torchao (31 messagesšŸ”„):

  • Loading a buffer containing int4 using torchao
  • Saving a tensor into a safetensors file
  • Dequantizing tensors using torchao
  • Handling packed int4 arrays in Python
  • torchao's handling of unexpected keyword arguments
  • Handling packed int4 arrays in Python efficiently: Members discussed interpreting packed int4 arrays by first converting to uint8 and then bit-shifting and stacking the tensor using torch.frombuffer(). They emphasized understanding the bit-layout of the buffer before parsing it.
  • Dequantizing tensors using Python techniques: A member asked about dequantizing tensors with quantization scales, which involved creating tensors and performing element-wise operations like dequant_tensor = quant_tensor * scale using PyTorch.
  • torchao’s buffer loading and unexpected keyword handling: Discussion centered around torchao’s function to() and how it interprets arguments, revealing issues with unexpected keyword arguments being passed to __new__(). They noted the importance of configuring parameters correctly to avoid errors during tensor operations.

Link mentioned: Gemini-Nano/playground/converter.py at master Ā· ethanc8/Gemini-Nano: Contribute to ethanc8/Gemini-Nano development by creating an account on GitHub.


CUDA MODE ā–· #llmdotc (84 messagesšŸ”„šŸ”„):

  • Memory efficiency comparison with PyTorch
  • Visualization of model weights and training issues
  • New GitHub PRs and bug fixes
  • Experiments with and observations on muP
  • Schedule-free optimization discussion
  • Memory Efficiency: Our models significantly outperform PyTorch: Our batch 16 run fits very comfortably, even batch 24 fits fine, while PyTorch struggles with batch size 8. This highlights significant memory savings compared to PyTorch.
  • Visualization and Training Bugs: Issues observed with integer division by zero error during training on batch sizes < 4 in the HellaSwag eval dataloader. Fix implemented in GitHub PR #667.
  • muP Experiments Look Promising Yet Challenging: Preliminary muP results are stable with various learning rates, with extensive experiments filling up to 80 GB of VRAM per GPU. Further hyperparameter sweeps are being planned.
  • Efficient Inference with HF Transformers: Sample generation from Hugging Face GPT2 models is extremely slow, taking 4 minutes for 512 steps due to inefficient eager mode and dynamic key-value concatenations. Active effort noted to improve Transformers documentation.
  • Excitement Over Schedule-Free Optimizers: Schedule-free optimizers by Facebook Research show surreal convergence on various tasks. Claims suggest it may be a breakthrough in practical and theoretical optimization research.

Links mentioned:


Stability.ai (Stable Diffusion) ā–· #general-chat (145 messagesšŸ”„šŸ”„):

  • Anti-AI art software debate
  • Tips for low-resolution pixel art training
  • Job postings on Discord
  • Improving prompt techniques and comparisons
  • Comparisons between SD models *MixofExperts* and segmoe
  • Anti-AI Art Software Discussed: Members discussed the feasibility of anti-AI art software that could protect artists’ work from being used in AI training. Suggestions included existing tools like Glaze and Nightshade, but community members pointed out that these methods are easily defeated.
  • Training Low-Resolution Pixel Art Models: A user inquired about training AI for 16x16 pixel art, and members recommended upscaling images to 512x512 for training. Crystalwizard noted the potential inefficiency but suggested trial and error as a cost-effective method.
  • Job Postings and Freelance Work: A user asked if there was a job-posting channel for recruiting, and another inquired about upwork account rentals, highlighting the demand for freelance opportunities.
  • Effectiveness of Prompt Techniques: Members discussed different prompting techniques and their effectiveness in generating images using text2img. Variations like [A|B], C vs. [A, B, C] were mentioned, alongside a comparison of model capabilities like SD1.5 vs. segmoe and MixofExperts.
  • Comparing SD Models: MixofExperts vs segmoe: Discussions included the segmoe model in ComfyUI and its substantial improvement in prompt understanding. Comparisons were made with SD1.5 finetunes and newer models, emphasizing prompt accuracy and performance.

Links mentioned:


OpenRouter (Alex Atallah) ā–· #announcements (1 messages):

  • Big update to /models page
  • Changing Google Token Sizes for Gemini and PaLM models
  • Deprecation of Default Model in settings page
  • Deprecation of custom auth headers for OpenAI API keys
  • Big Update Coming to /models Page: A significant update to the /models page is coming soon, with a sneak peek shared. Members are encouraged to provide feedback in the dedicated channel.
  • Google Token Sizes Changing for Gemini and PaLM Models: The Gemini and PaLM models will have their token lengths changed to match GPT-equivalent sizes, increasing token size roughly 3x and reducing context limits, leading to higher pricing but with the same model and API.
  • Deprecating the Default Model on Settings Page: The Default Model option on the /settings page is being deprecated as most apps set models themselves or use the auto router. Users with a valid use case are encouraged to provide feedback.
  • Deprecating Custom Auth Headers for API Keys: The use of custom auth headers for sending OpenAI API keys is being deprecated, with a replacement coming soon. This feature was used by a few people in mid-June but was never officially documented.

OpenRouter (Alex Atallah) ā–· #app-showcase (3 messages):

  • Quick and dirty wrapper shared by lastrosade
  • Feedback on the non-streamed response
  • Quick and dirty wrapper shared: lastrosade announced the creation of a quick and dirty wrapper, offering it to anyone interested in the community. No additional technical details or links were provided.
  • Feedback on the non-streamed response: A community member, clarie_starr, commented on the wrapper mentioning, ā€œSo all that for a non-streamed response. I’ll give it to you it’s um detailed.ā€ followed by lastrosade agreeing that the wrapper ā€œsucksā€.

OpenRouter (Alex Atallah) ā–· #general (77 messagesšŸ”„šŸ”„):

  • 500 errors with Claude 3.5
  • Self-moderation issues with Claude
  • Different frontends for using and jailbreaking Claude
  • OpenRouter privacy settings and logging policies
  • Google models token size change announcement
  • 500 Errors with Claude 3.5: Several users reported intermittent 500 errors while using Claude 3.5 on OpenRouter. Temporary fixes include switching to different versions like Claude 3.0.
  • OpenRouter Privacy and Logging Issues Addressed: Users discussed OpenRouter’s privacy settings, clarifying that some providers log requests while others do not, with an emphasis on NovitaAI and Infermatic not retaining data. Alex Atallah provided insights into the different privacy policies for third-party providers.
  • Google Models Token Size Update Clarification: A discussion on the token size change in Google models raised concerns about potential cost increases. LouisGV clarified that total pricing remains roughly the same despite the token size adjustment.
  • Exploring Different Frontends for Claude: Users explored various frontends like SillyTavern and LibreChat for jailbreaking or prefilling Claude models. Typingmind and Pal Chat were suggested as alternatives for a smoother user experience.
  • Quantization of LLM Models on OpenRouter: Questions about the quantization of deployed LLM models on OpenRouter were raised, focusing on whether models are in FP16 or other precisions. The discussion highlighted that models remain in their native precision unless specified otherwise by the provider.

Links mentioned:


Latent Space ā–· #ai-general-chat (40 messagesšŸ”„):

  • Magic.dev $500M to $1.5B valuation, 20 employees, no product, no revenue.
  • New paper on persona-driven data synthesis with 1 billion personas.
  • First real-time Audio LLM by Kyutai, 'Moshi'.
  • OpenDevin founders start All Hands AI.
  • Sentient's $85M seed round for open AGI platform.
  • Magic.dev valuation skyrockets to $1.5B despite no product: Magic.dev jumps in valuation from $500M to $1.5B with just 20 employees, no product, and no revenue.
  • Synthetic data creation gets boost with 1 billion personas: Persona Hub introduces 1 billion personas to scale synthetic data creation, showing massive gains in mathematical problem solving and diverse scenarios.
    • Presented by Aran Komatsuzaki, the new approach leads to significant improvements, particularly on the MATH benchmark.
  • Kyutai launches real-time Audio LLM ā€˜Moshi’: Moshi claims to be the first real-time Audio LLM with minimal delay, although the quality remains slightly robotic.
    • ā€˜Moshi’ demo by Kyutai shows promise despite its current limitations, as it sometimes interrupts the user in eagerness to respond.
  • Founders of OpenDevin launch All Hands AI: OpenDevin founders announce the formation of All Hands AI to accelerate AI software development for everyone in an open-source manner.
  • Sentient secures $85M seed round for open-source AI platform: Sentient announces $85M seed round co-led by Founders Fund to support the development of a community-built open AGI platform aimed at equitable AI development.
    • Prominent investors like Peter Thiel are backing the initiative, which aims to distribute AI benefits globally.

Links mentioned:


Latent Space ā–· #llm-paper-club-west (34 messagesšŸ”„):

  • openai AV issues during AIEWF demo
  • migration to Zoom for better accessibility
  • Discord's incompatibility with Linux and proposed alternatives
  • OpenAI’s AV Struggles during AIEWF Demo: Members experienced significant AV issues with Discord during the OpenAI AIEWF demo, leading to frustration and multiple participants being unable to see the screen. This resulted in a suggestion to switch platforms.
  • Zoom Migration for Better Accessibility: Due to continuous AV issues on Discord, members unanimously agreed to migrate to Zoom for the Paper Club (West) session. This shift aimed to resolve visibility problems and improve meeting quality.
  • Discord Incompatibility with Linux: Participants highlighted a known issue with Discord’s compatibility on Linux, leading to additional accessibility challenges. Alternatives were briefly discussed, suggesting a need for a more reliable platform moving forward.

Link mentioned: Join our Cloud HD Video Meeting: Zoom is the leader in modern enterprise video communications, with an easy, reliable cloud platform for video and audio conferencing, chat, and webinars across mobile, desktop, and room systems. Zoom …


LAION ā–· #general (2 messages):

  • jaan.li introducing their work at onefact.org and usb.club
  • san.tosh inquiring about updates on open GPT-4o
  • jaan.li builds decentralized edge transformers: jaan.li announced their work at onefact.org and usb.club on decentralized edge transformers. Contact them anytime at [email protected].
  • Inquiry on updates for open GPT-4o: san.tosh asked if there are any updates regarding open GPT-4o. The query remains open.

LAION ā–· #research (59 messagesšŸ”„šŸ”„):

  • Ablation tests and justification of changes for Terminator models
  • Discussions on slow-fast networks and their advantages
  • Released code for Terminator on GitHub
  • Introduction of FORA for accelerating Diffusion transformers
  • Critiques and suggestions for the HyperZā‹…Zā‹…W paper
  • Terminator model ablation tests criticized: Members discussed the lack of sufficient ablation tests and justification of changes in the Terminator models, though benchmarks showed impressive performance. They emphasized the need for detailed ablation studies to highlight the impact of individual components (e.g., residuals, dot product attention, intermediate pooling).
  • Debate on QKV redundancy in ViT: There was a heated debate on whether QKV in Vision Transformers (ViT) is redundant, with suggestions that Q & K might be unnecessary for attention matrix generation. Some members believe a proper evaluation and proof are required to validate this theory.
  • Terminator code released on GitHub: Terminator code has been released on GitHub for public access, contrary to some claims of it being vaporware. Users can now explore the official repository here.
  • FORA speeds up Diffusion transformers: A new approach, Fast-FORward CAching (FORA), was introduced to accelerate Diffusion transformers by caching and reusing intermediate outputs, significantly reducing computational overhead. This method integrates seamlessly with existing models and provides faster processing with minimal trade-offs in quality. Read more.
  • HyperZā‹…Zā‹…W paper receives mixed reviews: The HyperZā‹…Zā‹…W paper by @harvie_zhang received both praise and critique from the community, noting it as a rough submission with novel ideas for achieving SOTA. The author acknowledged the feedback, indicating future revisions and potential ablation studies to demonstrate the redundancy of QKV in ViT. Read the survey by Schmidhuber here.

Links mentioned:


tinygrad (George Hotz) ā–· #general (26 messagesšŸ”„):

  • Image dtype special treatment
  • Runtime error in tinygrad
  • UNMUL pattern matcher issue
  • Frontend fuzzer idea
  • Loop optimization bug
  • Tinygrad faces runtime error on UOps.UNMUL: A member reported a RuntimeError in tinygrad’s codebase: ā€˜failed to render UOps.UNMUL’. George Hotz suggested treating the issue as an assert and that it should ā€˜never happen’.
    • Chenyuy proposed adding flat_l4.realize() as a potential workaround and suggested making loop collapse optional to mitigate the impact on users.
  • Frontend fuzzer proposal for tinygrad: Chenyuy proposed the idea of a frontend fuzzer for tinygrad, possibly using an LLM to port torch code. The suggestion aims to catch more edge cases and unexpected behaviors during development.
  • Handling the UNMUL pattern matcher bug: Chenyuy noted that altering input dimensions sometimes prevents triggering the bug, which underscores incomplete loop optimization. Yosifrost found that specific dimensions impact the occurrence of the bug, suggesting an issue with heuristic boundary behaviors.
    • Members discussed the possibility of writing a minimal reproduction (repro) test to isolate the bug and expressed intentions to leave the PR open for further investigation and focused testing.
  • Error handling and dev tooling improvements before 1.0: Yosifrost stressed the necessity for better error messages and dev tooling in tinygrad before version 1.0. Various members collaborated to reproduce the error and develop a minimal test case for further debugging.

tinygrad (George Hotz) ā–· #learn-tinygrad (33 messagesšŸ”„):

  • Equivalent of torch.no_grad() in tinygrad
  • ā€œ-= operator incompatibility with gradient enabled in tinygrad
  • Handling gradient accumulation issues leading to CUDA memory errors
  • Slowdown and memory issues with TinyJit during gradient accumulation
  • Behavior of Tensor creation methods in tinygrad vs PyTorch
  • Tinygrad has torch.no_grad() equivalent: A user asked about the equivalent of torch.no_grad() in tinygrad, and it was explained that there’s Tensor.no_grad = True and the @Tensor.inference_mode() decorator with examples provided here.
  • Incompatibility of -= operator with gradient enabled: A user noted that a -= lr * a.grad will assert while a = a - lr * a.grad works due to gradient restrictions in tinygrad. The issue was illustrated with a reference to the code here.
  • Issues with gradient accumulation and CUDA memory: Users discussed problems with gradient accumulation leading to CUDA memory out of bounds errors. Suggestions included detaching the loss and addressing assert t.grad is not None from the optimizer.
  • Errors caused by TinyJit in optimization steps: It was revealed that TinyJit fails with assert t.grad is not None when not used on the entire step, causing inefficiencies. Users suggested returning realized gradients from the jit function and calculating the step externally.
  • Tensor creation method inconsistency in tinygrad: A user observed that Tensor.randn/randint create contiguous tensors, whereas Tensor.full creates non-contiguous tensors, differing from PyTorch’s behavior. It was confirmed as expected behavior in tinygrad, with potential improvements discussed for future versions.

Link mentioned: tinygrad/tinygrad/tensor.py at master Ā· tinygrad/tinygrad: You like pytorch? You like micrograd? You love tinygrad! ā¤ļø - tinygrad/tinygrad


LlamaIndex ā–· #blog (3 messages):

  • Building a RAG pipeline on Raspberry Pi
  • OpenContracts AI-powered document analytics tool
  • Webinar on RAG experimentation and evaluation with Weights & Biases
  • RAG on Raspberry Pi šŸ“šŸ”Ž: A tutorial by @pavan_mantha1 demonstrates how to build a RAG pipeline on a Raspberry Pi using Docker and Ollama. Check out the tweet for more details.
    • This project showcases how the RAG pipeline can be efficiently run on a small, embedded device like the Raspberry Pi.
  • OpenContracts Launches ✨: OpenContracts, an open-source AI-powered document analytics tool by @johnscrudato, allows users to analyze, annotate, and share documents using LLMs and Llama Index. More information can be found here.
    • This project is genAI native and aims to democratize document analytics by using AI efficiently.
  • Webinar on RAG Experimentation 🚨: A webinar on RAG Experimentation and Evaluation is being held in partnership with Weights & Biases. The session aims to teach how to build, evaluate, and iterate on RAG pipelines, as detailed in this tweet.
    • With over a year of RAG development, this webinar addresses the challenge of proper evaluation in the field.

LlamaIndex ā–· #general (49 messagesšŸ”„):

  • DocumentSummaryIndex issues with Pinecone limits
  • Code snippets and potential fixes for metadata exclusion
  • Alternative vector stores to Pinecone
  • LlamaIndex's single LLM support
  • Parsing issues with PDF tables
  • DocumentSummaryIndex hits Pinecone limits: A user reported issues with Big Docs exceeding Pinecone limits when creating a DocumentSummaryIndex. The issue arises from the first node’s metadata being too large, and embed exclusion filters seemingly not applied properly. GitHub link
    • Another user suggested not including too much metadata in documents/nodes and offered a potential code fix to exclude embedding metadata keys. Additionally, they suggested considering alternatives to Pinecone like qdrant or pg_vector.
  • LlamaIndex supports only OpenAI LLMs: Multiple users noted that LlamaIndex currently only supports OpenAI as its LLM, which was met with some dissatisfaction. They suggested that expanding support to other LLMs would be beneficial.
    • ā€˜Looks like it currently only supports OpenAI as LLM … if so then šŸ‘Žā€™
  • PDF to Markdown parsers struggle with tables: A user tried converting PDFs to Markdown using Marker but found that ā€˜funkily’ formatted tables caused parsing issues. They are looking for a better local or open-source solution but mentioned that Azure Document Intelligence performs better.
    • Another user recommended trying Unstructured or llamaparse, even though they are not open-source. These tools seem to handle complex table structures better.

Links mentioned:


LlamaIndex ā–· #ai-discussion (5 messages):

  • Agentic RAG with LlamaIndex, Claude-3.5 Sonnet, and MongoDB
  • Toolio for running private AI/LLM agents and tool-calling workflows on Mac
  • Agentic RAG gains interest: Unleashing AI Potential: Agentic RAG with LlamaIndex, Claude-3.5 Sonnet, and MongoDB article discusses innovative strategies in the realm of AI. A member hinted it will be promoted soon.
  • Toolio simplifies private AI workflows on Mac: Toolio allows users to run private AI/Large Language Model (LLM) agents & tool-calling workflows on their Mac with ease, supporting JSON schema constraints and offering fast inference. An advocate claimed tool-calling is the ā€˜true magic with LLMs’ and anticipates meaningful innovations in this area.

Nous Research AI ā–· #off-topic (1 messages):

  • Tortoise-TTS converted to ggml
  • Optimization for real-time inference
  • Open-source projects on GitHub
  • CUDA and CPU support for Tortoise-TTS
  • Tortoise-TTS Conversion to GGML: A member converted Tortoise-TTS to ggml and is seeking assistance to improve inference time for real-time text-to-speech. The repository already supports CUDA and CPU.
  • Optimization Opportunity for AI Developers: The Tortoise-TTS ggml project offers a great chance to practice optimizing transformers and diffusion models. The goal is to speed up the inference process.

Link mentioned: GitHub - balisujohn/tortoise.cpp: A ggml (C++) re-implementation of tortoise-tts: A ggml (C++) re-implementation of tortoise-tts. Contribute to balisujohn/tortoise.cpp development by creating an account on GitHub.


Nous Research AI ā–· #general (42 messagesšŸ”„):

  • Private channels on decentralized training
  • Tool calling in vLLM for Hermes 2 Pro
  • Discussion on handling tool calls and text content in Hermes 2 Pro
  • Tool calling working properly in vLLM for Hermes 2 Pro: A member announced that tool calling is now functioning properly in vLLM for Hermes 2 Pro. They indicated the project is very close to completion.
  • Hermes 3 training includes <scratch_pad>: The team discussed the addition of <scratch_pad> before tool calls in Hermes 3 training, aiming for improved parsing that extracts between <scratch_pad> and handles both ā€˜content’ and ā€˜tool_calls’.
    • Discussions included handling text content before tool calls and ensuring compatibility with OpenAI’s spec.

Link mentioned: Neural Networks: Zero to Hero: no description found


Nous Research AI ā–· #ask-about-llms (3 messages):

  • Creating conversational dataset from documents
  • Instruction-generation from documents
  • Genstruct 7B model for generating instructions
  • Creating datasets from documents: Creating conversational datasets from documents depends on the document and budget. Options include generating datasets using a language model or using tools like those from Anthropic.
  • Genstruct 7B generates instructions: The Genstruct 7B model generates valid instructions from a raw text corpus, useful for creating instruction finetuning datasets. It’s inspired by the Ada-Instruct model.

Link mentioned: NousResearch/Genstruct-7B Ā· Hugging Face: no description found


Nous Research AI ā–· #rag-dataset (5 messages):

  • Huggingface's PR on Cohere's CommandR model
  • Microsoft's GraphRAG release
  • Huggingface PR on Cohere CommandR Expands Tool-Use: Huggingface opened a PR on Cohere’s CommandR model, focusing on technical improvements for tool-use and RAG templates. The system prompt is formulated using a preamble and dynamic content arrangement with Jinja templating.
  • Microsoft Releases GraphRAG: Microsoft released a modular graph-based Retrieval-Augmented Generation (RAG) system called GraphRAG, designed to boost information retrieval and generation. The tool is available on GitHub and aims to enhance modularity and effectiveness in RAG systems.

Link mentioned: GitHub - microsoft/graphrag: A modular graph-based Retrieval-Augmented Generation (RAG) system: A modular graph-based Retrieval-Augmented Generation (RAG) system - microsoft/graphrag


Modular (Mojo šŸ”„) ā–· #general (7 messages):

  • Installation issues on Ubuntu 24.04/Python 3.12.3
  • Implementation workaround for Mojo/max on Ubuntu 24.04/Python 3.12.3
  • Mojo implicit conversion bug
  • Casting bug in Mojo
  • Facing installation issue on Ubuntu 24.04: A user reported an installation issue on Ubuntu 24.04/Python 3.12.3, receiving errors for max-engine due to version mismatches.
    • Another user shared a step-by-step guide to resolve this by installing Python 3.11 and adjusting alternatives.
  • Mojo’s odd implicit conversion: A user noticed that multiplying integers by np.pi in Mojo produces an unexpected negative integer result due to an implicit conversion bug.
    • The discussion pointed out that this is related to casting bugs already tracked as #3065 and #3167.

Links mentioned:


Modular (Mojo šŸ”„) ā–· #šŸ’¬ļø±twitter (2 messages):

  • ā€œ
  • Modular Shares Exciting Updates on Twitter: Modular tweeted interesting updates recently. More details can be found in their latest Twitter announcements.
  • Further Announcements from Modular on Twitter: Modular shared additional updates a few days later. Check their Twitter for the full posts.

Modular (Mojo šŸ”„) ā–· #āœļø±blog (1 messages):

  • Mojo N-Body Example Benchmark
  • Single-core Numeric Performance in Mojo
  • Symplectic Integrator in N-body.js
  • Vectorization in N-Body Example
  • Ordinary Differential Equation Solver

Link mentioned: Modular: A Brief Guide to the Mojo N-Body Example: We are building a next-generation AI developer platform for the world. Check out our latest post: A Brief Guide to the Mojo N-Body Example


Modular (Mojo šŸ”„) ā–· #šŸ”„mojo (8 messagesšŸ”„):

  • Mojo List types and printing issues
  • Printing RepresentableCollectionElement types
  • Printing errors inline in Mojo
  • Impact of excessive empty lines on startup time
  • Variable startup times in Mojo programs due to loop unrolling in bench_matmul
  • Mojo List Confusion: A member expressed confusion about List[String] in Mojo, noting that although it contains strings, it lacks the Stringable trait, affecting its printability. They provided a GitHub link for reference.
  • Printing RepresentableCollectionElement types in Mojo: A suggestion was made to use print(list.__str__()) to print Lists of RepresentableCollectionElement types in Mojo, providing a GitHub link for more details.
  • Variable Startup Times Due to Loop Unrolling: A user noted that their Mojo program had variable startup times, initially linking it to excessive empty lines. Another user clarified that it was due to bench_matmul unrolling a lot of loops, making it slow to compile.

Links mentioned:


Modular (Mojo šŸ”„) ā–· #nightly (1 messages):

melodyogonna: The joys of early tooling


Modular (Mojo šŸ”„) ā–· #mojo-marathons (31 messagesšŸ”„):

  • Parallel processing in MLM
  • Matrix Multiplication Optimization
  • Strassen Algorithm Performance
  • SPIRAL Project
  • Numerical Stability in Matrix Multiplication
  • SPIRAL Project Aims at Automated High-Performance Libraries: The SPIRAL project focuses on automating software and hardware development for DSP algorithms and other numerical kernels, often outperforming MKL in direct hardware tasks.
  • Strassen Algorithm vs Naive Vectorized Approach: Strassen Algorithm achieved around 50 GFlops while the naive vectorized and parallelized version hit 70 GFlops for 1024x1024 matrices. Details on GitHub.
    • Optimizations like vectorizing adds/subs and parallelizing sub-matrix multiplications added 3-5 extra GFlops, while fewer intermediate allocations and better guardrails for non-square matrices are potential future improvements.
  • Parallel and Vectorized Operations Conceptual Challenge: Discussions highlighted the conceptual difficulty beyond parallel, vectorize, unroll, and parts of tiling for tuning algorithms.
  • Numerical Stability Impact in Strassen Algorithm: The Strassen Algorithm reportedly reduces numerical stability, leading to test failures when different types and sizes were used.
  • Recursive vs Iterative Algorithms for Cache Locality: It’s suggested that different type sizes might benefit from recursive algorithms over iterative ones for better cache locality.

Links mentioned:

  • SPIRAL Project: Home Page: no description found
  • Matrix Multiplication: Sheet1 Contstraints,Parameters / Tuning Vectorization,Contiguous Access,Nelts, Unrollable Parallelization,Unrollable Unrolling,Contiguous Operations Tiling Square Optimized,Amorized Increase,Recursiv...

Interconnects (Nathan Lambert) ā–· #news (29 messagesšŸ”„):

  • Apple getting a board observer seat at OpenAI
  • Microsoft investments in OpenAI and comparison with Apple's partnership
  • Kyutai Labs' new real-time audio LLM 'Moshi'
  • Training details and technical specifics of 'Moshi'
  • Kyutai Labs' open model releases and future plans
  • Apple Gets a Board Observer Seat at OpenAI: Apple will get a board observer seat at OpenAI later this year as part of its partnership for Apple Intelligence, with Phil Schiller occupying the seat, as reported by Bloomberg.
  • Microsoft vs. Apple in OpenAI Partnerships: Community members compared Microsoft’s billion-dollar investment in OpenAI to Apple’s partnership, noting that Apple seems to be getting more benefits, including an app and iPhone integration, while Microsoft does not.
  • Kyutai Labs Unveils ā€˜Moshi’: The Real-Time Audio LLM: Kyutai Labs introduced Moshi, the first real-time Audio LLM with 150ms latency, during a live update, showcasing its multimodal LM capabilities with the potential for on-device use.
  • Moshi’s Impressive Technical Details: Moshi is built on a 7B multimodal LM, has a VQ-VAE-based speech codec with a 300x compression factor, and delivers superhuman response times, according to multiple community members.
  • Open Model Releases and Future Plans for ā€˜Moshi’: Kyutai Labs plans to release open models including the 7B multimodal LM, audio codec, and an optimized stack. Users have already begun testing the model in real-world scenarios and discussed its response latency and potential uses.

Links mentioned:

  • Tweet from Lucas Beyer (bl16) (@giffmana): Kyutai Moshi - first real-time Audio LLM. Basically no delay - the LLM even interrupted the speaker a few times. It was actually a bit eager to answer very quick. :) All to be open-sourced. Quality ...
  • Tweet from Bartok Gabriel (@BartokGabi17): @markgurman Microsoft invests bilion in open Ai dosent get an app Apple pays litterally in exposure, open Ai makes a great app an big iPhone integration Profit?? Tim Apple it's a genius
  • Tweet from Xeophon (@TheXeophon): Quoting kyutai (@kyutai_labs) Join us live tomorrow at 2:30pm CET for some exciting updates on our research! https://www.youtube.com/live/hm2IJSKcYvo
  • Tweet from Vaibhav (VB) Srivastav (@reach_vb): Fuck yeah! Moshi by @kyutai_labs just owned the stage! šŸ‡ŖšŸ‡ŗ/acc. Architecture 1. 7B Multimodal LM (speech in, speech out) 2. 2 channel I/O - Streaming LM constantly generates text tokens as well as a...
  • Tweet from Mark Gurman (@markgurman): NEW: Apple will get a board observer seat at OpenAI later this year as part its partnership for Apple Intelligence. The person getting the seat: Phil Schiller, the head of the App Store and former mar...

Interconnects (Nathan Lambert) ā–· #posts (12 messagesšŸ”„):

  • SB 1047 and First Amendment challenges
  • Protection of 3D gun designs under the First Amendment
  • Model weights and code as protected speech
  • Claude 3.5 admiration
  • Use Claude TM
  • SB 1047 faces potential First Amendment challenge: A discussion emerged on whether SB 1047 would survive a First Amendment challenge if it passes, particularly comparing it to the protection of 3D gun designs and code as free speech, referencing the EFF’s case on code as speech.
  • Courts on code as protected language: A quote highlighted that there is ā€˜no meaningful difference between computer language and German or French,’ citing that both communicate information and thus are protected under the First Amendment, likening them to ā€˜music and mathematical equations.’
  • Debate on model weights as protected speech: Debate continued on whether model weights could be considered protected under the First Amendment, with discussions around their classification as ā€˜high-level languages’ or merely ā€˜mathematical equations,’ and comparisons to publishing random sequences of words.
  • Claude 3.5 hype: ā€œwoo Claude 3.5 admiration let’s goā€ – sparked excitement among members regarding the recent release of Claude 3.5.
  • Promoting Claude TM: A suggestion was made to promote Claude TM, comparing its campaigns to those of American Airlines with the enthusiastic remark, ā€˜I never would go back.’

Link mentioned: Deeplinks Blog: no description found


LangChain AI ā–· #general (30 messagesšŸ”„):

  • RAG strategies for answering general questions
  • Rate limiting issues with AzureAIDocumentIntelligenceLoader
  • Parsing PDFs with different tools and libraries
  • LangSmith tracing issues
  • General help and troubleshooting in LangChain
  • Azure and PDF Loaders Clash: A user switched from PyPDFium2Loader to AzureAIDocumentIntelligenceLoader and encountered a consistent 429 error (Too Many Requests). This suggests possible rate limiting issues due to the method AzureAIDocumentIntelligenceLoader processes documents.
  • PDF to Markdown Pain Points: A member tried using marker to convert PDFs to markdown but faced issues with parsing tables with ā€˜funky’ formatting like merged cells. They noted that Azure Document Intelligence worked better for such documents but expressed a preference for a local or open-source alternative.
  • LangSmith Stops Tracing Calls: A user reported that LangSmith stopped tracing their calls without any clear reason. This suggests potential bugs or issues in the tracing mechanism of LangChain.

Links mentioned:


LangChain AI ā–· #langserve (2 messages):

  • uploading CSV files directly vs. providing file path
  • output not displaying in CSV playground
  • code improvement for CSV handling
  • FastAPI endpoint for file uploads
  • Chroma vectorstore usage and issues
  • Enable Users to Upload CSV Files: A user seeks help to enable CSV file uploads in their project, instead of users specifying file paths. They need a way to implement this in their FastAPI setup for better usability.
  • No Output in CSV Playground: In the csv/playground/ directory, there’s an issue where no output is displayed even though the code seems correct. This indicates a potential problem in the file handling or output rendering logic.
  • Improving CSV Handling Code: The user is looking for guidance to improve their existing code, which currently requires users to set the file path. Suggestions to enhance code efficiency and usability are needed.

LangChain AI ā–· #share-your-work (2 messages):

  • OpenAI CriticGPT paper discussion
  • Toolio open source project for private LLMs
  • OpenAI Introduces CriticGPT to Identify GPT-4 Mistakes: A member shared a YouTube video discussing OpenAI’s latest paper on CriticGPT, which is designed to correct errors made by GPT-4. The video highlights the key features of CriticGPT and its significance in improving the reliability of AI-generated code.
  • Toolio Empowers Private LLM Workflows on Mac: A member announced the release of Toolio, an open source project that enables running private LLM agents and tool-calling workflows on Mac with ease. The project, showcased in a YouTube video, also features JSON schema output constraints and fast inference capabilities.

Link mentioned: OpenAI releases CriticGPT to correct GPT-4’s mistakes | Read the paper with me: OpenAI has unveiled CriticGPT, a new AI model based on GPT-4 designed to identify errors in code generated by ChatGPT, marking a significant step towards imp…


LangChain AI ā–· #tutorials (1 messages):

dracount: hi, is there a beginner langchain/langraph tutorial that anyone can recommend?


Mozilla AI ā–· #llamafile (25 messagesšŸ”„):

  • Hardware recommendations for running llamafile
  • VRAM and CPU usage for large language models
  • Syncthread trick for CPU inference
  • Running llama3 70B on high-end workstation
  • Issues with Rockchip RK3588 NPUs support for llamafile
  • Building the Best Llamafile Setup: Members discussed hardware recommendations for a new Linux computer to run llamafile effectively. Recommended GPUs include 3090/4090 for consumer use and A6000/RTX 6000 Ada for workstations; at the higher end, older EPYC CPUs were suggested for their cores and PCIe lanes support.
  • VRAM Needs for Large Models: To run large models efficiently, more VRAM is crucial; for example, 24GB VRAM can handle 33B parameters on q4. FP16 mode is discouraged due to its massive VRAM requirements compared to minimal quality loss.
  • Syncthread Trick in CPU Inference: Members discussed the potential of CPU-learning with the syncthread trick used for CPU inference. A YouTube talk was shared explaining the concept.
  • Running Llama3 70B on Threadripper: A member successfully ran llama3 70B on a high-end Threadripper CPU workstation. This shows the capability of CPU for handling large models when adequate specifications are met.
  • Issues with Rockchip RK3588 NPUs: There were questions about running llamafile on Rockchip RK3588 NPUs. It was suggested to use version v0.8.9 to avoid address space issues on such hardware.

Links mentioned:


Torchtune ā–· #general (22 messagesšŸ”„):

  • phi mini new weights same repo
  • torchtune evaluation using eleutherai's eval harness
  • evaluation basics and logs on wandb
  • discussion about gradients and epochs for training
  • FullModelHFCheckpointer and conversion between HF format and torchtune
  • New weights for phi mini: phi mini has received new weights but is reusing the same repository as before. Users are assuming the old recipe will work without updates for torchtune.
  • Training strategies and evaluations: Discussion on different gradients 8 vs 16 and batch size 2 to determine which works better for a dataset and whether to adjust epochs. Wandb logs and evaluation tutorials were shared to aid in better understanding and tracking performance metrics.
  • Conversion arguments for HF formats: Queries about why conversion needs parameters like num_heads, num_kv_heads, and dim. Conversion is needed to switch between HF grouped multihead layers and torchtune’s separate layers.
  • Checkpointers for torchtune: torchtune’s FullModelHFCheckpointer automatically converts checkpoints to HF format. Details were shared about how the checkpointer ensures compatibility with various tools and handles different formats.

Links mentioned:


Cohere ā–· #general (14 messagesšŸ”„):

  • Training LLM with Stockfish data
  • Usage of tools like Stockfish for reasoning in LLMs
  • GitHub notebook code
  • Chess strategy and LLMs
  • Cohere API tools
  • **Mixing LLMs with Stockfish for Better Planning: **: A user raised a question about using Stockfish data to improve LLM reasoning and planning abilities, or to develop a quick chess engine.
  • **Fine-Tuning LLMs with Chess Data: **: One member shared their experience of fine-tuning LLMs with chess data, highlighting that it requires significant overfitting, which can be problematic. There’s a debate on the effectiveness and utility of such an approach.
  • **Using Tools for Chess in Cohere’s Platform: **: An interesting perspective suggested that LLMs should use tools like Stockfish for better results in chess understanding, rather than being trained directly on chess data.

Link mentioned: GitHub: Let’s build from here: GitHub is where over 100 million developers shape the future of software, together. Contribute to the open source community, manage your Git repositories, review code like a pro, track bugs and fea…


Cohere ā–· #project-sharing (5 messages):

  • Creation of a Cohere Slack bot
  • Discussion on Slack's request handling
  • Lack of documentation on Slack integration
  • Offer to share script and create documentation
  • Cohere Slack Bot Simplifies Workspace Interaction: A member shared their creation of a Cohere Slack bot designed for workspace convenience, enhancing accessibility.
  • Slack’s Efficiency Challenges Bot Creation: With Slack’s requirement to complete requests in under 3 seconds, this showcases the impressive speed of the Cohere models utilized.
    • ā€œI have a Cloudflare Worker handling requestsā€, demonstrating practical integration solutions despite initial documentation complexities.
  • Community Eagerly Awaits Bot Documentation: The community showed enthusiasm for the bot creation, seeking guidance and documentation on replicating the process.
    • ā€œI will work on a tutorial with documentation and publish it to one of my domainsā€, indicating future resource availability for interested members.

OpenInterpreter ā–· #general (18 messagesšŸ”„):

  • Kyutai Moshi - real-time Audio LLM
  • Various Open Interpreter compatible projects
  • Experience modding games for Open Interpreter
  • Pull request for Open Interpreter Labs
  • MIke Bird and blurryboi discussion on Kyutai Moshi
  • Kyutai Moshi launches real-time Audio LLM: Kyutai Moshi released the first real-time Audio LLM, with no delay but some robotic quality, which will be open-sourced. Live demo is available.
    • Mikebirdtech noted it was FAST, almost too fast, interrupting if pauses were too long.
  • Open Interpreter projects suggested by Techfren: Several projects like Open interpreter, taxyai, clickolas cage, self-operating computer, pywinassistant, GPT computer assistant were listed as Open Interpreter compatible. They may require some configuration but are promising options.
  • Seeking modder experience for Open Interpreter: Nonadjective.eth_55058 expressed interest in modding a game to be compatible with Open Interpreter and sought advice from others with similar experience. They are open to clunky interfaces to create a proof of concept.

Links mentioned:

  • moshi.chat: no description found
  • Open Interpreter Labs : Open Interpreter Labs and Experiments (not directly affiliated with OI) - Open Interpreter Labs
  • Tweet from Lucas Beyer (bl16) (@giffmana): Kyutai Moshi - first real-time Audio LLM. Basically no delay - the LLM even interrupted the speaker a few times. It was actually a bit eager to answer very quick. :) All to be open-sourced. Quality ...

OpenInterpreter ā–· #O1 (1 messages):

johnlenflure: Isn’t there a way to integrate 01 into glasses?


OpenAccess AI Collective (axolotl) ā–· #general (1 messages):

  • Weighted cross entropy in the trainer
  • Discovery of Weighted Cross Entropy Section: A member noticed the whole weighted cross entropy section in the trainer. They mentioned they would look into it further.
  • Lack of Additional Topics Noted: No other significant topics were discussed in the given message history. Only weighted cross entropy in the trainer was mentioned.

OpenAccess AI Collective (axolotl) ā–· #general-help (6 messages):

  • Differences between LoRA and QLoRA quantization
  • Explanation of 8-bit quantization in LoRA
  • Efficiency of QLoRA in fine-tuning large models
  • LoRA is quantized to 8-bit: LoRA applies 8-bit quantization while QLoRA goes further to use 4-bit quantization as clarified by users in the discussion.
    • The conversation highlighted how quantization is not just about decomposing matrices, referencing the QLoRA paper for a comprehensive explanation.
  • QLoRA facilitates efficient fine-tuning: A member shared that the QLoRA paper demonstrates how QLoRA enables finetuning a 65B parameter model on a single 48GB GPU with near-full 16-bit finetuning performance.
    • This paper introduces innovations like 4-bit NormalFloat (NF4) and double quantization to minimize memory usage without sacrificing performance, according to a user discussing the content.

Link mentioned: QLoRA: Efficient Finetuning of Quantized LLMs: We present QLoRA, an efficient finetuning approach that reduces memory usage enough to finetune a 65B parameter model on a single 48GB GPU while preserving full 16-bit finetuning task performance. QLo…


OpenAccess AI Collective (axolotl) ā–· #axolotl-help-bot (2 messages):

  • torch.cuda.OutOfMemoryError on Google Colab
  • Axolotl running issues
  • GPU memory allocation
  • VRAM requirements
  • CUDA Memory Woes in Google Colab: A member encountered a torch.cuda.OutOfMemoryError while attempting to allocate 172.00 MiB GPU memory running axolotl on Google Colab.
  • Axolotl Demands More VRAM: In response to a CUDA memory error, another member advised that more VRAM is needed to avoid such issues.

OpenAccess AI Collective (axolotl) ā–· #axolotl-phorm-bot (5 messages):

  • Quantization and its impact on model performance
  • LoRA versus QLoRA configuration specifics
  • Memory footprint and inference speed improvements with 8-bit quantization
  • 8-bit Quantization Explained: Quantization reduces model precision from 32-bit or 16-bit down to 8-bit (int8), significantly reducing memory footprint and speeding up inference times. The load_in_8bit option in lora.yaml enables this quantization process for deploying large models on limited hardware.
  • Differences between LoRA and QLoRA: While LoRA focuses solely on parameter-efficient fine-tuning, QLoRA combines low-rank adaptation with quantization. The inclusion of load_in_8bit in QLoRA configurations signifies the use of 8-bit quantization, as seen in various example files (qlora.yml).

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


AI Stack Devs (Yoko Li) ā–· #ai-town-discuss (11 messagesšŸ”„):

  • Docker port for AI Town
  • GitHub Page for AI Town Windows Setup with WSL
  • API communication issues with Docker port of AI Town
  • Convex automatic download via Docker for AI Town
  • Testing Docker integration for AI Town
  • Docker request for AI Town: A member highlighted the need for a Docker image for AI Town as ā€˜amazing’ and useful, encouraging submissions to the main repo.
  • GitHub setup guide for AI Town on WSL: A member shared a GitHub page for setting up AI Town on Windows using WSL, and it was suggested to submit a PR for incorporating it into the main repo.
  • API issues in Docker port for AI Town: While working on a Docker port, a member noted, ā€˜I can run AI town with no problem except one: Ollama and other API communication,’ and promised updates once a solution is found.
  • Automation of Convex download in Docker: A member is finalizing adjustments so that Convex automatically downloads via Docker, aimed at simplifying user experience and planning to go live around 8 p.m. UTC+4.
  • Request for help with Docker integration testing: A request for testing Docker integration was made before submitting a PR, but the member decided to proceed after successful tests on their Legion Go.

Link mentioned: GitHub - Ikkitsuna/AI-Town-Windows-Setup-WSL-method: Guide for setting up AI Town on Windows using WSL: Guide for setting up AI Town on Windows using WSL. Contribute to Ikkitsuna/AI-Town-Windows-Setup-WSL-method development by creating an account on GitHub.


LLM Finetuning (Hamel + Dan) ā–· #🟩-modal (3 messages):

  • struggling to deploy RAG app using gradio on Modal
  • post on Modal Slack for help
  • Struggling with RAG App Deployment: A member mentioned facing issues deploying a RAG app using Gradio on Modal, despite it running fine locally and being deployed on Huggingface Spaces. They have exhausted all options and are looking for resources to figure out what went wrong.
  • Modal Slack to the Rescue: Another member suggested posting the issue on the Modal Slack for further assistance. The original member acknowledged the reminder and expressed their intent to do so.

LLM Finetuning (Hamel + Dan) ā–· #paige_when_finetune (1 messages):

shamik_53759: Yep, it’s up now. Thanks!


LLM Finetuning (Hamel + Dan) ā–· #axolotl (1 messages):

  • DeepSpeed configuration for data sharding and disabling model sharding
  • Assistance with DeepSpeed configurations
  • Confusion about DeepSpeed settings
  • Navigating DeepSpeed Configuration for Sharding: A member expressed confusion over selecting the appropriate DeepSpeed configuration to enable data sharding and disable model sharding.
  • Request for Assistance with DeepSpeed Settings: A request was made for assistance in choosing the correct DeepSpeed settings for enabling data sharding while disabling model sharding.

LLM Finetuning (Hamel + Dan) ā–· #freddy-gradio (1 messages):

  • Sharing private code deployments
  • Running code on multiple platforms
  • Limitations of private deployments on Hugging Face
  • Struggles with Private Code Deployments: A user expressed difficulty in sharing private code with colleagues using Hugging Face private spaces, noting that sharing=True is not supported in private deployments.
  • Challenges Running Code on Modal: Having issues with getting code to run on Modal, a user mentioned their struggles and expressed interest in alternative solutions for private code sharing.

LLM Perf Enthusiasts AI ā–· #eval (1 messages):

  • Evaluating LLM accuracy in legal contract review
  • Screens tool achieving 97.5% accuracy
  • Methodologies for assessing LLMs in the legal domain
  • Impact of different LLMs and methods on AI accuracy in legal tasks
  • Screens Report Evaluates LLM Accuracy in Legal Domain: Screens’ new evaluation report discusses treating LLM evaluations like traditional ML classification problems in the legal domain, particularly for contract reviews. They claim a 97.5% accuracy rate for their system, highlighting its potential use in playbook execution and workflow routing.
  • Accuracy Challenges and Methodologies Explored: The report details the challenges in objectively evaluating long-form, free-text responses, proposing a methodology that uses classification standards to assess LLM performance. The approach can significantly aid in legal tasks like negotiation, redlining, and summarization.

Link mentioned: Screens Accuracy Evaluation Report: Evaluating the accuracy of large language models (LLMs) on contract review tasks is critical to understanding reliability in the field. However, objectivity is a challenge when evaluating long form, f…


LLM Perf Enthusiasts AI ā–· #prompting (1 messages):

  • ā€œ
  • Seeking Simple Prompt Tuning Tools: Evan_04487 is looking for a user-friendly tool for tuning templates prompts, tailored for non-technical stakeholders like designers and product managers. They need a hosted, freemium product that can run variations of templated prompts and manually inspect responses.
  • Requirement for Freemium Hosted Tool: Evan_04487 specified a preference for a freemium hosted tool for prompt tuning that can handle a couple dozen variables. They mentioned having a more robust self-service infrastructure for high-stakes exercises, but need something simpler for lower stakes.

Datasette - LLM (@SimonW) ā–· #ai (1 messages):

derekpwillis: https://thescoop.org/archives/2024/06/22/all-foreign-gifts-around-us/index.html





{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}