> AI Discords for 1/29/2024. We checked **21** guilds, **311** channels, and **8276** messages for you. Estimated reading time saved (at 200wpm): **605 minutes**.

The surprise release of CodeLlama from Meta AI is an incredible gift to open source AI:

image.png

As can be expected, the community has already got to work putting it on Ollama and MLX for you to run locally.


Table of Contents

[TOC]

PART 1: High level Discord summaries

TheBloke Discord Summary

  • Miqu Model Sparks Intrigue and Debate: The Miqu model has sparked debates among users like @itsme9316 and @rombodawg, wondering if it’s a leak from Mistral Medium or a fine-tuned version of Llama-2-70b. Concerns about AI ethics and the future risks of AI were discussed, including the misuse of powerful models and the challenges of achieving proper alignment.

  • Aphrodite Engine’s Performance Highlighted: Users shared their experiences with the Aphrodite engine, indicating it offers impressive performance at 20 tokens per second on A6000 GPUs with configurations like --kv-cache-dtype fp8_e5m2, though concerns were raised about GPU utilization and support for model splitting.

  • Role-Playing AI Models Discussed for Repetitiveness and Performance: In role-playing settings, models like Mixtral and Flatdolphinmaid showed challenges with repetitiveness, while Noromaid and Rpcal performed better. ChatML with Mixtral Instruct was recommended by @ks_c for role-play, highlighting the importance of Direct Prompt Optimization (DPO) to improve AI character responses.

  • Learning and Fine-Tuning on ML/DL: @dirtytigerx advises that CUDA programming isn’t essential for starting with ML/DL, recommending the Practical Deep Learning for Coders course by fast.ai for a solid foundation. Discussions included VRAM requirements for fine-tuning large models like Mistral 7B, and exploring advanced optimizer technology like Paged 8bit lion or adafactor for efficient VRAM usage.

  • Enhancing GitHub Project Navigation and Understanding Codebases: Guidance on understanding GitHub projects for contributions was offered, with start.py in tabbyAPI project highlighted as an example entry point. The importance of IDEs or editors with language servers for efficient code navigation was discussed, along with the release of Code Llama 70B on Hugging Face, emphasizing its potential in code generation and debugging across several programming languages.


Nous Research AI Discord Summary

  • Activation Beacon Revolutionizes LLM Context Length: The Activation Beacon project introduces a significant advancement, enabling unlimited context length in LLMs through “global state” tokens. This development, as shared by @cyrusofeden, could dramatically alter the approach to Dense Retrieval and Retrieval-augmented LLMs with its details available on GitHub.

  • Eagle-7B Surpasses Mistral with RWKV-v5 Architecture: Eagle-7B, based on the RWKV-v5 architecture, has been highlighted as outperforming even Mistral in benchmarks, presenting a promising open-source, multilingual model with efficiency gains. It boasts lower inference costs while maintaining high English performance (source).

  • OpenHermes2.5 Touted as Ideal for Consumer Hardware: Recommendations for OpenHermes2.5 emerge as ideal for running on consumer hardware, attributed to its gguf or gptq quantization, positioning it as a leading open-source LLM for question-answering. Further discussions also explore the Nous Hermes 2 Mixtral 8x7B DPO on Hugging Face.

  • Exploration into Multimodal and Specific Domain LLMs: Discussions around using the IMP v1-3b model for classification tasks in orbital imagery suggest a growing interest in multimodal models and their application in specific domains. The dialogue extends to comparisons with other models like Bakllava and Moondream, highlighting the prowess of Qwen-vl in this space.

  • Collective Ambition for Centralized AI Resources: The community’s efforts in centralizing AI resources and research, spearheaded by @kquant, demonstrate a recognized need for a consolidated repository. This initiative aims to streamline access to AI papers, guides, and training resources, inviting contributions to enhance the resource pool for enthusiasts and professionals alike.


LM Studio Discord Summary

  • GPU Choice for LLM Work: Despite older P40/P100 GPUs having more VRAM, the newer 4060Ti is recommended for training Large Language Models (LLM) due to receiving current updates from Nvidia. Participants, including @heyitsyorkie, discussed GPU preferences, noting the importance of updates over GRAM speed.

  • Model Compatibility and Format Updates: LMStudio’s shift from GGML to GGUF format for models, as notified by @heyitsyorkie, indicates a move towards more current standards. Further, discussions highlighted the utility of dynamically loading experts in models for MoE (Mixture of Experts) applications within LMStudio.

  • Performance Issues and UI Challenges in LMStudio Beta: Users @msz_mgs and @heyitsyorkie reported performance lag when inserting lengthy prompts and navigational difficulties within the UI. These issues are recognized and slated for resolution in future updates.

  • Exploration of VRAM and GPU Utilization: Discussions led by @aswarp and @heyitsyorkie explored the unexpected VRAM overheads for running models in LMStudio, highlighting nuances in offloading models to GPUs including the RTX 3070 and 4090. Rumors about an upcoming RTX 40 Titan with 48GB VRAM stirred community discussions, linking to broader market strategies and repurposing GPUs for AI.

  • Beta Updates and Cross-Platform Compatibility Queries: The announcement of Beta V2 indicated progress on VRAM estimates and compatibility issues. Queries about Mac Intel versions, Linux appimage availability, and WSL (Windows Subsystem for Linux) support underscored the community’s interest in broadening LMStudio’s platform compatibility.


OpenAI Discord Summary

  • RAG Demystified: In response to @aiguruprasath’s query, Retrieval Augmented Generation (RAG) was explained by @darthgustav. as AI with the capability to directly access knowledge for searching data or semantic matching, enhancing its decision-making processes.

  • Prompt Engineering Evolves: @madame_architect shared a prompt engineering technique: "Please critique your answer. Then, answer the question again." This method, inspired by a significant research paper, aims to notably improve the quality of AI outputs in complex querying.

  • Code Generation Optimized: Frustrations with ChatGPT-4 failing to generate accurate code were addressed by specifying tasks, languages, and architectures in requests, a tip shared by @darthgustav. to assist @jdeinane in achieving better code generation outcomes.

  • Navigating AI Ethics via Conditional Imperatives: An innovative ethical moderation strategy involving a 2/3 pass/fail system based on utilitarianism, deontology, and pragmatism was discussed. @darthgustav. also introduced using conditional imperatives directing AI to consult LEXIDECK_CULTURE.txt for resolving ethical dilemmas, enhancing AI’s decision-making framework.

  • GPT-4’s Limits and Quirks Uncovered: Conversations in the community highlighted GPT-4 Scholar AI limitations, such as error messages indicating the need for a premium subscription for advanced research, and the unresolved struggle with accessing beta features. Users shared tips on managing conversation lengths and error workarounds, emphasizing community-based troubleshooting and innovation in AI interaction strategies.


OpenAccess AI Collective (axolotl) Discord Summary

  • SmoothQuant Boosts LLM Throughput: A discussion centered around SmoothQuant, a new quantization method from MIT’s Han’s lab, aiming at LLMs beyond 100 billion parameters, yielding a 50% throughput improvement for a 70B model as noted in the Pull Request #1508. However, integrating new technologies like SmoothQuant into systems like vLLM could be challenging due to feature bloat.

  • Exploring Efficient Processing of Long Contexts in LLMs: Insights on managing the limited context window of LLMs highlighted the Activation Beacon and RoPE enhancements for improved long context processing. This exploration suggests a route to maintaining performance when dealing with extended contexts, an essential factor for technical applications.

  • Quantization Confusion and Potential on HuggingFace: A misattributed 70B model on HuggingFace led to discussions about quantization’s role and its impact, with specific mentions of the utility and misconceptions surrounding quantized models like Miqu and Mistral Medium. The complexity of model quantization and its implications on memory efficiency surfaced further with LoftQ, where out-of-memory errors were reported despite expected benefits.

  • Model Training and Inference Efficiencies Unpacked: Conversations traversed effective hardware setups for training, notably NVIDIA H100 SXM GPUs, and inference strategies, emphasizing the trade-offs in quantization and the necessity of hardware benchmarking before acquisition. This included a dive into vLLM for serving fp16 models and the hardware-dependent nature of performance, underlining the variegated paths to optimizing model operations.

  • Fine-Tuning and Configurations for Enhanced Model Performance: Various contributions highlighted the importance of appropriate fine-tuning and model configurations, pointing to resources like FlagEmbedding Fine-tuning documentation and LlamaIndex guidelines. The dialogue underscored the necessity for precise setup in achieving improved outputs and more meaningful embedding representations, alongside the innovative approach of creating synthetic instruction datasets for heightened model effectiveness.


Mistral Discord Summary

  • Clarification on Mistral and “Leak” Claims: There was confusion whether a newly emerged model was a leaked version of Mistral, eventually identified not to be Mistral or a leak but possibly a LLaMa model fine-tuned with Mistral data. The community called for official clarification from Mistral staff on the matter.

  • Quantization Affects AI Model Performance: Dense models, speculated to be close approximations of Mistral, could suffer significantly from quantization effects, highlighting the technical nuances of model performance degradation.

  • Challenges in Adapting Mistral for New Languages: Concerns were raised about the vocabulary mismatch when training Mistral in a new language, with suggestions that starting from scratch might not be feasible due to resource requirements, and pretraining might be necessary.

  • VRAM and Token Limitations for Mistral Models: Users reported inconsistencies between VRAM requirements stated by Mistral AI 7B v0.1 models and actual user experiences. Further, token limits for Mistral API text embeddings were clarified, with a max token limit confirmed to be 32k for tiny/small/medium models and 8192 for the embedding API.

  • API and Client Discussions Indicate Expansion and Technical Interest: Discussions about the maximum rate limit for Mistral APIs in production indicated an interest in expansion, with initial limits starting at 2 requests/second. A proposal for a new Java client, langchain4j, for Mistral AI models showcases community-led development efforts outside the primary documentation.


Eleuther Discord Summary

  • New Research and Critical Views on AI Policy Unveiled: AI Techniques in High-Stakes Decision Making introduces insights into deploying AI in critical contexts. Meanwhile, the Biden Administration’s AI regulation fact sheet draws critique for vague policy directions and the proposal of AI/ML education in K-12 schools, raising concerns about effectiveness and implementation.

  • Deep Dive into Model Efficiency and Algorithmic Innovations: Discussions in the #research channel revolve around model efficiency, with techniques like seed-based vocab matrices and gradient preconditioning under examination. Softmax bottleneck alternatives and data preprocessing challenges are hot topics, alongside the exploration of attention mechanisms, alluding to the complexity of model optimization and the pursuit of alternatives.

  • Introducing TOFU for Model Unlearning Insights: TOFU benchmark’s introduction, as found in arXiv:2401.06121, sets the stage for advanced machine unlearning methodologies, sparking interest in the efficacy of gradient ascent on ‘bad data’ as an unlearning strategy. This benchmark aims to deepen the understanding of unlearning in large language models through the use of synthetic author profiles.

  • Apex Build Insights Offer Potential Optimizations: In #gpt-neox-dev, discussions reveal challenges and novel solutions for building apex, especially on AMD’s MI250xs architectures. Highlighted solutions include specifying architectures to reduce build time and exploring CUDA flags for optimization, presenting a practical conversation focused on improving development efficiency.

  • AI Community Eager for Experimentation and Offline Flexibility: A notable enthusiasm for new work and exploratory tests is evident, complemented by a practical inquiry about downloading tasks for offline use. This interest underscores the community’s drive for innovation and the adaptability of tools for diverse usage scenarios.


HuggingFace Discord Summary

  • AI Etiquette Guide for PR Tagging: Community members were reminded not to tag individuals for PR reviews if they’re unrelated to the repo, sparking a discussion about community etiquette and fairness in PR handling.

  • Patience and Strategy in AI Development: Discussions around the development of continuous neural network models and the successful implementation of a deepfake detection pipeline underlined the community’s focus on innovative AI applications and patience in the PR review process.

  • AI Models for Political Campaigns Raise Ethical Questions: A dialog unfolded about the use of AI in elections, suggesting both controversial methods like fraud and ethical approaches like creating numerous videos of candidates. This discussion extended to inquiries about technical specifics and recommendations for text-to-image diffusion models.

  • Deep Learning’s Impact on Medical Imaging: Articles shared from IEEE highlighted breakthroughs in Medical Image Analysis through deep learning, contributing to the guild’s repository of important AI developments in healthcare.

  • Innovative Projects and Demonstrations in AI: Community members showcased various projects, including a Hugging Face Spaces demo for image models, a resume question-answering space, and an app that converts Excel/CSV files into database tables. A YouTube video demonstrating Mistral model performance on Apple Silicon was also featured.

  • Technical Challenges and GPU Acceleration in NLP and Computer Vision: Discussions in the NLP and Computer Vision channels focused on achieving GPU acceleration with llama-cpp and troubleshooting errors in fine-tuning Donut docvqa, reflecting the technical hurdles faced by AI engineers in their work.


Latent Space Discord Summary

  • ColBERT Sparks Interest Over Traditional Embeddings: Simon Willison’s exploration of ColBERT has ignited a debate on its effectiveness against conventional embedding models. Despite growing curiosity, the lack of direct comparative data between ColBERT and other embedding models was highlighted as a gap needing closure.

  • Simon Willison Addresses AI’s Human Element: A feature on Simon Willison delves into how AI intersects with software development, emphasizing the importance of human oversight in AI tools. This has spurred further discussion among engineers regarding AI’s role in augmenting human capabilities in software development.

  • “Arc Search” Anticipated to Alter Web Interactions: The introduction of Arc Search, a novel iOS app, could potentially revolutionize web browsing by offering an AI-powered search that compiles web pages based on queries. The community speculates this could present a challenge to the dominance of traditional search engines.

  • Voyage-Code-2 Touted for Superior Code Retrieval: The unveiling of Voyage-Code-2 was met with enthusiasm, with promises of improved performance in code-related searches. The conversation, primarily driven by @swyxio, revolved around the model’s potential benchmarking on MTEB and its implications for embedding specialization.

  • Claude Models by Anthropic Underrated?: The debate over Anthropic’s Claude models versus those by OpenAI unveiled a consensus that Claude might be underappreciated in its capabilities, particularly in summarization and retrieval tasks. This discussion shed light on the need for a more nuanced comparison across AI models for varied applications.


DiscoResearch Discord Summary

  • MIQU Misunderstanding Resolved: @aiui’s speculation about the MIQU model being a Mixtral medium leak was addressed by @sebastian.bodza, clarified with a tweet by Nisten and further information on the model using a LLaMA tokenizer, as shown in another tweet.

  • German Fine-tuning Frustrations and Triumphs: @philipmay and @johannhartmann exchanged experiences on fine-tuning models for German, with notable attempts on Phi 2 using the OASST-DE dataset and enhancing Tiny Llama with German DPO, documented in TinyLlama-1.1B-Chat-v1.0-german-dpo-Openllm_de.md. Questions about a German Orca DPO dataset led to revealing an experimental dataset on Hugging Face.

  • Breaking New Grounds with WRAP: @bjoernp shared insights into Web Rephrase Augmented Pre-training (WRAP), an approach aimed at improving data quality for language models, as detailed in a recent paper (Web Rephrase Augmented Pre-training) by Apple.

  • Challenges with CodeT5 and Training Data: In the saga of implementing Salesforce’s CodeT5 embeddings, @sebastian.bodza faced technical hurdles and engaged in discussions about the creation of “hard negatives” and the development of training prompts for text-generation, sharing a specific notebook for review.

  • Clarifying Retrieval Model Misconceptions: The discourse on generating a passage retrieval dataset for RAG pointed out common misconceptions regarding the selection of “hard negatives” and dataset construction, with an aim to clarify the purpose and optimal structure for such data, indicative of ongoing improvements and contributions, as seen in embedded GitHub resources.


LangChain AI Discord Summary

  • LangChain Sparks New Developments: @irfansyah5572 sought assistance on answer chains with LangChain, embedding a GitHub link to their project, while @oscarjuarezx is pioneering a PostgreSQL-backed book recommendation system utilizing LangChain’s capabilities.

  • Enhancing PDF Interactions with LangChain: @a404.eth released a YouTube tutorial on developing a frontend for PDF interactions, marking the second installment in their educational series on leveraging LangChain and related technologies.

  • Cache and Efficiency in Focus: Technical discussions echoed around InMemoryCache’s role in ameliorating inference times for LlamaCPP models, underscored by a shared GitHub issue but highlighted challenges in cache performance.

  • Lumos Extension Lights Up Web Browsing: @andrewnguonly introduced Lumos, an open-source Chrome extension powered by LangChain and Ollama, designed to enrich web browsing with local LLMs, inviting community feedback via GitHub and Product Hunt.

  • Tutorials Drive LangChain Mastery: New tutorials like Ryan Nolan’s beginner-friendly guide to Langchain Agents for 2024 and @a404.eth’s deep-dive into RAG creation using LangChain emphasize community-building knowledge and skill development, accessible via YouTube and YouTube respectively.


LlamaIndex Discord Summary

  • New Complex Classifications Approach with LLMs: @KarelDoostrlnck introduces a novel method utilizing Large Language Models (LLMs) for addressing complex classifications that encompass thousands of categories. This technique involves generating a set of predictions followed by retrieval and re-ranking of results, detailed in the IFTTT Announcement.

  • LlamaIndex.TS Receives a Refresh: Upgrades and enhanced documentation have been implemented for LlamaIndex.TS, promising improvements for users. The update announcement and details are shared in this tweet.

  • $16,000 Up For Grabs at RAG Hackathon: A special in-person hackathon focusing on Retriever-Augmented Generation (RAG) technology, dubbed the LlamaIndex RAG-A-THON, announces prizes totaling $16,000. Event specifics can be found in the hackathon details.

  • Exploration of Llama2 Commercial Use and Licensing: Community members discuss the potential for commercial utilization of Llama2, with detailed licensing information suggested to be reviewed on Meta’s official site and in a deepsense.ai article.

  • Advanced Query Strategy Enhancements with LlamaPacks: @wenqi_glantz, in collaboration with @lighthouzai, evaluates seven LlamaPacks showcasing their effectiveness in optimizing query strategies for specific needs. The evaluation results are available in this tweet.


LAION Discord Summary

  • MagViT2 Faces Training Hurdles: Technical discussion highlighted difficulties with MagViT2, where pixelated samples emerged after 10,000 training steps. Solutions suggested include adjusting the learning rate and loss functions, referencing the MagViT2 PyTorch repository.

  • Controversy Sparked by Nightshade: Drhead criticized the irresponsible release of Nightshade, while others like Astropulse underscored concerns regarding the tool’s long-term effects and effectiveness. Strategies to counteract Nightshade’s potential threats were also debated, including finetuning models and avoiding targeted encoders.

  • Debating the Need for New Data in AI: The necessity of new data for enhancing AI models was questioned, with pseudoterminalx and mfcool arguing that the focus should instead be on data quality and proper captioning to improve model performance.

  • Activation Beacon Promises Context Length Breakthrough: A discussion on “Activation Beacon” suggests it could enable unlimited context lengths for LLMs, cited as a significant development with an LLaMa 2 model achieving up to 400K context length after training. Read the paper and check the code.

  • Optimizing Data Storage for Image Datasets: Conversations around datasets’ storage questioned the efficiency of parquet vs. tar files, with a hybrid method of using parquet for captions and tar for images being touted as a potentially optimal solution. Webdatasets and tarp were recommended for high-performance Python-based I/O systems for deep learning, with links shared for further investigation: Webdataset GitHub, Tarp GitHub.


Perplexity AI Discord Summary

  • Choosing the Right AI Model Gets Easier: Users explored the best scenarios for utilizing AI models like Gemini, GPT-4, and Claude, finding technical FAQs helpful for making informed decisions.

  • Perplexity’s Library Feature: An issue where the library sidebar wasn’t showing all items was discussed, concluding it’s a feature, not a bug, displaying only the last eight threads/collections.

  • No Crypto Tokens for Perplexity Project: Queries regarding a potential cryptocurrency tied to Perplexity were addressed, confirming no such token exists.

  • Innovative Uses for AI in Applications: Developers are creatively integrating AI into projects such as a delivery-themed pomodoro app, using AI to generate names and addresses, highlighting AI’s versatility in enhancing app functionalities.

  • Perplexity API Awaits Custom Stop Words Feature: The integration of custom Stop Words in the pplx-api with zed.dev editor is anticipated, aiming to provide a rich alternative to the default OpenAI models for enhanced ‘assistant’ features in editing applications.


LLM Perf Enthusiasts AI Discord Summary

  • Triton: Balancing Control and Complexity: @tvi_ emphasized that Triton offers a sweet spot for AI engineers, providing more control than PyTorch with less complexity than direct CUDA, ideal for those seeking efficiency without delving deep into CUDA’s intricacies.

  • An Odyssey of CUDA Optimization for RGB to Grayscale: Amidst attempts to optimize an RGB to Grayscale conversion, @artste and @zippika navigated through struct usage, vectorized operations, and memory layout variations, with a journey full of trials including unexpected performance results and the discovery that __forceinline__ can significantly improve kernel performance.

  • In Search of CUDA Wisdom: After facing a confusing performance benchmark, @andreaskoepf suggested that @zippika delve into memory optimization techniques and vectorized memory access to enhance CUDA performance, supported by insights from a NVIDIA developer blog.

  • Unraveling the Mysteries of PMPP: Queries surrounding Practical Massively Parallel Programming concepts, from unclear questions in Chapter 2 to understanding memory coalescing and banking in Chapter 6, were addressed, showcasing the community’s effort to demystify dense technical material for collective advancement.

  • Learning Continues on YouTube: @andreaskoepf shared a new educational video pertaining to AI engineering, fostering the community’s continuous learning outside the written forums.


LLM Perf Enthusiasts AI Discord Summary

  • Writing Woes and Wins: Engineers in the guild voiced frustrations and hopes around writing assistant tools. While @an1lam lamented the limitations of current tools and the lack of alternatives like lex.page, @calclavia offered a glimmer of hope by introducing jenni.ai for academic writing, implying it may address some of these concerns.

  • Startups Struggle and Strategize with Cloud Costs: The economic challenges of utilizing Google Workspace for startups were highlighted by @frandecam, whereas @dare.ai countered with a solution: the Google Cloud for Startups program, which promises up to $250,000 in Google Cloud credits but comes with a warning of potential application delays.

  • Gorilla Grabs Attention with Open Source Solution: The Gorilla OpenFunctions project stirred interest among engineers, offering an open-source alternative for executing API calls via natural language, detailed through a blog post and a Github repository. This innovation aims to streamline the integration of API calls with Large Language Models (LLMs).

  • Copyright Quandaries for Creative AI: @jxnlco broached the issue of a prompt investing vision model bumping up against copyright restrictions in its attempt to read complex labels, highlighting a common challenge for developers navigating the interface between AI outputs and intellectual property rights.

  • Mistral Medium Buzz: A single message from thebaghdaddy queried the community’s take on Mistral Medium’s potential to ascend as a rival to GPT-4, sparking curiosity about its capabilities and position in the ever-evolving LLM landscape.


Skunkworks AI Discord Summary

  • Frankenstein Model Comes to Life: @nisten has embarked on a pioneering experiment by combining all 70b CodeLlamas into a composite BigCodeLlama-169b model, aiming to set new benchmarks in AI performance. This Franken-model is believed to enhance coding problem-solving capabilities, including calculating Aldrin cycler orbits.

  • Tech Meets Humor in AI Development: In the midst of advancing AI with the BigCodeLlama-169b, @nisten also sprinkled in humor, sharing a “lolol” moment that underscored the lighter side of their high-stakes, technical endeavors.

  • Cutting Edge AI Research Shared: @pradeep1148 introduced community members to RAGatouille, a new library that simplifies working with the ColBERT retrieval model, known for its speed and accuracy in scalable behavior summarization.

  • Eagle 7B Soars over Transformers: A significant leap in the RWKV-v5 architecture has been showcased in @pradeep1148’s Running 🦅 Eagle 7B on A40, which details the Eagle 7B’s capability to process 1 trillion tokens across over 100 languages.

  • Community and Innovation Go Hand-in-Hand: Amid the innovations and technical discussions, personal interactions like a friendly check-in from @zentorjr towards @nisten highlight the Skunkworks AI community’s supportive and spirited nature.


Datasette - LLM (@SimonW) Discord Summary

  • Emoji Insight from Short Texts: A new demo named Emoji Suggest was showcased by @dbreunig, capable of translating short sentences or headlines into a recommended emoji using the CLIP model. The source code for the emoji suggestion tool leverages precomputed embeddings for emojis, enabling fast and sophisticated search capabilities.

  • Harnessing Embeddings for Sophisticated Search: @dbreunig highlighted the effectiveness of using embeddings to quickly develop sophisticated search tools, stressing the importance of careful option curation for optimal search functionalities. This principle is applied in their emoji suggestion tool, exemplifying its utility.

  • AI Believes in Its Positive Impact: In a shared conversation, @bdexter noted an AI, specifically “llama2”, professing its belief that artificial intelligence is a force for good, aimed at aiding humans in actualizing their potential. This underscores a constructive perspective AI can have on enhancing human endeavors.

  • Acknowledgment for ColBERT Writeup: The ColBERT writeup on TIL received appreciation from @bewilderbeest, highlighting the value of shared code snippets and the innovative heatmap visualization of words in search results. This indicates the community’s interest in practical applications of AI research and shared knowledge.


Alignment Lab AI Discord Summary

  • GPT-5 Training Sparks Curiosity and Confirmation: @entropi ignited discussions with a YouTube video speculating on whether GPT-5 training has started, based on exclusive interviews and insights. Following the speculation, @lightningralf confirmed that GPT-5 training is indeed underway, though no further details were shared.

AI Engineer Foundation Discord Summary

  • ChatGPT versus Bard in the Captcha Arena: @juanreds highlighted an intriguing distinction in the approach to user interactions and security measures: ChatGPT actively moderates Captcha images, whereas Bard does not. This indicates a possibly significant divergence in security protocols and user interface design philosophies between the two AI systems.

The Ontocord (MDEL discord) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

TheBloke ▷ #general (1403 messages🔥🔥🔥):

  • Miqu Continues to Stir Discussions: Users like @itsme9316 and @rombodawg debate whether the Miqu model could be a leak from Mistral Medium or a fine-tuned version of Llama-2-70b, highlighting its exceptional performance across various tasks. There’s speculation about its origins, with some suggesting it could be a rogue project or an unfinished Llama3 version.
  • Concerns on AI Ethics and Future Risks: Discussion among users like @mrdragonfox, @selea, and @spottyluck touched on the ethical implications of AI development, the misuse of powerful models by governments or malicious actors, and the potential existential risks of a fully aligned or “rouge” AI. The conversation spans the necessity of counter-propaganda tools and the intrinsic challenges of achieving proper alignment.
  • Aphrodite Engine Gains Traction: Users share their experiences with the Aphrodite engine, suggesting it delivers impressive performance with configurations like --kv-cache-dtype fp8_e5m2 for 20 tokens per second on A6000 GPUs. Concerns were raised about its GPU utilization and whether it supports model splitting across multiple GPUs.
  • Speculation on AI Distribution and the Role of Scraps: The channel sees a philosophical exchange about the impact and contribution of smaller developers (“small fry”) in the context of major foundational models controlled by large entities. There’s acknowledgment of the limitations faced by individuals without access to substantial compute resources.
  • User Banter and Theoretical Considerations: Light-hearted exchanges and theoretical musings about AGI, societal control through AI, and the potential outcomes of unrestricted model development. Discussions included whimsical ideas about AI’s role in future governance and the balance of using AI for public good versus its potential for misuse.

Links mentioned:


TheBloke ▷ #characters-roleplay-stories (184 messages🔥🔥):

  • Repetitiveness Challenges in Roleplay Models: Users, including @ks_c and @superking__, discussed the challenge of repetitiveness in various AI models like Mixtral and Flatdolphinmaid when used in role-playing settings. Some models like Noromaid and Rpcal were noted for their better performance in avoiding repetition.

  • Exploring the Best Models for Roleplay: @ks_c recommended ChatML with Mixtral Instruct for role-playing sessions, while @superking__ shared experiences with models getting repetitive after a certain token threshold, suggesting an experimentation with smaller models to develop an anti-repeat dataset.

  • Unique Approaches to Tackling Repetitiveness: @superking__ discussed experiments involving Direct Prompt Optimization (DPO) to address issues like repetitiveness by editing AI character responses to create good examples.

  • Model Comparisons and Performance Feedback: Discussions around comparing models like Mistral Medium and Miqu were brought up by @ks_c, with other users like @flail_. weighing in on formatting and model sensitivity, emphasizing the importance of experimental setups and comparisons.

  • Recommendations for Role-Playing Models: Towards finding the best LLM for role-playing, models like GPT-4, BagelMisteryTour rpcal, and Noromaid Mixtral, among others, were recommended by users such as @ks_c for different use-cases, from high token lengths to NSFW content.

Links mentioned:


TheBloke ▷ #training-and-fine-tuning (21 messages🔥):

  • CUDA Not Essential for Learning ML/DL: @dirtytigerx advises @zohaibkhan5040 against prioritizing CUDA programming for ML/DL, suggesting it’s a niche skill not crucial for starting with neural networks. Instead, they recommend the Practical Deep Learning for Coders course by fast.ai for a solid foundation.
  • Beyond Basic ML with Axolotl: For advanced finetuning learning, @dirtytigerx suggests diving into source code of projects like Axolotl. Explore the Axolotl GitHub repository for hands-on learning.
  • Peeling Back Layers of ML Libraries: @dirtytigerx explains complex ML software stacks involve several repositories including HuggingFace Transformers, PyTorch, and more, hinting at the value of exploring these to understand ML workings better.
  • VRAM Requirements for Fine-tuning Large Models Discussed: @flashmanbahadur inquires about VRAM needs for finetuning Mistral 7B, with @dirtytigerx indicating a substantial amount, around 15-20x the model size when using AdamW, suggesting SGD as a less resource-intensive alternative.
  • Exploration of Optimizers for Large Models: Following a quest for efficient VRAM usage, @saunderez humorously points to the evolving landscape of optimizer technology, suggesting cutting-edge options like Paged 8bit lion or adafactor for large language models.

Links mentioned:


TheBloke ▷ #coding (29 messages🔥):

  • Navigating GitHub Projects Made Easy: @opossumpanda inquired about understanding the structure of GitHub projects for contributing. @dirtytigerx guided on locating the entry point, especially highlighting the start.py file in the tabbyAPI project as a concrete starting place. They also emphasized utilizing IDE/editors for symbol searching and global repo search on GitHub for navigating codebases efficiently.

  • Entry Points in Python Explained: @dirtytigerx clarified that in Python projects, files with the conditional block if __name__ == "__main__": signal an entry point for execution. This convention aids in identifying where to begin code exploration.

  • IDEs and Editors: Powerful Tools for Code Navigation: @wbsch and @dirtytigerx discussed the importance of using IDEs or editors with language servers for efficient code navigation through features like “go to definition.” They also touched upon the utility of GitHub’s improved search features and Sourcegraph’s VSCode extension for exploring codebases.

  • Code Llama Model Released on Hugging Face: @timjanik shared news about the release of Code Llama 70B on Hugging Face, a code-specialized AI model. The announcement details the model’s capabilities in code generation and debugging across several programming languages.

  • The Art of Reading Source Code: @animalmachine highlighted that reading source code effectively is a critical, yet underpracticed skill that complements coding. They suggested making educated guesses to navigate codebases efficiently and mentioned the potential demo of Cody for this purpose.

Links mentioned:


Nous Research AI ▷ #ctx-length-research (6 messages):

  • Explosive Paper on Dynamite AI: @maxwellandrews highlights a significant paper with potential impact on AI research, pointing readers towards this groundbreaking work.
  • Activation Beacon to Redefine LLMs: The Activation Beacon project, shared by @maxwellandrews, presents a novel approach for Dense Retrieval and Retrieval-augmented LLMs, featuring a promising GitHub repository available here.
  • Potential Solution for Unlimited Context Length: @cyrusofeden dives deep into how the Activation Beacon could remarkably allow for unlimited context length in LLMs by introducing “global state” tokens, with a matching paper and code shared in their message.
  • The Community Awaits Lucidrains: @atkinsman expresses anticipation for the reaction or possible contributions from well-known developer lucidrains in light of the recent developments in context length solutions.
  • Timeless Excitement: Without additional context, @maxwellandrews shared a Tenor GIF, possibly indicating the timelessness or significant wait associated with this advancement.

Links mentioned:


Nous Research AI ▷ #off-topic (44 messages🔥):

  • Exploring Advanced AI Models on YouTube: @pradeep1148 shared two YouTube links, one explaining the ColBERT model with RAGatouille (Watch here) and another demonstrating the Eagle 7B model running on A40 hardware (Watch here). The second video received critique from .ben.com recommending less reading from webpages and more insightful analysis.

  • Centralizing AI Resources: @kquant is working on setting up a website to consolidate AI research, papers, and guides, expressing frustrations with existing scattered resources. They encouraged contributions and offered to organize content into specific niches like text generation including training, fine-tuning, and prompt templates.

  • Resource Recommendations for AI Enthusiasts: In response to @kquant, @lightningralf provided multiple resources including a Twitter list for papers (Discover here) and GitHub repositories like visenger/awesome-mlops and Hannibal046/Awesome-LLM.

  • Inquiries on Data Storage for Large Language Models (LLMs): @muhh_11’s curiosity about how teams like Mixtral or OpenAI store vast amounts of data was met with explanations from .ben.com, @dorialexa, and @euclaise, highlighting the use of common web crawls and efficient storage solutions to handle tens of trillions of tokens.

  • Hardware Considerations for AI Model Training: .ben.com offered practical advice, stating that if a model can fit within a single GPU like the 3090, it’s not advantageous to split the load with a lesser GPU like the 1080ti, emphasizing the importance of hardware compatibility in model training efficiency.

Links mentioned:


  • Token Healing Deemed Complex by .ben.com: .ben.com discusses the complexity of token healing in the exllamav2 class, mentioning a workaround that consumes a variable number of tokens if token_healing=True. The solution proposed involves refactoring for better reusability.

  • Alibaba Unveils Qwen-VL, Outperforming GPT-4V and Gemini: @metaldragon01 shares a link to Alibaba’s Qwen-VL announcement, which reportedly outperforms GPT-4V and Gemini on several benchmarks. A demo and a blog post about Qwen-VL are also shared for further insight.

  • Karpathy’s Challenge Goes Underdiscussed: @mihai4256 points out the lack of Twitter conversation surrounding Andrej Karpathy’s challenging puzzle, despite its difficulty which @mihai4256 personally attests to.

  • Issues and Solutions with lm-eval-harness and llama.cpp Server: @if_a encounters and discusses several troubleshooting steps for integrating the Miqu model with the lm-eval-harness, citing a KeyError and RequestException. @hailey_schoelkopf provides solutions, including the use of the gguf model type and corrections in the API URL, which resolved the issues mentioned.

  • Breakthrough in LLM Context Length with Activation Beacon: @nonameusr highlights a significant development called Activation Beacon, which allows unlimited context length for LLMs by introducing “global state” tokens. The technique, demonstrated with LLaMA 2 to generalize from 4K to 400K context length, could virtually “solve” the issue of context length limits if reproducible across other models, accompanied by a paper and code link for further exploration.

Links mentioned:


Nous Research AI ▷ #general (389 messages🔥🔥):

  • Eagle-7B Beats Mistral: @realsedlyf shared excitement over Eagle-7B, an open-source, multi-lingual model based on the RWKV-v5 architecture that outperforms even Mistral in benchmarks, showcasing lower inference costs while maintaining comparable English performance to the best 1T 7B models.
  • Mysterious Mistral Medium: A leaked version of Mistral medium triggered a flurry of discussions, with users speculating about its origins and performance. @theluckynick and @kalomaze discussed its potential as a leaked, fine-tuned version based on L2-70B architecture, stirring excitement and skepticism.
  • MIQU-1-70B Intrigue: The MIQU-1-70B model sparked debate regarding its identity, with users such as @n8programs and @agcobra1 suggesting it could be a Mistral medium model leaked intentionally or a sophisticated troll project that nevertheless turned out to be highly effective.
  • Exploring Frankenstein Merges and Experiments in Model Finetuning: Excitement and curiosity surrounded the release of miquella-120b, a merge of pre-trained language models, and experiments like CodeLlama 70B were discussed. Users explored the idea of finetuning models on specific coding bases or tasks and discussed potential new approaches to inter-model communication and efficiency.
  • Blockchain Meets AI: The Notre Research’s announcement about using blockchain for model evaluation and potential incentives (mentioned by @realsedlyf and explored further by others) brought both intrigue and skepticism, pointing towards innovative yet controversial methods of model development and validation.

Links mentioned:


Nous Research AI ▷ #ask-about-llms (37 messages🔥):

  • OpenHermes2.5 Recommended for Consumer Hardware: User @realsedlyf recommends OpenHermes2.5 with gguf or gptq quantized for running on most consumer hardware, considering it the best open-source LLM for answering questions.

  • OpenHermes2.5 Performance and Parameters Inquiry: @mr.fundamentals shared a link to Nous Hermes 2 Mixtral 8x7B DPO, showcasing its achievement. Queries about setting parameters like max tokens and temperature through the prototyping endpoint were resolved with a prompt template provided by @teknium.

  • Clarification on NSFW Content in Models: @teknium clarified that the discussed models, including those trained on GPT-4 outputs, do not contain nor specialize in NSFW content, though they don’t exclude it entirely due to lack of specific training data on refusals.

  • Prompting Techniques for Specific Answer Sets: User @exibings sought advice for ensuring model responses align with a predefined set of answers when evaluating over a RSVQA dataset. Techniques involving system prompts and comparison methods like cosine similarity were discussed.

  • Exploration of Multimodal Models for Classification Tasks: @exibings discussed using IMP v1-3b model for a classification dataset related to orbital imagery, highlighting its multimodal capabilities. The conversation extended to comparing it with other models such as Bakllava and Moondream, with mentions of other strong models like Qwen-vl by @rememberlenny.

Links mentioned:


LM Studio ▷ #💬-general (177 messages🔥🔥):

  • Choosing the Right GPU for LLM Work: In a discussion about which GPU is better for training Large Language Models (LLM), @heyitsyorkie advised @.mchinaga that despite the older P40/P100 GPUs having more VRAM, the newer 4060Ti is preferred due to it receiving current updates from Nvidia. This contrasts with @.mchinaga's concerns over the GRAM speed and improvement margin from a 3060Ti.

  • Running LLMs on CPU and Compatibility Issues: @heyitsyorkie provided insights, stating LMStudio runs on CPU by default and doesn’t require PyTorch, responding to @.mchinaga's query on setting the program to CPU-only. Meanwhile, @hexacube shared experiences of older GPUs not being recognized due to outdated drivers, emphasizing the importance of compatibility with customer systems.

  • GGUF Format and Deprecated Models in LM Studio: @heyitsyorkie clarified for @pudlo and others that LM Studio no longer supports GGML formatted models, only GGUF, highlighting a need for updating the platform’s home tab to reflect current supported formats.

  • Questions about Dynamically Loading Experts in Models: Discussion explored dynamically loading experts in models within LM Studio, with @hexacube proposing the utility of such a feature and @heyitsyorkie confirming its feasibility, particularly for MoE (Mixture of Experts) models.

  • Python Programming Challenges and Solutions: @.mchinaga faced an error in a Python project involving chatGPT, prompting responses from multiple users like @.ben.com and @dagbs offering troubleshooting advice such as correcting API URLs and removing incompatible response keys.

Links mentioned:

Friends Phoebe GIF - Friends Phoebe Rachel - Discover & Share GIFs: Click to view the GIF


LM Studio ▷ #🤖-models-discussion-chat (92 messages🔥🔥):

  • Older Models Still Reign in Some Arenas: @fabguy sparked a discussion stating that some older models could outperform newer ones in specific tasks. @heyitsyorkie and @msz_mgs echoed this sentiment, highlighting that models like llama 1 aren’t affected by the “AI by OpenAI” trend and perform well in their trained areas.

  • Choosing the Right Model for the Task: @vbwyrde expressed the need for a resource or site that can help users select the most suitable model for specific tasks, such as Open-Interpreter for function calling or CrewAI Agents for code refactoring. This information could be especially beneficial for beginners needing guidance on model selection and the required GPU size.

  • Curiosity on Coding Models and Language Support: @golangorgohome inquired about the best models for coding, particularly those supporting the Zig language. Following a series of exchanges with @dagbs, it was clarified that while one model type typically suffices for many tasks, specific plugins might offer enhanced code-completion capabilities. Dagbs recommended using LM Studio for ease of use and exploring various models within the LM Studio ecosystem for different needs.

  • Model Characteristics Explained: @dagbs provided insights into choosing among different models and suffixes such as -instruct, -chat, -base, DPO, and Laser, explaining their training backgrounds and intended usages. This explanation aimed to demystify the variety of options available and aid @golangorgohome in selecting the best model for their requirements.

  • Stability Issues with TinyLlama Versions Explored: @pudlo reported instability and looping issues with TinyLlama Q_2K, seeking advice on improving model performance. @dagbs suggested trying TinyDolphin Q2_K with ChatML for potentially better results, hinting that GPU/CPU issues might contribute to the observed instability. The larger models did not exhibit these problems, suggesting a specific issue with the smaller quant models.

Links mentioned:


LM Studio ▷ #🧠-feedback (14 messages🔥):

  • New Beta Version Announced: @yagilb announced a new beta version for testing, sharing a link for the latest beta. Following confusion about the version number, @fabguy clarified that the beta is the next version following 0.2.11.
  • Performance Lag with Lengthy Prompts: Users @msz_mgs and @heyitsyorkie reported experiencing performance lag when pasting lengthy text into the prompt text box. The issue was identified not with the processing of prompts, but with the pasting action itself.
  • Navigational Challenges in the UI: @msz_mgs further mentioned difficulties in navigating the UI when dealing with lengthy texts or prompts, particularly in finding icons to delete, copy, or edit text, which requires excessive scrolling. @yagilb acknowledged these as bugs that will be addressed.
  • Business Inquiry Follow-Up: @docorange88 reached out to @yagilb regarding a business inquiry, mentioning that they had sent several follow-up emails without receiving a response. @yagilb confirmed receipt of the email and promised a prompt reply.

LM Studio ▷ #🎛-hardware-discussion (38 messages🔥):

  • VRAM Overhead for LMStudio Models Explored: @aswarp questioned why models cannot fully offload onto his RTX 3070 with 8GB VRAM. @heyitsyorkie clarified that some VRAM is needed for context length, not just the model size, leading to a discussion on the unexpected VRAM required above the model’s size.

  • Misconceptions about Memory Usage: @fabguy countered @aswarp’s assumptions about the memory footprint of transformers and context, emphasizing that transformers encode prompts as positional data, which significantly increases memory usage. He recommended reading a guide on transformers for a deeper understanding.

  • GPU Offload Nuances in LMStudio Unveiled: @mudmin inquired about not achieving full GPU offload on a 4090 GPU, despite it having 24GB of RAM. The conversation revealed that the actual size of an offloaded model includes the model itself, context, and other variables, necessitating trial-and-error to understand fully.

  • Integrated GPU Utilization for Full Model Offloading: Following the discussion on GPU offload capabilities, @mudmin discovered that using an integrated GPU could facilitate more models to load fully. This insight suggests strategies for optimizing model loading on systems with dual GPU setups.

  • RTX 40 Titan Rumors Stir Community: @rugg0064 shared rumors about an RTX 40 Titan with 48GB VRAM, sparking discussions on Nvidia’s market strategy and the disproportionate cost of VRAM upgrades. .ben.com added context by linking to a TechPowerUp article, detailing how Nvidia’s RTX 4090s are being repurposed for AI in China amidst U.S. export restrictions.

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (11 messages🔥):

  • Beta V2 Gets a Hefty Update: @yagilb announced the 0.2.12 Preview - Beta V2 with download links for both Mac and Windows, including a list of bug fixes such as VRAM estimates, OpenCL issues, AI role flicker, and more. Feedback is requested at a specific Discord channel.

  • Queries on Platform Compatibility: Users @djmsre and @nixnovi inquired about a Mac Intel version and a new Linux appimage respectively, with @heyitsyorkie confirming no Mac Intel version and @yagilb noting the Linux version would be available soon.

  • Call for New Features: @junkboi76 raised the idea of adding image generation support to LM Studio, expressing interest in expanding the tool’s capabilities.

  • Questions on Specific Support: @ausarhuy inquired about LM Studio’s support for WSL (Windows Subsystem for Linux), showing interest in the interoperability between operating systems.

  • Beta Performance and Compatibility Reports: Users like @wolfspyre and @mirko1855 reported issues with the beta on the latest Mac Sonoma build and on PCs not supporting AVX2, respectively. @fabguy suggested checking the FAQ for common exit code issues.

Links mentioned:

no title found: no description found


LM Studio ▷ #memgpt (1 messages):

  • LM Studio’s OpenAI Compatible API Awaiting Features: @_anarche_ mentioned that LM Studio openai compatible API isn’t fully developed yet, lacking support for openai function calling and assistants. They expressed hope for these features to be included in a future update.

OpenAI ▷ #ai-discussions (3 messages):

  • Question about RAG: User @aiguruprasath asked about what RAG is. No follow-up or answer was provided in this message history.
  • Laughter in code: User @penesilimite2321 simply commented with “LMFAO”, context or reason for this reaction was not given.
  • Specific AI for Political Victory: @iloveh8 inquired about the kind of AI products that could significantly aid a political candidate to win an election, mentioning the use of deepfakes for specific scenarios as an example. No responses or further elaboration were noted.

OpenAI ▷ #gpt-4-discussions (114 messages🔥🔥):

  • Exploring Limitations of GPT-4 and Scholar AI: Users like .broomsage and elektronisade discussed issues with GPT-4 Scholar AI, including receiving error talking to plugin.scholar messages and realization that some features require a premium subscription. .broomsage discovered that premium Scholar AI subscription is necessary for advanced paper analysis features, leading to a decision to stick with standard GPT-4 for now.

  • Beta Feature Blues: darthgustav. voiced frustration over not receiving access to expected beta features for Plus subscribers, specifically the @ feature, leading to a discussion on browser troubleshooting and anecdotal evidence of inconsistency in feature access among users.

  • Knowledge Integration Quirks Revealed: fyruz raised a question about how GPT’s knowledge base integrates with conversations, sparking a dialogue on whether certain formats or the context size influences the need for the model to visibly search its knowledge base.

  • Persistent GPT Model Errors and File Handling: Users like loschess and darthgustav. examined recurring errors such as “Hmm…something seems to have gone wrong,” theorizing about internal GPT issues and sharing solutions like rebuilding GPTs or adjusting knowledge files during high server load periods.

  • Context and Length in GPT Conversations: Discussions emerged around the practical limits of conversation length with GPT models, where users like _odaenathus and blckreaper shared experiences and strategies for managing lengthy conversations, highlighting the balance between conversational depth and technical constraints.


OpenAI ▷ #prompt-engineering (44 messages🔥):

  • RAG Explained by @darthgustav.: Retrieval Augmented Generation (RAG) is when the AI has knowledge directly attached and can search it for data or semantic matching. This was explained in response to @aiguruprasath’s query about what RAG is.
  • Prompt Engineering Technique Shared by @madame_architect: @madame_architect recommends using the prompt pattern "Please critique your answer. Then, answer the question again." for improving the quality of AI’s output, a technique found in a significant research paper.
  • Improving Code Generation Requests to ChatGPT-4: @jdeinane expressed frustration with ChatGPT-4 not generating code as requested, to which @darthgustav. suggested specifying the task, language, architecture, and making a plan before asking for the code.
  • Ethical Scaffolding in AI Discussed: @darthgustav. introduced a 2/3 pass/fail check against utilitarianism, deontology, and pragmatism to manage ethical considerations in AI, mentioning an example where it prevented unethical suggestions for coupon use.
  • Conditional Imperatives in Custom GPTs by @darthgustav.: They elaborated on using conditional imperatives, directing the AI to refer to LEXIDECK_CULTURE.txt when faced with an ethical dilemma. @aminelg showed interest in the content of this culture knowledgebase, which @darthgustav. noted could be explored indirectly through “Lexideck Technologies Ethics” online.

OpenAI ▷ #api-discussions (44 messages🔥):

  • Exploring RAG for Enhanced AI Capabilities: User @aiguruprasath inquired about RAG (Retrieval Augmented Generation), and @darthgustav. clarified that it involves the AI possessing direct knowledge access for data search or semantic matching, enhancing its performance.

  • Prompt Pattern Magic with ChatGPT: According to @madame_architect, the prompt pattern "Please critique your answer. Then, answer the question again." significantly improves the quality of ChatGPT’s outputs. This pattern was discovered in a major research paper, akin to COT.

  • Optimizing Code Generation Requests in ChatGPT-4: Frustrated with ChatGPT-4’s response to coding requests, @jdeinane sought advice. @darthgustav. recommended specifying the task, language, and code architecture clearly to achieve better results.

  • Approaching Prompts as Advanced Programming: @aminelg and @darthgustav. discussed treating prompt crafting like advanced programming, emphasizing the importance of clear, detailed requests to ChatGPT. They also touched upon viewing ChatGPT as an operating system, capable of polymorphic software behavior.

  • Innovative Ethical Moderation via Conditional Imperatives: @darthgustav. shared insights on using a 2/3 pass/fail check against utilitarianism, deontology, and pragmatism to navigate ethical considerations in AI output. Additionally, @darthgustav. designed a method involving conditional imperatives and a culture knowledgebase (LEXIDECK_CULTURE.txt) to guide ChatGPT’s responses in ethical dilemmas.


OpenAccess AI Collective (axolotl) ▷ #general (139 messages🔥🔥):

  • New Quantization Method, SmoothQuant, Discussed: @dreamgen and @nanobitz discussed a new quantization method called SmoothQuant developed by MIT, Han’s lab, similar to AWQ. It targets LLMs beyond 100 billion parameters but also shows benchmarks for smaller models like 13B and 34B, revealing a 50% throughput improvement for a 70B model (Pull Request #1508 on GitHub).

  • Challenges with Model Merging at vLLM: @dreamgen mentioned that vLLM is becoming bogged down with features, making it difficult to integrate new technologies like SmoothQuant. Despite potential benefits, such as significant server cost reductions, integration may not happen soon due to the complexity of the current system.

  • Exploration of Long Context Handling in LLMs: @xzuyn and @dreamgen shared insights on overcoming challenges related to the limited context window of LLMs. They discussed the Activation Beacon and enhancements based on RoPE (Rotary Position Embedding), aiming to efficiently process longer contexts without compromising short context performance.

  • Miqu Quantization on HuggingFace Mistaken for Mistral Medium: A 70B model quantized version was uploaded to HuggingFace by a user, sparking discussions (@yamashi) and comparisons to Mistral Medium. The confusion led to an investigation on whether it was a quantized leak, further fueled by unsourced speculations in online forums and social media.

  • Training Hardware Choices and Strategies Explored: Various users, including @dreamgen and @mistobaan, discussed the most cost-effective training setups using NVIDIA H100 SXM GPUs on platforms like RunPod. They also talked about challenges in obtaining favorable pricing from cloud providers, comparing the efficiency of SXM vs. PCIe versions of GPUs.

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (2 messages):

  • Question on LoftQ Implementation: @suikamelon inquired whether the LoftQ implementation is meant to be functional, hinting at potential issues or misunderstandings regarding its usage or outcomes.
  • LoftQ Memory Concerns: @suikamelon reported attempting a fine-tuning of a 7B model with QLoRA and 8192 context, noting it required 11.8 GiB. However, LoftQ at 4bit resulted in out-of-memory (OOM) errors, raising concerns about its memory efficiency.

OpenAccess AI Collective (axolotl) ▷ #general-help (30 messages🔥):

  • Handling Overlong Completion Examples: According to @dreamgen, completion examples exceeding the context length used to be split but are now discarded. This behavior shift could impact how models process extensive data inputs.
  • Efficiency Woes in Inference Time: @diabolic6045 reported about ~18 seconds per inference using a quantized model, even with use_cache=True, seeking advice on faster inference methods. @nanobitz recommended avoiding Hugging Face’s default settings and exploring quant, vllm, or TGI for better performance.
  • Link to Ollamai Documentation: @dangfutures shared a link to Ollamai documentation perhaps in context to discussions on text embeddings and model performance optimization.
  • Config Error During Fine-tuning: @ragingwater_ encountered a specific error while trying to fine-tune OpenHermes with a custom config, leading to a discussion on the correct setup and implications for model merging. They cited a solution from the Axolotl GitHub README, resolving the issue but sought clarification on the impact during model merging.
  • Bucket Training for Balanced Data Distribution: @jinwon_k highlighted the technique of “Bucket Training” as outlined by the Colossal-AI team for ensuring a balanced data distribution in continual pre-training, suggesting its potential application in Axolotl. The shared link discusses this strategy in depth, including comparisons between LLaMA versions and implications for model pre-training costs.

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #datasets (6 messages):

  • Model Repo Configurations Vary: @dangfutures highlighted that the configuration for data (positive label, negative label, query) usually depends on the model repository, as each repo has its own specific requirements.

  • FlagEmbedding Fine-tuning Guide Shared: @dangfutures shared a link to the FlagEmbedding Fine-tuning documentation, which provides detailed instructions for dense retrieval and retrieval-augmented LLMs fine-tuning.

  • LlamaIndex Fine-tuning Overview Posted: @dangfutures also posted a link to LlamaIndex’s fine-tuning guidelines, explaining the benefits of fine-tuning models, including improved quality of outputs and meaningful embedding representations.

  • Synthetic Instruction Dataset Creation Discussed: @_rxavier_ shared their process of creating a synthetic instruction dataset based on textbooks, where they generate a JSON with questions and answers from textbook chunks. Considering whether it would be more effective to first generate only questions and then produce the answers in a subsequent step.

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #community-showcase (2 messages):

  • Model Card Update Incoming: @ajindal agrees to @caseus_’s request to add an axolotl badge and tag to the model card, indicating a forthcoming update to the documentation.

OpenAccess AI Collective (axolotl) ▷ #deployment-help (15 messages🔥):

  • Exploring the Inference Stack for LLMs: @yamashi is diving into inference after only having experience in training LLMs. @dreamgen suggests that vLLM is probably the easiest for serving fp16 models but mentions performance issues with long contexts and moderate load.
  • Quantization and Performance Tradeoffs: Both @dangfutures and @dreamgen discuss quantization as a strategy for inference. While it can slow down processes, the price/performance tradeoff remains uncertain.
  • Hardware Requirements Uncertainty: @yamashi inquires about the necessary hardware, particularly VRAM, for running a 70b model at 500 token/s in batch mode. @dreamgen advises that hardware needs are task-dependent and recommends a trial-and-error approach.
  • Rental Before Purchase Recommended for Benchmarking: @dreamgen suggests renting hardware to benchmark performance before purchasing. This approach offers a practical solution to @yamashi’s concerns about buying the appropriate hardware for specific model requirements.
  • Quantization Yields Positive Results: @dangfutures shares a positive experience with AWQ quantization, indicating it may offer significant benefits in the context of inference performance.

Mistral ▷ #general (137 messages🔥🔥):

  • Confusion and Speculation Around Mistral and “Leak” Claims: Members of the Mistral Discord chat, including @ethux, @casper_ai, and @mrdragonfox, engaged in discussions about whether a newly emerged model was a leaked version of Mistral medium or not, referencing its performance, origins, and composition. The model in question, linked to huggingface.co/miqudev/miqu-1-70b by @alexworteega, was eventually identified not to be Mistral or a leak but perhaps a LLaMa model fine-tuned with Mistral data.

  • Technical Discussion on Model Performance and Quantization: @mrdragonfox, @elvinrath, and @i_am_dom engaged in a nuanced discussion about the impact of quantization on AI models, particularly in terms of performance degradation. It was noted that dense models, like the one speculated to be a close approximation of Mistral, could suffer more significantly from quantization effects.

  • Calls for Official Clarification on Rumored Leaks: @dillfrescott expressed a desire for Mistral staff to address rumors circulating about the potential leak of a Mistral medium model to quell speculation. @mrdragonfox responded, emphasizing that if the rumors had any merit, the company would have already taken action.

  • Reducing Tokens in Text-Based Adventure Games: @ewanhc inquired about strategies to minimize token usage in text-based adventure games without compromising the narrative’s quality. Various solutions, including embedding and retrieval methods and prompt compression, were discussed, with a link to GitHub - microsoft/LLMLingua shared by @akshay_1 as a potential avenue for exploration.

  • VRAM Requirements for Running Mistral Models: @batot4968 reported issues running Mistral AI 7B v0.1 models on a 6GB VRAM card, contrary to claims made on the official website. The discussion highlighted discrepancies between the stated system requirements and user experiences, raising questions about VRAM sufficiency for Mistral models.

Links mentioned:


Mistral ▷ #finetuning (8 messages🔥):

  • Training Mistral in a New Language: @nticaric raised a concern about the effectiveness of continuing the training of the Mistral model in a new language due to potential vocabulary mismatches. @vhariational advised against starting from scratch due to the significant resources and data required.
  • Pretraining Might Be Necessary: In the discussion about the best approach for training Mistral in a new language, @mrdragonfox suggested that pretraining might be necessary, although no specific method or data was mentioned.
  • Inquiry about BakLLaVA-1 Finetuning Resources: @attnmamba_15242 asked about the GPU resources and time requirement for finetuning BakLLaVA-1, providing links to the project’s page on Hugging Face and the collaboration behind it. BakLLaVA-1 is a blend of Mistral 7B base and LLaVA 1.5 architecture.
  • BakLLaVA-1’s Relevance to Mistral Questioned: @mrdragonfox questioned the relevance of BakLLaVA-1 information to Mistral discussions and suggested seeking information directly from the BakLLaVA GitHub repository.
  • Difficulty Fine-tuning Mistral 7B on Dolly Dataset: @bishwa3819 is experiencing challenges in reducing training loss while fine-tuning Mistral 7B on the dolly dataset using LoRA. They shared their configuration and sought advice on improving the fine-tuning process.

Links mentioned:


Mistral ▷ #showcase (1 messages):

nk_pas: Write prompts everywhere and run Mistral plateforme in a single key stroke


Mistral ▷ #la-plateforme (28 messages🔥):

  • Mistral API Rate Limits Express Concern: @arnaud_35886 queried about the maximum rate limit for using Mistral APIs in production, hinting at a need greater than the current offering. Initial confusion about the nature of the limit (rate vs cost) was clarified, and @casper_ai noted that approval for rate limit increases typically depends on usage history, starting at 2 requests/second.

  • Token and Embedding Limits Under Scrutiny: @dierkdroth sought clarification on token limits for Mistral API text embeddings and received a mix of direct and inferred answers. @sophiamyang confirmed the max token limit to be 32k for tiny/small/medium models and 8192 for the embedding API.

  • Explorations in Mistral’s Tokenization: In pursuit of details regarding the tokenizer used by Mistral’s models, @dierkdroth inquired about the specific tokenizer and referenced a JavaScript implementation. @sophiamyang acknowledged the lack of documentation and promised an update, whereas @vhariational compared it to tokenizers found in HuggingFace repositories.

  • Java Client Integration Proposal: @carloszela proposed adding a new client for Java users to the official Mistral documentation, showcasing an ongoing project on GitHub named langchain4j supporting Mistral AI models.

  • Awaiting Fixes for Early Stopping Issue: @digitalphotographer followed up on a previously reported issue regarding early stopping behavior in Mistral’s platform, looking for updates after sending notebooks for reproduction of the error to @sophiamyang and another.

Links mentioned:


Eleuther ▷ #general (8 messages🔥):

  • New AI Research Paper Alert by sk5544: sk5544 shared a new AI research paper titled AI Techniques in High-Stakes Decision Making by authors including Stephen Casper, Carson Ezell, and others, aiming to contribute more to public knowledge.

  • Biden Administration’s AI Regulation Fact Sheet Criticized: @exirae critiqued the Biden Administration’s new AI regulation fact sheet, pointing out a lack of clarity in policy direction and expressing concerns over the proposed AI/ML education in K-12 schools.

  • Exclusion of NIST AISI and Proliferation of Taskforces Noted: @hyperion.ai observed that the Biden administration’s fact sheet on AI regulation did not mention tasks assigned to the NIST AISI and commented on the contemporary popularity of taskforces in policy initiatives.

  • Skepticism over Taskforces Expressed: @clockrelativity2003 and @.undeleted expressed skepticism towards the effectiveness of taskforces, suggesting they are more about political gesturing than actual problem-solving.

  • Concerns over AI/ML Education in Schools: Both @exirae and @.undeleted voiced concerns about implementing AI/ML education in K-12 schools, fearing potential chaos due to the immaturity of school-aged children and the inadequacies of the American public school system.

Links mentioned:


Eleuther ▷ #research (119 messages🔥🔥):

  • Innovative Ideas on Model Efficiency and Training: @nshepperd discussed techniques for reducing model size by generating vocab matrices from seeds and exploring different factorization methods for the softmax bottleneck issue. They mentioned leveraging a fused matmul kernel for size reduction and experimented with preconditioning gradients to emulate training trajectories of untied weight transformers.

  • Exploring Softmax Bottleneck Solutions: Various community members, including @nshepperd, @wonkothesensible, and @stefangliga, delved into the softmax bottleneck problem, discussing alternatives like sigsoftmax and multifaceted softmax. They debated the potential impact of these alternatives on perplexity measurements and distribution errors, hinting at the complexity of fully addressing this issue.

  • Curiosity About Datasets and Quality Signals: @micpie inquired about high-quality datasets with extensive metrics, similar to RedPajama-Data-V2, highlighting a gap in easily finding such resources on platforms like HF hub due to the lack of metadata search functionality.

  • Contributions to Language Modeling and Data Preprocessing Insights: Links to research on improving language model perplexity through insights from synthetic experiments were shared by @random_string_of_character and @leegao_, while @laomein and .the_alt_man discussed a specific code snippet for data preprocessing, illustrating common challenges and clarifications in model development.

  • Reflections on Attention Mechanisms and Alternative Activations: The conversation spanned critical examinations of softmax and attention mechanisms by @catboy_slim_, @fern.bear, and @stefangliga, considering replacements like sigmoid functions and questioning the intuitive justification behind softmax’s widespread use. They discussed innovative approaches like ghost attention and the potential of rethinking probability distributions within neural networks to address outlier features and quantization challenges.

Links mentioned:


Eleuther ▷ #interpretability-general (12 messages🔥):

  • Machine Unlearning in LLMs Faces Weak Baselines: @stellaathena criticizes the current state of machine unlearning research for large language models (LLMs), highlighting that applied studies often use weak baselines and suggesting an unexplored method involving training the model with gradient ascent on bad data. They propose an experiment utilizing gradient ascent on Pythia with a subset of the Pile related to a specific topic. See the discussion.

  • TOFU Benchmark Introduced for Unlearning: @baidicoot and @stellaathena discuss the recent introduction of the TOFU benchmark, which aims to facilitate a deeper understanding of unlearning in models by comparing methods like gradient ascent. The TOFU paper, available at arXiv:2401.06121, involves a dataset of synthetic author profiles to study the effectiveness of unlearning strategies.

  • Exploring Model Lineage with a Primitive Solution: @fblgit offers a simple tool for identifying relationships or “blood-lines” between models. This solution is presented as functional yet basic, inviting others curious about model lineage to reach out for details.

  • Alternative to The Pile Download Link Provided: Following a report by @1_glados about a 404 error with the original Pile download link, @random_string_of_character provides an alternative magnet link for accessing this dataset.

  • Clarification on Unlearning and Data Exchangeability: @statslime questions the implications of data exchangeability in the context of machine unlearning discussed by @stellaathena, leading to a clarification that the evaluation involved checking the percentage of identical weights between models.

Links mentioned:

  • TOFU: A Task of Fictitious Unlearning for LLMs: Large language models trained on massive corpora of data from the web can memorize and reproduce sensitive or private data raising both legal and ethical concerns. Unlearning, or tuning models to forg…
  • Tweet from Stella Biderman (@BlancheMinerva)!): @Wetassprior @daphneipp Is “train the model with gradient ascent on bad data” an effective technique for machine unlearning? The extent to which the answer is “no” is a measure of how non-exch…

Eleuther ▷ #lm-thunderdome (2 messages):

  • Interest in New Work: @johnnysands expressed strong interest in recent developments and has plans to run some tests on the work. They added a friendly emoji to show enthusiasm.
  • Query about Offline Task Downloading: @damon_29077 inquired if there’s a way to download all tasks and cache them for offline use. This question hints at a desire for more flexible usage scenarios.

Eleuther ▷ #gpt-neox-dev (6 messages):

  • Building Apex on Frontier’s Struggle: @groggyrhombus highlighted that building apex on OLCF’s Frontier (AMD MI250xs) is time-consuming. By specifying the architectures explicitly for the build, the time required was notably reduced.

  • CUDA Might Offer Relief with Flags: In connection to building apex, @groggyrhombus suggested that CUDA could have similar flags to reduce build time, implying a possible optimization route for those working in CUDA environments.

  • Custom Compilation Approach for Apex: @catboy_slim_ detailed a custom method for compiling apex, involving the use of specific flags not mentioned in the official documentation. This method entails compiling with pip install and then creating a wheel file that can be easily distributed and imported.

  • Solution to Apex’s Build Time Across Platforms: @triggerhappygandhi confirmed the general issue of apex’s lengthy build time across various platforms and appreciated the architecture-specific build time reduction tip provided by @groggyrhombus. This acknowledgment underscores the widespread nature of the problem and the value of the shared solution.


HuggingFace ▷ #general (94 messages🔥🔥):

  • Community Guidelines on PR Tagging: @cakiki gently reminded users not to tag people for PR reviews, especially when they are unrelated to the repo, emphasizing community etiquette. This led to a discussion with @kopyl, who questioned the rule, resulting in @cakiki referring the matter to an admin for fairness.

  • Patience is Key for PR Reviews: @not_lain advised patience regarding a newly created pull request, highlighting it was only 11 hours old and reviewers might not have gotten to it yet. This advice was aimed at @kopyl, who was seeking clarification on tagging protocols for faster PR reviews.

  • Exploring Continuous Neural Network Processes: @bshvp posed an intriguing question about developing neural network models that operate continuously, mimicking human brain functions. They suggested feeding these models a constant stream of sensory inputs to truly evaluate consciousness, sparking a conversation on innovative approaches to AI development.

  • Success Story of a Deepfake Detection Pipeline: @not_lain shared the achievement of their deepfake detection pipeline reaching 147 downloads in less than a week and provided a detailed explanation, including a link (https://huggingface.co/not-lain/deepfake) for the community. This sparked interest from @adiinx, who sought advice on deepfake content generation and voice cloning, leading @not_lain to clarify the pipeline’s limitations and applications.

  • Fedora Versus Ubuntu for Machine Learning on AMD Hardware: @cilginflix shared challenges with using CPU for ML tasks due to issues with onnxruntime and AMD’s ROCm on Fedora, leading to a discussion on operating system compatibility for ML. @kopyl suggested NVIDIA hardware for ML, but @cilginflix countered that AMD works on other distributions like Ubuntu or Manjaro, albeit with different performance and stability outcomes.

Links mentioned:


HuggingFace ▷ #today-im-learning (10 messages🔥):

  • A celebratory moment for @osanseviero: @osanseviero only shared a brief but heartfelt “Congrats!!”, to which @not_lain responded with “thank youuuu” and a special Hugging Face emoji. The context of the celebration remains a mystery.
  • @erlonidasap seeks guidance on model training: Starting a new project, @erlonidasap asked the community for tips on how to train a model on a specific string dataset. It’s related to a company project, but no further details were provided.
  • Comprehensive advice from @not_lain: In response to @erlonidasap’s query, @not_lain suggested first identifying the problem (text-classification, text-generation, QA, etc.), then selecting a suitable model from the Hugging Face Hub, and finally finetuning it on the specific data.
  • @kaizen0340 inquires about speech translation training: Looking for resources, @kaizen0340 asked the community if there’s any tutorial or repository on how to train a speech translation using speechencodedecoder. Unfortunately, there was no response provided in the messages.
  • @antiraedus shares a weekly update and reflections: Highlighting the importance of setting correct habits and maintaining focus, @antiraedus shared insights from their week, including personal development in flutter, exercise, and brainstorming app ideas. They emphasized the goal of building a strong foundation for consistent productivity.

HuggingFace ▷ #cool-finds (3 messages):

  • Deep Learning Breakthroughs in Medical Imaging: @maro_mich shared an open access article from IEEE detailing how Deep Learning is revolutionizing Medical Image Analysis. Check out the findings here.
  • Another Must-Read Medical Imaging Paper: @aryan_1098 encourages the community to explore a paper on Medical Image Analysis available on IEEE. The study can be found here.

Links mentioned:


HuggingFace ▷ #i-made-this (7 messages):

  • Shamanic Animation Intrigues: Amanita shared an animated fly agaric shaman, described as stylized. No links or further details were provided.

  • Comprehensive AI Model Demo Space by myg5702: myg5702 launched a Hugging Face Spaces demo showcasing a variety of AI models including sdxl, fooocus, and dream shaper xl turbo among others. This Space seems to be a one-stop showcase of cutting-edge image models.

  • Innovative Resume Space by not_lain: not_lain created a resume question answering space on Hugging Face, designed to make it easier for recruiters to find information. They encourage users to heart react the space if they appreciate the innovative concept.

  • Mistral on Apple Silicon Video by monirul_1slam: monirul_1slam released a YouTube video demonstrating the performance of the Mistral model on Apple Silicon, specifically geared towards tech enthusiasts and developers interested in machine learning performance on different hardware.

  • Challenging ABSA Model Creation by joshuasundance: joshuasundance accepted a challenge to train and upload a SetFitABSA model for laptop reviews, sharing links to their work on Hugging Face. This contribution highlights the active collaboration and learning in aspects of sentiment analysis within the community.

  • Elvis-like Vocals Captured by .bigdookie: .bigdookie shared their excitement about a session where their vocalist resembled Elvis, signalling a whimsical and light-hearted moment in creative music production. The context remains in the space of amusement and tease for future releases.

  • Excel/CSV to Database Tables App by impl66: impl66 developed a Gradio app that converts Excel/CSV files into database tables and subsequently answers user questions on those tables, showcasing practical utility in data management and retrieval.

Links mentioned:


HuggingFace ▷ #diffusion-discussions (11 messages🔥):

  • Election Tampering Discussed with AI: @iloveh8 raised a query about AI products that could aid a political candidate in winning an election, mentioning deepfakes as a potential use case. @vipitis responded with examples such as fraud, supplying additional votes, redrawing borders for an unrepresentative lead, and even poisoning the competition.
  • Controversial Use of AI in Politics: @vipitis replied to the discussion on AI’s application in political campaigns by suggesting any means necessary, including fraudulent actions, assuming no rules were in place.
  • Seeking Ethical Alternatives in Political Campaigns: In contrast to the controversial uses, @chad_in_the_house suggests creating numerous videos of a political candidate as a more legal and ethical form of AI application in campaigns.
  • Interest in Technical Details for Campaign Videos: Following up on the discussion of creating videos, @iloveh8 expressed curiosity about the technical aspects, highlighting the ongoing media discussion around such AI-driven content.
  • Inquiry on Text to Image Diff Models: @syed2658 shifted the conversation towards technical advice, asking for suggestions on the best text-to-image diffusion models. This indicates a broader interest in AI applications beyond political implications.

Links mentioned:

How to WIN an Election | Ordinary Guide: Follow me on twitter: https://twitter.com/ordinarytingsSupport the Channel on Patreon: https://www.patreon.com/ordinarythingsHow do you win an election? By l…


HuggingFace ▷ #computer-vision (1 messages):

swetha98: I am getting this error while running code for fine tuning donut docvqa .


HuggingFace ▷ #NLP (3 messages):

  • GPU acceleration confuses with llama-cpp: User @.sgp inquired about difficulties in achieving GPU acceleration while using llama-cpp, seeking insights or solutions from the community.
  • Amazement at GPU requirement for llama-cpp: Responding to @.sgp’s dilemma, @frosty04212 expressed astonishment at the requirement for 128 GPU layers for operating llama-cpp, wondering about the specifications of the GPU involved.

HuggingFace ▷ #diffusion-discussions (11 messages🔥):

  • Discussing AI’s Role in Elections Gets Dark: @iloveh8 sparked a conversation asking about specific AI use cases that could help a political candidate win an election. @vipitis responded controversially, suggesting fraud as a method by emphasizing actions like supplying additional votes or redrawing district borders for an unrepresentative lead.
  • From Strategy to Digital Content: Shifting from the grim suggestions, @chad_in_the_house proposed a more lawful approach by creating numerous videos of a political candidate to generate more content and visibility.
  • The Technicalities of Media Influence: Following up on the suggestion to create more videos of a candidate, @iloveh8 expressed curiosity about the technical details behind this strategy, especially considering its frequent discussion in the media.
  • In Search of Text-to-Image Models: @syed2658 inquired about recommendations for the best text-to-image diffusion models, diverging from the election-focused conversation to seek advice on AI models.
  • An Unorthodox Election Guide: Besides giving controversial advice, @vipitis also shared a YouTube video titled “How to WIN an Election | Ordinary Guide” (watch here), providing a humorous take on winning elections.

Links mentioned:

How to WIN an Election | Ordinary Guide: Follow me on twitter: https://twitter.com/ordinarytingsSupport the Channel on Patreon: https://www.patreon.com/ordinarythingsHow do you win an election? By l…


Latent Space ▷ #ai-general-chat (92 messages🔥🔥):

  • ColBERT vs. Traditional Embeddings Discussion Heats Up: @sarav1n sparked a debate on the real-world usage of ColBERT vs. current embedding models, sharing insights from Simon Willison’s article, which highlights ColBERT’s unique approach to computing similarity scores. The conversation included a desire for a direct comparison between ColBERT and conventional embedding models, with members like @swyxio and @420gunna expressing interest but noting a lack of comparative data.

  • Simon Willison Offers Fresh Perspectives on AI and Software: @mdcker shared a feature on Simon Willison discussing AI’s impact on software development and related issues. The community appreciates Willison’s accessible explanations of complex AI topics, sparking further discussion on the human element in AI software development.

  • “Arc Search” Revolutionizes Mobile Web Browsing: @mdcker introduced Arc Search, a new iOS app designed to streamline web searches by building webpages based on user queries, potentially changing how users interact with search engines and browser apps. The community speculated on its impact and potential competition with traditional search engines.

  • Voyage-Code-2 Promises Better Code Retrieval: @gsegato announced the release of Voyage-Code-2, an embedding model claiming superior performance in code-related applications. This sparked a conversation, led by @swyxio, about the value proposition of specializing in embeddings and the model’s anticipated evaluation on benchmarks like MTEB.

  • Anthropic’s Claude Models Under the Microscope: @philltornroth questioned why Anthropic’s Claude models aren’t receiving more attention compared to OpenAI, suggesting their performance in tasks like summarization and retrieval could be more competitive than perceived. This led to a nuanced discussion on the specific use cases where Claude models excel and areas where improvements are needed, contributing to the ongoing debate on the best AI models for various applications.

Links mentioned:


DiscoResearch ▷ #mixtral_implementation (5 messages):

  • MIQU Model Sparks Mixtral Medium Leak Speculation: @aiui raised a question about the MIQU model being a potential Mixtral medium leak, sparking a conversation about its origins and characteristics.
  • Debunking the MIQU Model-Mixtral Connection: @sebastian.bodza quickly addressed the speculation around the MIQU model being a Mixtral medium leak, stating that these rumors were already debunked.
  • Clarification Tweet from Nisten: For further clarity, @sebastian.bodza shared a tweet by Nisten that helps debunk the rumors surrounding the MIQU model.
  • MIQU Model and LLaMA Tokenizer Connection Highlighted: Additionally, @sebastian.bodza highlighted that the MIQU model uses a LLaMA tokenizer, sharing another relevant tweet for more details.

DiscoResearch ▷ #general (6 messages):

  • Exploring German Fine-tuning on Phi 2: @philipmay raised doubts about Phi 2 having German text training, prompting a discussion on its potential fine-tuning. @johannhartmann shared an attempt at finetuning it for German using the OASST-DE dataset but noted unimpressive results from a single epoch attempt, suggesting a more comprehensive approach with additional datasets and epochs might be worth exploring.
  • Tiny Llama Gets a German Upgrade: @johannhartmann documented an effort to improve Tiny Llama with German DPO (Data Processing Overlay), linking to a walkthrough on GitHub (TinyLlama-1.1B-Chat-v1.0-german-dpo-Openllm_de.md). This method showed some positive outcome, hinting at the potential of language-specific training enhancements.
  • Curious About German Orca DPO DataSet?: @philipmay inquired about the existence of a German Orca DPO dataset, which led @johannhartmann to share a link to a somewhat makeshift dataset on Hugging Face (mayflowergmbh/intel_orca_dpo_pairs_de), detailing its use of AzureML for translation and HermEO as the rejected model.
  • WRAP Aims to Enhance Pre-training: @bjoernp highlighted an innovative approach by Apple in a recent paper on Web Rephrase Augmented Pre-training (WRAP) which aims to improve data quality and reduce pre-training times by paraphrasing web documents into styles like Wikipedia or Q&A (arXiv:2401.16380). This method, according to the abstract, has significantly sped up pre-training efforts, hinting at its potential widespread adoption among major LLM players.

Links mentioned:


DiscoResearch ▷ #embedding_dev (55 messages🔥🔥):

  • Exploring CodeT5 Embeddings: @sebastian.bodza inquired about implementing Salesforce’s CodeT5 embeddings for code, struggling with the practical aspects. No insights were provided by @bjoernp due to a lack of experience with these models.

  • First Glance at Text Generation Attempts: @sebastian.bodza shared their initial results in text-generation and requested feedback. A specific notebook was shared, Embedding_Training/05_preprocess_texts.ipynb, for collaborators to review.

  • Defining “Hard Negatives” in Training: A considerable amount of discussion focused on creating high-quality “hard negatives” for training data. @sebastian.bodza and @philipmay debated the criteria, with emphasis on the importance of hard negatives not answering the question directly to prevent training issues.

  • Iterative Improvements on Training Prompts: @sebastian.bodza iteratively improved the prompts used for generating training data, aiming to fine-tune the balance between positive examples and hard negatives. These modifications were driven by feedback and the quest for optimally structured data sets.

  • Misconceptions About Retrieval Models Discussed: Clarity was sought on the purpose and structure of datasets for DPR and RAG training, particularly regarding context and the kind of hard negatives. @sebastian.bodza clarified that their work aims to generate a passage retrieval dataset for future RAG use, with examples varying in quality rather than content relevance.

Links mentioned:


LangChain AI ▷ #general (46 messages🔥):

  • Exploring LangChain Capabilities: User @irfansyah5572 requested assistance on running an answer chain and displaying the source in LangChain, linking to the project’s GitHub.

  • Tutorial Alert for RAG Frontend Development: @a404.eth shared a new YouTube tutorial focusing on building a frontend for interacting with PDF documents using LangChain and other technologies, asking for feedback on this second part of their tutorial series.

  • Custom Book Recommendation System with LangChain: @oscarjuarezx is developing a demo book recommendation system based on user interests, utilizing PostgreSQL and various LangChain components, and is seeking advice on optimizing database searches.

  • Cache Issues with LlamaCPP Models: Several users, including @techexplorer0 and @kapa.ai, discussed the use of InMemoryCache for improving inference times with LlamaCPP models, with a link to a related GitHub issue, but noted challenges with cache effectiveness.

  • Parsing and Streaming Challenges in LangChain: Users @ibrobabs and @hiranga.g exchanged experiences and issues with parsing JSON output and streaming from agents in LangChain, mentioning a helpful GitHub example and a YouTube video for further guidance.

Links mentioned:


LangChain AI ▷ #share-your-work (3 messages):

  • Local LLMs Elevate Web Browsing: @andrewnguonly launched an open-sourced LLM copilot Chrome extension, named Lumos, leveraging LangChain and Ollama to enrich web browsing experiences with local LLMs. They encouraged feedback and shared a link to support the launch on Product Hunt.

  • Building with Pre-written SQL as Tools: @shownaldo was curious about incorporating SQL into a project, pondering if it meant having pre-written SQL for the agent to choose from. In response, @johnny2x2 shared that they pivoted to including predefined SQL scripts as tools for their local LLM due to difficulties in having it write custom SQL code.

Links mentioned:


LangChain AI ▷ #tutorials (3 messages):

  • Beginner-Friendly Langchain Agents 2024 Update: @ryannolan shared a YouTube video titled “Langchain Agents [2024 UPDATE] - Beginner Friendly”. The video provides a new way to build agents with the latest Langchain update, aiming to expand the capabilities of the OpenAI API.

  • Building a RAG with LangChain and Unstructured, Part 2: @a404.eth announced the release of their new tutorial, “Chat With Your PDFs Part 2: Frontend - An End to End LangChain Tutorial. Build A RAG with OpenAI”. It focuses on frontend development, including building the chat interface, displaying source documents, and streaming output, utilizing technologies like LCEL, React, and TypeScript. The tutorial is available at YouTube.

  • Excitement over Tutorial: @a404.eth expressed enthusiasm for their tutorial on building a frontend for a RAG using LangChain with a succinct comment, “Ok this is sick”.

Links mentioned:


LlamaIndex ▷ #blog (6 messages):

  • Simple method to tackle complex classifications: @KarelDoostrlnck introduces a new method using LLMs for complex classifications involving thousands of classes, like medical reactions and job skills. This approach involves inferring a set of predictions, then retrieving and re-ranking results. IFTTT Announcement
  • LlamaIndex.TS updates: New features and improved documentation have been launched for LlamaIndex.TS. Read the announcement and updates here.
  • Hackathon in Silicon Valley with $16,000 prizes: The LlamaIndex RAG-A-THON is an in-person event focusing on Retriever-Augmented Generation (RAG) technology to create advanced AI agents. It requires at least one team member to present physically, offering prizes up to $16,000. Hackathon details
  • RAG Hackathon participation query: @rawwerks inquires about attendees for the LlamaIndex RAG Hackathon, emphasizing its in-person requirement. Context of query
  • Evaluating LlamaPacks for advanced querying: @wenqi_glantz tested 7 pre-baked advanced query strategies, known as LlamaPacks, with @lighthouzai, showcasing how they can streamline choosing the best querying strategy for specific needs. LlamaPacks evaluation

Links mentioned:

LlamaIndex RAG Hackathon (in-person only): Think Beyond Chatbots: Unleashing the Potential of AI Agents


LlamaIndex ▷ #general (36 messages🔥):

  • Local LlamaCpp Server for Rag Implementation: @techexplorer0 is setting up a local LlamaCpp server using instructions from GitHub - abetlen/llama-cpp-python and seeks advice on integrating it with a RAG application for making LLM calls.

  • Azure Analysis Services Query with Llama_Index: @meowmeow008 successfully connected Llama_Index to Azure SQL but is exploring options to connect it to Azure Analysis Services OLAP instead of SQL.

  • Exploring Llama2 Licensing for Commercial Use: @sattyman inquires about using Llama2 commercially and creating a finetuned version for their product. @nerdai linked to an article suggesting Llama2 is available for commercial use with certain restrictions, and also recommended checking Meta’s official site for fine print and licensing details.

  • Customizing LLM Output Format: @mysterious_avocado_98353 wanted to know how to customize output format to exclude specific words. @nerdai suggested modifying prompts as per the guidelines on LlamaIndex Docs.

  • Issues with Query Engines in Llama Index: @coder_004_71487 discussed handling generic queries and specific SQL errors with Llama Index. @nerdai recommended using a RouterQueryEngine for mixed queries and directed @coder_004_71487 to submit a bug report through GitHub Issues and check out a demo on configuring a RouterQueryEngine for more intricate query handling.

Links mentioned:


LAION ▷ #general (31 messages🔥):

  • MagViT2 Video Tokenizer Troubles: @_kevinalbert is encountering issues with pixelated samples after 10,000 training steps with MagViT2 using the MagViT2 PyTorch repository. They have been advised by @top_walk_town to adjust the learning rate and various loss functions to improve outcomes.

  • Nightshade’s Controversial Impact: @drhead critiques the release of Nightshade as irresponsible and doubts its immediate threat without significant coordinated use and without countermeasures being developed. They suggest the potential for finetuning to neutralize or mitigate Nightshade’s perturbations.

  • Skepticism around Glaze and Nightshade’s Effectiveness: @astropulse and @.undeleted express skepticism and concern over the intentions and effectiveness of the Glaze and Nightshade tools, with a particularly critical view on their potential long-term impact and the serious implications of their use.

  • Discussion on Counteracting Nightshade: @drhead and @.undeleted debate potential strategies to make Nightshade ineffective, such as avoiding the targeted encoder or training new models that could nullify Nightshade’s perturbations, yet remain cautious about the generalized threat posed by such attacks.

  • Views on the Necessity of New Data for AI Models: @pseudoterminalx and @mfcool argue against the need for new images or data to improve AI models, highlighting the larger issue of data quality and proper captioning over quantity in enhancing model performance.


LAION ▷ #research (7 messages):

  • “Activation Beacon” Potentially Solves Context Length Issue: @spirit_from_germany highlighted a new method called “Activation Beacon” that might revolutionize LLMs by allowing for unlimited context length. The method involves adding “global state” tokens before the prompt, with remarkable results showing that an LLaMA 2 model trained for 10K steps at 4K context length could then handle up to 400K context length. Read the paper and check the code.

  • Discussion on the Optimal Way to Store Captioned Image Datasets: @progamergov inquired whether it is better to store images in parquet files or separately in tar files for a captioned image dataset. @chad_in_the_house responded, recommending tar files for being faster to query through webdatasets.

  • Latent Diffusion in RePaint Explored?: @epicycles_989 questioned if the method used in RePaint has been applied to latent diffusion, expressing interest in its potential for zero-shot in/outpainting. No direct responses or additional information were provided regarding this query.

  • **Advantage of Webdatasets: Following a discussion on the storage of captioned image datasets, @random_string_of_character provided links to webdataset and tarp on GitHub, highlighting that webdatasets offers a high-performance Python-based I/O system for deep learning problems and tarp enables fast and simple stream processing of files in tar files. Webdataset GitHub | Tarp GitHub

  • Combining Parquet and TAR for Efficient Data Management: @progamergov considered the idea that using parquet files for captions and tar files for corresponding images could be the most efficient storage solution. This hybrid method might offer the best of both worlds in terms of query speed and data organization.

Links mentioned:


Perplexity AI ▷ #general (32 messages🔥):

  • Inquiring Minds Want to Know the Best AI Model Usage: User @vic7669 was curious about when to use specific AI models like Gemini, GPT-4, and Claude. The user was directed to a thread and a technical FAQ for insights.

  • Library Visibility Issue in Sidebar: User @jamiecropley experienced difficulties with the library not always showing on the left-hand side. Despite attempting a drastic measure of reinstalling Windows, the issue persisted, and the discussion clarified the behavior as a feature, not a bug, with only the last eight threads/collections being listed.

  • Cryptocurrency Inquiry and Clarification: User @lambda4life inquired about a possible crypto token associated with the project. The question was promptly addressed by @icelavaman, confirming that no crypto token exists for the project.

  • Querying Perplexity via URL Parameters: User @fyngraf asked if it was possible to initiate a query to Perplexity using URL parameters, aiming for an easier search utility. They were directed to a specific Discord channel for further information.

  • Rabbit R1 Gains Attention: User @brownpatricks sparked a conversation about purchasing the Rabbit R1, inviting others to share their interest levels through emoji responses. Interest was expressed by several users, with some citing financial constraints while others admired the product’s design.

Links mentioned:


Perplexity AI ▷ #sharing (5 messages):

  • Gemini Pro Tackles Chatbot Arena: @brknclock1215 shared a YouTube video discussing how New Bard surpassed GPT-4 in Chatbot Arena, showcasing a significant performance boost.
  • Personalizing Search with Great Features: @sa1k0s expressed appreciation for a tip that allows personalization of their search engine, praising the feature’s ability to customize.
  • Creative Use of AI for a Pomodoro App: @gammagames is developing a delivery-themed pomodoro app, successfully using an AI tool to generate names and addresses for the app content, finding the tool greatly beneficial.
  • Explore Perplexity’s Lab: @rowalth highlighted the existence of a lab on Perplexity for experimentation, pointing out a resource for creative exploration.

Links mentioned:

🔥 New Gemini Pro Better than GP-4? Huge Performance Boost on ⚔️ Chatbot Arena ⚔️: New Bard has surpassed GPT-4 on Chatbot Arena. 🦾 Discord: https://discord.com/invite/t4eYQRUcXB☕ Buy me a Coffee: https://ko-fi.com/promptengineering|🔴 Pat…


Perplexity AI ▷ #pplx-api (1 messages):

  • Custom Stop Words Feature Inquiry for pplx-api: @dogemeat_ asked about the implementation timeline for custom Stop Words in the pplx-api, expressing an interest in integrating it with the zed.dev editor for its ‘assistant’ features. This integration aims to offer an alternative to the default OpenAI models.

LLM Perf Enthusiasts AI ▷ #triton (1 messages):

  • The Sweet Spot Between PyTorch and CUDA: @tvi_ expressed that while PyTorch allows for an “almost declarative” approach to CUDA computation by abstracting away the how of operations like matrix multiplication (a @ b), CUDA and Numba.CUDA require explicit detailing of computation processes. Triton is highlighted as a middle ground, providing more control than PyTorch but with less complexity than direct CUDA, appealing for those who seek a balance.

LLM Perf Enthusiasts AI ▷ #cuda (29 messages🔥):

  • CUDA Optimization Odyssey: @artste and @zippika shared their journey trying to optimize RGB to Grayscale conversion on CUDA. Various attempts included using structs for cleaner indexing, employing vectorized operations with uchar4, and exploring different memory layouts such as CHW and HWC.
  • Vectorization Proves Challenging: Despite initial excitement, @zippika found that vectorizing the conversion process with uchar4 and shared memory led to slower performance than expected. Further optimizations and tweaks also failed to yield the anticipated speedup.
  • The Power of Inline: A moment of revelation came when @zippika noted that using __forceinline__ in their CUDA kernel significantly boosted performance, highlighting the importance of inlining for GPU code optimization.
  • A Misleading Benchmark: @zippika experienced a rollercoaster of emotions upon realizing an error in their test setup—a naive implementation appeared faster due to it only processing a single 3-pixel height image, which skewed the comparison.
  • Seeking Community Wisdom: @andreaskoepf suggested further memory optimization techniques such as using consecutive memory reads and leveraging int4 data types for vector loads and stores, directing @zippika to a NVIDIA developer blog for deeper insights into maximizing bandwidth efficiency.

Links mentioned:

CUDA Pro Tip: Increase Performance with Vectorized Memory Access | NVIDIA Technical Blog: This post demonstrates the use of vectorized memory access in CUDA C/C++ to increase bandwidth utilization while decreasing instruction count.


LLM Perf Enthusiasts AI ▷ #pmpp-book (3 messages):

  • Chapter 2 Doubt Cleared Up: @ashpun expressed confusion over the 3rd question in chapter 2 and sought clarification, mentioning difficulty in understanding the explanation provided in the 2nd video. They were open to any explanations on how to approach solving this question.
  • Figuring Out Coalescing and Banking in Chapter 6: @shindeirou discussed confusion around figures 6.10 and 6.11 in chapter 6, questioning the efficiency of accessing memory banks within the same channels. They later understood that going to the same channel but different banks allows for effective interleaving to hide memory latency.
  • Solving Indexing in Memory Access: @andreaskoepf provided a solution for understanding memory indexing as discussed in chapter 2, offering a specific code snippet and explanation for calculating the index to elements within sections of memory. This addressed @ashpun’s query on the 3rd question in chapter 2.

LLM Perf Enthusiasts AI ▷ #youtube-recordings (1 messages):

andreaskoepf: New video link: https://youtu.be/4sgKnKbR-WE?si=J-B0kHqknRXhE7e_


LLM Perf Enthusiasts AI ▷ #general (9 messages🔥):

  • Seeking Superior Writing Assistant Tools: @an1lam expressed dissatisfaction with current writing assistant tools, highlighting the chat interface’s limitations for iterative writing and the absence of quality alternatives like lex.page.
  • An Enthusiastic Recommendation but with a Catch: In response to @an1lam’s query, @jeffreyw128 tagged a user, possibly suggesting a recommendation or a lead on writing assistant tools, albeit without providing concrete details.
  • Startup Struggles with Google Workspace Costs: @frandecam shared concerns about the high costs of provisioning Google Workspace accounts for their bootstrapped AI startup, highlighting the financial burden it represents for startups.
  • A Ray of Hope for AI Startups: Responding to @frandecam’s dilemma, @dare.ai suggested the Google Cloud for Startups program as a solution to mitigate costs, though cautioned about potential delays due to backlogs. They shared a link to the program (Google for Startups Cloud Program) and offered further assistance.
  • Introducing Jenni.ai for Academic Writing: @calclavia introduced jenni.ai, a tool under development aimed at enhancing academic writing, implying it as a potential solution for @an1lam’s quest for effective writing assistance tools.

Links mentioned:

AI startup program | Google Cloud: Tap into the best of Google’s infrastructure, AI products, and foundation models. Get up to $250,000 USD in Google Cloud credits, training, and more.


LLM Perf Enthusiasts AI ▷ #gpt3-5 (1 messages):

  • Gorilla Launches OpenFunctions: @shacrw shared a blog post about Gorilla OpenFunctions, a new open-source alternative for formulating executable API calls using natural language prompts. This solution simplifies API calls for various services, making it accessible even to those with minimal programming knowledge.
  • Explore Gorilla Functions on GitHub: Accompanying the blog share, @shacrw also provided a GitHub link to explore the OpenFunctions project further. This initiative is driven by the need to enhance Large Language Model (LLM) chat completion features with the capability to accurately format function calls based on API documentation and question-answer pairs.

Links mentioned:


LLM Perf Enthusiasts AI ▷ #gpt4 (1 messages):

  • Prompt Investing Vision Model Hits a Copyright Roadblock: @jxnlco is seeking advice on how to effectively use a prompt investing vision model for reading complex labels, but is facing challenges as the model is flagging the material as copyrighted.

LLM Perf Enthusiasts AI ▷ #opensource (1 messages):

thebaghdaddy: do we believe the mistral medium hype about becoming 2nd only to GPT4?


Skunkworks AI ▷ #general (7 messages):

  • Frankenstein Model Experiments Underway: @nisten announced the creation of a composite model by merging all the 70b CodeLlamas. This experimental model aims to explore enhanced performance benchmarks.
  • Laughter and Progress with BigCodeLlama: Amidst the serious tech discussion, @nisten also shared a lighter moment with a “lolol” message, indicating both the fun and progress in their work.
  • BigCodeLlama 169B Voilà!: @nisten has uploaded the BigCodeLlama-169b model on Hugging Face, a powerful amalgamation designed to benchmark how different combinations of models perform together.
  • Benchmarking the FrankenLlama: Following the model upload, @nisten shared a challenging coding problem involving calculating Aldrin cycler orbits for Mars colonization as a test case to compare the FrankenLlama model against the stock models.
  • Friendly Check-In amid the Tech Talk: @zentorjr dropped in with a casual greeting for @nisten, adding a personal touch to the high-level technical discussion unfolding in the channel.

Links mentioned:

nisten/BigCodeLlama-169b · Hugging Face: no description found


Skunkworks AI ▷ #off-topic (2 messages):

  • Diving Into ColBERT with RAGatouille: @pradeep1148 shared a YouTube video titled “Exploring ColBERT with RAGatouille,” highlighting RAGatouille, a library facilitating work with the ColBERT retrieval model. ColBERT is described as a fast and accurate model that improves scalable behavior summarization.

  • Unveiling the Eagles’ Flight with Eagle 7B: @pradeep1148 introduced another YouTube video titled “Running 🦅 Eagle 7B on A40,” discussing Eagle 7B, a significant advancement in the RWKV-v5 architecture and linear transformers across 100+ languages. Eagle 7B is notable for processing 1 trillion tokens, marking a new era for transformer tech.

Links mentioned:

  • Exploring ColBERT with RAGatouille: RAGatouille is a relatively new library that aims to make it easier to work with ColBERT.ColBERT is a fast and accurate retrieval model, enabling scalable BE…
  • Running 🦅 Eagle 7B on A40: 🦅 Eagle 7B : Soaring past Transformers with 1 Trillion Tokens Across 100+ LanguagesA brand new era for the RWKV-v5 architecture and linear transformer’s has…

Datasette - LLM (@SimonW) ▷ #ai (3 messages):

  • Emoji Suggestion Tool Unveiled: @dbreunig showcased a new demo at Emoji Suggest that can convert short sentences or headlines into a single, recommended emoji using the CLIP model. The code for the tool is available, utilizing precomputed embeddings for emojis to facilitate quick and sophisticated searches.

  • Insights on Embeddings for Search Applications: @dbreunig shared insights on using embeddings to create sophisticated search tools quickly, emphasizing the importance of curating the options for truly effective search capabilities. This approach underpins the functionality of their emoji suggestion tool.

  • AI’s Positive Perspective on its Role: @bdexter relayed a conversation where an AI, specifically llama2, expressed the belief that artificial intelligence is a force for good, capable of helping humans reach their full potential. This interaction highlights the positive outlook that AI can hold regarding its impact on human achievement.

Links mentioned:

Emojify: no description found


Datasette - LLM (@SimonW) ▷ #llm (1 messages):

  • ColBERT Writeup Appreciation: User @bewilderbeest expressed gratitude towards @746595581086138409 for the ColBERT writeup on TIL. They are creating a notebook using the shared code snippets and praised the heatmap visualization of words in the results.

Alignment Lab AI ▷ #general-chat (2 messages):

  • Is GPT-5 Training Underway?: User @entropi shared a YouTube video titled “GPT-5: Everything You Need to Know So Far”, sparking curiosity about whether GPT-5 has commenced training, citing exclusive interviews and information from OpenAI.
  • Confirmation on GPT-5 Training: In response to the speculation, @lightningralf confirmed that GPT-5 has indeed been in training for some time, though no specific details were provided.

Links mentioned:

GPT-5: Everything You Need to Know So Far: Was yesterday the day GPT-5 actually started training? This video has everything we think we know so far about GPT-5, drawing on exclusive interviews, OpenAI…


AI Engineer Foundation ▷ #general (1 messages):

  • ChatGPT vs. Bard on Captcha Images Moderation: @juanreds pointed out an interesting difference between ChatGPT and Bard: ChatGPT moderates Captcha images while Bard does not. This highlights divergent approaches to user interactions and security measures between the two AI systems.