Frozen AI News archive

DeepSeek-V2 beats Mixtral 8x22B with >160 experts at HALF the cost

**DeepSeek V2** introduces a new state-of-the-art MoE model with **236B parameters** and a novel Multi-Head Latent Attention mechanism, achieving faster inference and surpassing GPT-4 on AlignBench. **Llama 3 120B** shows strong creative writing skills, while Microsoft is reportedly developing a **500B parameter** LLM called **MAI-1**. Research from Scale AI highlights overfitting issues in models like **Mistral** and **Phi**, whereas **GPT-4**, **Claude**, **Gemini**, and **Llama** maintain benchmark robustness. In robotics, **Tesla Optimus** advances with superior data collection and teleoperation, **LeRobot** marks a move toward open-source robotics AI, and **Nvidia's DrEureka** automates robot skill training. Multimodal LLM hallucinations are surveyed with new mitigation strategies, and **Google's Med-Gemini** achieves SOTA on medical benchmarks with fine-tuned multimodal models.

Canonical issue URL

More experts are all you need?

DeepSeek V2 punches a hole in the Mistral Convex Hull from last month:

image.png

Information on dataset is extremely light; all they say is it's 8B tokens (4x more than DeepSeek v1) with about 12% more Chinese than English.

Snowflake Arctic was the last very large MoE model with the highest number of experts (128) we'd seen in the wild; DeepSeek v2 now sets a new high water mark scaling up what was already successful with DeepSeekMOE, but also introducing a new attention variant called Multi-Head Latent Attention.

image.png

These result in much faster inference by caching compressed KVs ("reducing KV cache by 93.3%").

image.png

The paper details other minor tricks they find useful.

DeepSeek is putting their money where their mouth is - they are offering token inference on their platform for $0.28 per million tokens about half of the lowest prices seen in the Mixtral Price War of Dec 2023.


Table of Contents

[TOC]


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

LLM Developments and Releases

Robotics and Embodied AI

Multimodal AI and Hallucinations

Emerging Architectures and Training Techniques

Benchmarks, Frameworks, and Tools

Trends, Opinions, and Discussions


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Development and Capabilities

Societal Impact and Concerns

AI Applications and Developments

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries


PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord

GGUF Conversion Hiccups for Llama3: The Unsloth community encountered conversion issues with Llama3 models when using llama.cpp, notably affecting training data when transitioning to GGUF format. Issues weren't limited to FP16 conversions, implying deeper underlying problems than just precision loss.

New Lines, Big Problems: A recurrent theme in the glitches was linked to new line tokenization, with different behaviors across regex libraries leading to erratic tokenizer.json patterns. Potential solutions involving regex modifications were explored to fix the GGUF conversion challenges.

Llama Variant Takes on Genomic Data: The introduction of the LLaMA-3-8B-RDF-Experiment model by M.chimiste marks a push towards integrating LLMs with genomic data and knowledge graph construction.

Demand for Vision-Language Model Tuning Tools: Community request surfaced for a generalized method to fine-tune Language-Vision Models (LVLM), demonstrated by a member's interest in supporting Moondream, as detailed in their GitHub notebook.

Showcasing and Sharing Platform Growth: Proposals for a separate discussion channel on deploying large language models (LLMs) highlight a demand for shared learning. This aligns with showcases like Oncord's integration of Unsloth AI for web development AI tools and the release of models that enhance Llama-3 capabilities.


OpenAI Discord

Perplexity AI Pulls Ahead with Pages: Perplexity AI's new Pages feature garners attention for its ability to create comprehensive reports. Meanwhile, a healthy skepticism surrounds the potential of GPT-5 as engineers discuss the diminishing returns on investment.

AGI Concept Sparks Debate: The AI community on Discord is locked in a debate over the definition of AGI and whether AI models like ChatGPT are pioneering versions of AGI. Interest in AI-generated music indicates a growing appetite for creative AI applications, with reference to services like Udio.

Performance Frustration Hits GPT-4 Turbo: Significant increases in response latency are reported for GPT-4 Turbo, and users are seeking clarity about inconsistent message cap thresholds, suggesting possible dynamic adjustments during peak times.

Prompt Engineering Challenges and Strategies: Engineers share experiences and resources, recommend "Wordplay" by Teddy Dicus Murphy for prompt-crafting insights, and delve into the intricacies of using logit bias to manipulate token probabilities in the OpenAI API.

Fine-Tuning AI for Queries: A lively discussion revolves around fine-tuning models to generate questions rather than answers, including strategies for improving GPT-4-TURBO prompts for product information extraction, backed by a logit bias tutorial.


Stability.ai (Stable Diffusion) Discord


Nous Research AI Discord


Perplexity AI Discord


LM Studio Discord


CUDA MODE Discord

Backpack Packs a Punch: BackPACK, a PyTorch extender for extracting additional information from backward passes, has been discussed, highlighting its potential for PyTorch developers. Details are in the publication "BackPACK: Packing more into Backprop" by Dangel et al., 2020.

DoRA Delivers on Fusion: A new fused DoRA layer implementation decreases the number of individual kernels and has been optimized for GEMM and reduction operations, detailed in a GitHub pull request. Enthusiasm was noted for upcoming benchmarks focused on these enhancements.

Custom CUDA Extensions Customization: Members discussed best practices for installing custom PyTorch/CUDA extensions, sharing multiple GitHub pull requests like PR#135 and a sample setup.py for reference, aiming for cleaner installation processes.

Streaming Ahead with CUTLASS Interest has bubbled around stream-K scheduling techniques used in CUTLASS, with suggestions of diving deeper into its workings in a future talk.

GPU Communication Goes to School: Upcoming sessions on GPU Collective Communications with NCCL have been announced, with a focus on distributed ML concepts.

Must-Read ML Systems Papers: For newcomers to machine learning systems, an ML Systems Onboarding list on GitHub provides a curated selection of informative papers.

Overcoming CUDA Compiling Conundrums: Issues with CUDA compilers like nvcc 11.5 throwing errors for operations in bfloat16 have been addressed in a fix proposal, aiming to support older GPUs and toolkits. Multi-GPU training hangs have also been discussed, linked to Issue #369, with a separate branch maintaining functionality.

LLaMa's Lean Learning: Discussions around memory efficiencies during LLaMa 2 70B model training highlighted configurations that allow for reduced memory usage. A tool named HTA was mentioned for pinpointing performance bottlenecks in PyTorch.

Post-training Peaks with Quantization: A YouTube video was shared, detailing the process and benefits of quantization in PyTorch.

GreenBitAI Goes Global: A toolkit called green-bit-llm was introduced for fine-tuning and inferencing GreenBitAI's language models. Attention was drawn to BitBlas for rapid 2-bit operation gemv kernels, along with a unique approach to calculating gradients captured in the GreenBitAI's toolkit.


Modular (Mojo 🔥) Discord

Tune in to Mojo Livestream for MAX 24.3 Updates: Modular's new livestream video titled "Modular Community Livestream - New in MAX 24.3" invites the community to explore the latest features of MAX Engine and Mojo, along with an introduction to the MAX Engine Extensibility API.

Community Projects Zoom Ahead: Noteworthy updates include NuMojo's improved performance and the introduction of Mimage for image parsing. The Basalt project also reached a milestone of 200 stars and released new documentation.

Mojo Compiler Evolves: Mojo compiler sees nightly updates with changes to better fit current practices, such as the move away from 80-column width and transitioning to types more suited for register passability.

AI Engineers Seek Don Hoffman's Consciousness Exploration: Interest in Donald Hoffman's work at UCI linked to consciousness research correlates with AI, as parallels are drawn between sensory data limitations seen in split-brain patients and AI hallucinations.

Mojo's Growing Ecosystem & Developer Guidance: Discussion on contribution processes to Mojo, inline with GitHub's pull request guidelines, and insights into the development workflow with tutorials on parameters demonstrate the active support for contributors to the rapidly expanding Mojo ecosystem.


HuggingFace Discord

Moondream and BLOOM Make Waves: The HuggingFace community has spotlighted new advancements including Moondream 2 batch processing and FLUENT's newest iteration, as well as tools for multilingual support. Particularly noteworthy is the BLOOM multilingual chat and AutoTrain's support for YAML configs, simplifying the training process for machine learning newcomers. Check out the community highlights.

When Audio Models Sing: There's interest in audio diffusion models for generative music with Whisper being fine-tuned for Filipino ASR, prompting discussions on optimization. However, a user faced challenges converting PyTorch models into TensorFlow Lite due to size limits.

AI's Frontline: Cybersecurity took center stage as the Hugging Face Twitter account was compromised, underlining the need for robust AI-related security. Members also exchanged GPU utilization tips for variance in training times between setups.

Visions of Quantum and AI Unions: In computer vision, the emphasis was on improving traditional methods like YOLO for gap detection in vehicle parts and adapting models like CLIP for image recognition with rotated objects. GhostNet's pre-trained weights were sought after, and CV members pondered the contemporary relevance of methods like SURF and SIFT.

Graph Gurus Gather: Recent papers on using LLMs with graph machine learning propose novel ways to integrate the two, with a paper](https://arxiv.org/abs/2404.19705) specifically teaching LLMs to retrieve information only when needed via the <RET> token. The reading group provided additional resources for those eager to learn more.

Showcasing Synthesis and Applied AI: From the #i-made-this section, there's the launch of tools like Podcastify and OpenGPTs-platform, along with models like shadow-clown-BioMistral-7B-DARE using mergekit.

NLPer's Quandaries and Queries: In NLP, a user offered compensation for custom training on Mistral-7B-instruct and concerns were raised about LLMs evaluating other LLMs. The GEMBA metric for translation quality using GPT 3.5+ was introduced, with a link provided to learn more.


OpenInterpreter Discord

Integrating OpenInterpreter with Groq LLM: Engineers discussed challenges with integrating Groq LLM onto Open Interpreter, highlighting issues such as uncontrollable output and erroneous file creation. The connection command shared was interpreter --api_base "https://api.groq.com/openai/v1" --api_key "YOUR_API_KEY_HERE" --model "llama3-70b-8192" -y --max_tokens 8192.

Microsoft Hackathon Seeks Open Interpreter Enthusiasts: A team is forming to participate in the Microsoft Open Source AI Hackathon utilizing Open Interpreter; the event promises to offer hands-on tutorials and the sign-up details are available here.

Open Interpreter Gets an iOS Reimagining: Discussions revolved around reimplementation of TMC protocol for iOS on Open Interpreter and troubleshooting issues with setting up with Azure Open AI models, with one member sharing a GitHub repository link for the iOS app in development here.

Local LLMs Challenge Developers: Personal testings on local LLMs like Phi-3-mini-128k-instruct were shared, indicating significant performance variances and calling out for better optimization methods in future implementations.

AI Vtuber's STT Conundrum: Implementing Speech-to-Text for AI powered virtual streamers brought up practical challenges, with engineers considering using trigger words and working towards AI-driven Twitch chat interactions through a separate LLM instance, aiming for comprehensive responses. For those tackling similar integrations, a member pointed to a main.py file on their GitHub as a resource.


Eleuther Discord


OpenRouter (Alex Atallah) Discord


LlamaIndex Discord

Boosting Agent Smarts: LlamaIndex 0.10.34 ushers in introspective agents capable of self-improvement through reflection mechanisms, detailed in a notebook which comes with a content warning for sensitive material.

Agentic RAG Gets an Upgrade: An informative video demonstrates the integration of LlamaParse + Firecrawl for crafting agentic RAG systems, and the release can be found through this link.

Trust-Scored RAG Responses: "Trustworthy Language Model" by @CleanlabAI introduces a scoring system for the trustworthiness of RAG responses, aiming to assure accuracy in generated content. For more insights, refer to their announcement here.

Local RAG Pipeline Handbook Hits Shelves: For developers seeking independence from cloud services, a manual for setting up a fully local RAG pipeline with LlamaIndex is unveiled, promising a deeper dive than quickstart guides and accessible here.

Hugging Face, Now Hugging LlamaIndex Tightly: LlamaIndex declares support for Hugging Face TGI, enabling optimal deployment of language models on Huggingface with enhanced features like function calling and improved latency. Shed light on TGI's new capabilities here.

Creating Conversant SQL Agents: AI engineers are contemplating the use of HyDE to craft NL-SQL bots for databases brimming with tables, eyeing ways to elevate the precision of SQL queries by the LLM; meanwhile, introspective agent methodologies are making waves, with further reading at Introspective Agents with LlamaIndex.


OpenAccess AI Collective (axolotl) Discord

Hermes 2 Pro Llama 3 Speed Test Results: Hermes 2 Pro Llama 3 has showcased impressive inference speed on an Android device with 8GB RAM, boosted by enhancements in llama.cpp.

Anime’s Role in AI Conversations: Members humorously discussed the rise of anime as it relates to increasing capabilities in AI question-answering and image generation tasks.

Gradio Customization Achievements: Adjustments in Gradio now allow dynamic configuration set through a YAML file, enabling the setting of privacy levels and server parameters programmatically.

Datasets for AI Training Spotlighted: A new dataset containing 143,327 verified Python examples (Python Dataset) and difficulties in improving mathematical performance of Llama3, even with math-centric datasets, were discussed, highlighting dataset challenges in AI training.

AI Training Platform Enhancements and Needs: There was a call to refine Axolotl's documentation, particularly regarding merging model weights and model inference, accessible at Axolotl Community Docs. Additionally, issues with gradient clipping configurations were addressed, and Phorm offered insights into customizing TrainingArguments for gradient clipping and the chatbot prompt.


Latent Space Discord


AI Stack Devs (Yoko Li) Discord


LAION Discord

CLIP vs. T5: The Model Smackdown: There's a spirited discussion about integrating CLIP and T5 encoders for training AI models; while the use of both encoders shows promise, some argue using T5 alone due to prompt adherence issues with CLIP.

Are Smaller Models the Big Deal?: In the realm of model size, enhancement of smaller models is being prioritized, as evidenced by the focus on the 400M DeepFloyd, with technical conversations touching upon the challenges in scaling up to 8B models.

Releasing SD3: Keep 'Em Waiting or Drop 'Em All?: The community's reaction to Stability AI's hinted gradual rollout of SD3 models—from small to large—was a mix of skepticism and eagerness, reflecting on whether this release strategy meets the community's anticipation.

LLama Embeds Strut into the Spotlight: Debates over the efficacy of using LLama embeds in model training emerged, with some members advocating for their use over T5 embeds, and sharing resources like the LaVi-Bridge to illustrate modern applications.

From Concept to Application: A Data Debate: The conversation dove into why synthetic datasets are favored in certain research over real-world datasets such as MNIST and ImageNet, alluding to the value of interpretability in AI methods and sharing resources like the StoryDiffusion website for insights.


LangChain AI Discord

Code Execution Finds an AI Buddy: Enthusiastic dialogues emerged around using AI to execute generated code, highlighting methods like Open Interpreter and developing custom tools such as CLITOOL. These discussions are pivotal for those crafting more interactive and automated systems.

Langchain Learns a New Language: The Langchain library's expansion into the Java ecosystem via langchain4j marks a crucial step for Java developers keen to harness AI assistant capabilities.

Langchain Gets a High-Performance Polish: The coupling of LangChain and Dragonfly has yielded impressive enhancements in chatbot context management, as depicted in a blog post detailing these advancements.

Decentralized Search Innovations: The community is buzzing with the development of a decentralized search feature for LangChain, promising to boost search functionalities with a user-owned index network. The work is showcased in a recent tweet.

Singularity Spaces with Llama & LangGraph: A contributor shared a video on Retrieval-Augmented Generation techniques without a vectorstore using Llama 3, while another enriches the dialogue with a comparison between LangGraph and LangChain Core in the execution realm.


tinygrad (George Hotz) Discord

Clojure Captures Engineer's Interest in Symbolic Programming: Engineers are discussing the ease of using Clojure for symbolic programming compared to Python, suggesting the use of bounties to ramp up on tinygrad, and debating the merits of Julia over Clojure in the ML/AI space.

tinygrad's UOps Puzzle Engineers: A call for proposals was made to reformat tinygrad's textual UOps representation to be more understandable, potentially resembling llvm IR, alongside an explanation that these UOps are indeed a form of Static Single Assignment (SSA).

Optimizing tinygrad for Qualcomm's GPU Playground: It was highlighted that tinygrad runs efficiently on Qualcomm GPUs by utilizing textures and pixel shaders, with the caveat that activating DSP support might complicate the process.

Single-threaded CPU Story in tinygrad: Confirmation from George Hotz himself that tinygrad operates single-threaded on the CPU side, with no threads bumping into each other.

Understanding tinygrad's Tensor Tango: A user's curiosity about the matmul function and transposing tensors spurred explanations, and another user shared their written breakdown on computing symbolic mean within tinygrad.


Mozilla AI Discord


DiscoResearch Discord

Mixtral Woes Spiral: The mixtral transformers hit a snag due to bugs impacting finetune performance; references include Twitter, Gist, and a closed GitHub PR. There's ambiguity whether the bug affects only training or generation as well, necessitating further scrutiny.

Quantized LLaMA-3 Takes a Hit: A Reddit post reveals quantization deteriorates LLaMA-3's performance notably compared to LLaMA-2, with a potentially enlightening arXiv study available. Meta's scaling strategy may account for LLaMA-3's precision reduction woes, while GitHub PR #6936 and Issue #7088 discuss potential fixes.

Meet the New Model on the Block: Conversations indicate 8x22b Mistral is being leveraged for current engineering tasks, though no performance metrics or usage specifics were disclosed.


Interconnects (Nathan Lambert) Discord


LLM Perf Enthusiasts AI Discord


Skunkworks AI Discord


Datasette - LLM (@SimonW) Discord


Cohere Discord


AI21 Labs (Jamba) Discord


Alignment Lab AI Discord


PART 2: Detailed by-Channel summaries and links

Unsloth AI (Daniel Han) ▷ #general (791 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (107 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (1215 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (80 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (3 messages):

Link mentioned: moondream/notebooks/Finetuning.ipynb at main · vikhyat/moondream: tiny vision language model. Contribute to vikhyat/moondream development by creating an account on GitHub.


OpenAI ▷ #ai-discussions (854 messages🔥🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (40 messages🔥):


OpenAI ▷ #prompt-engineering (30 messages🔥):


OpenAI ▷ #api-discussions (30 messages🔥):


Stability.ai (Stable Diffusion) ▷ #general-chat (919 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ctx-length-research (1 messages):

I apologize for the confusion, but as an AI, I do not have direct access to Discord servers, channels, or messages. Thus, I am unable to summarize the content from the Nous Research AI Discord channel named ctx-length-research. If you can provide the text from specific Discord messages that you'd like to be summarized, I'd be happy to assist you.


Nous Research AI ▷ #off-topic (20 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (47 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (717 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (60 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #rag-dataset (2 messages):

Link mentioned: Cynde/README.md at main · Neural-Dragon-AI/Cynde: A Framework For Intelligence Farming. Contribute to Neural-Dragon-AI/Cynde development by creating an account on GitHub.


Nous Research AI ▷ #world-sim (74 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):


Perplexity AI ▷ #general (814 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (43 messages🔥):

Link mentioned: Perplexity Wants To Help You Find Better Answers On The Internet | Forbes: Google Search or Wikipedia may be the go-to methods for finding out information on the Internet. Perplexity aims to help you go deeper to find concise answer...


Perplexity AI ▷ #pplx-api (59 messages🔥🔥):

Links mentioned:


LM Studio ▷ #💬-general (396 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (234 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (8 messages🔥):

Link mentioned: 👾 LM Studio - Discover and run local LLMs: Find, download, and experiment with local LLMs


LM Studio ▷ #📝-prompts-discussion-chat (8 messages🔥):


LM Studio ▷ #⚙-configs-discussion (56 messages🔥🔥):

Link mentioned: Udio | AI Music Generator - Official Website: Discover, create, and share music with the world. Use the latest technology to create AI music in seconds.


LM Studio ▷ #🎛-hardware-discussion (123 messages🔥🔥):

Links mentioned:


LM Studio ▷ #autogen (1 messages):


LM Studio ▷ #langchain (1 messages):

drjflamez: Secrets don't make friends


LM Studio ▷ #amd-rocm-tech-preview (28 messages🔥):

Links mentioned:


LM Studio ▷ #model-announcements (1 messages):


LM Studio ▷ #🛠-dev-chat (53 messages🔥):

Links mentioned:


CUDA MODE ▷ #general (15 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (15 messages🔥):

Links mentioned:


CUDA MODE ▷ #cuda (64 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #torch (19 messages🔥):

Links mentioned:


CUDA MODE ▷ #announcements (1 messages):


CUDA MODE ▷ #algorithms (5 messages):

Links mentioned:


CUDA MODE ▷ #beginner (9 messages🔥):


CUDA MODE ▷ #pmpp-book (4 messages):

Link mentioned: An Efficient Matrix Transpose in CUDA C/C++ | NVIDIA Technical Blog: My last CUDA C++ post covered the mechanics of using shared memory, including static and dynamic allocation. In this post I will show some of the performance gains achievable using shared memory.


CUDA MODE ▷ #youtube-recordings (6 messages):

Link mentioned: Optimizing GPU performance | Apple Developer Documentation: Find and address performance bottlenecks using the Metal debugger.


CUDA MODE ▷ #jax (1 messages):

Link mentioned: Using JAX in multi-host and multi-process environments — JAX documentation: no description found


CUDA MODE ▷ #off-topic (12 messages🔥):


CUDA MODE ▷ #hqq (4 messages):

Links mentioned:


CUDA MODE ▷ #llmdotc (630 messages🔥🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #general (102 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (1 messages):

ModularBot: From Modular: https://twitter.com/Modular/status/1786483510141657384


Modular (Mojo 🔥) ▷ #📺︱youtube (1 messages):

Link mentioned: Modular Community Livestream - New in MAX 24.3: MAX 24.3 is now available! Join us on our upcoming livestream as we discuss what’s new in MAX Engine and Mojo🔥 - preview of MAX Engine Extensibility API for...


Modular (Mojo 🔥) ▷ #ai (3 messages):


Modular (Mojo 🔥) ▷ #🔥mojo (172 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (22 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-blogs-vids (6 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #📰︱newsletter (1 messages):

Zapier: Modverse Weekly - Issue 32 https://www.modular.com/newsletters/modverse-weekly-32


Modular (Mojo 🔥) ▷ #nightly (92 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #announcements (2 messages):

<ul>
    <li><strong>Community Highlights Get an Update</strong>: Community highlight #56 introduces <a href="https://huggingface.co/spaces/Csplk/moondream2-batch-processing">Moondream 2 batch processing</a>, <a href="https://huggingface.co/spaces/fluently/Fluently-Playground">FluentlyXL v4</a>, Portuguese translation of HF Audio course's first chapters, <a href="https://huggingface.co/spaces/unography/image-captioning-with-longcap">BLIP fine-tune</a> for long captions, and many other projects. A comprehensive Portuguese list and retrospective of highlights is also available <a href="https://iatalk.ing/destaques-comunidade-hugging-face/">here</a>.</li>
    <li><strong>New Advances in AI Shared</strong>: Latest spaces feature <a href="https://huggingface.co/spaces/as-cle-bert/bloom-multilingual-chat">BLOOM multilingual chat</a>, an <a href="https://huggingface.co/spaces/tonyassi/inpainting-sdxl-sketch-pad">inpainting sketch pad</a>, and a link prediction <a href="https://github.com/Lama-West/PnPR-GCN_ACM_SAC_24/tree/main">repository</a>. Additionally, the HuggingFace alignment handbook task can now be run in the cloud with dstack as tweeted <a href="https://twitter.com/dstackai/status/1785315721578459402">here</a>.</li>
    <li><strong>Cool Stuff Unveiled by Community</strong>: A wide range of topics is covered from <a href="https://huggingface.co/blog/AmelieSchreiber/protein-optimization-and-design">protein optimization with Generative AI</a> to <a href="https://huggingface.co/blog/AviSoori1x/seemore-vision-language-model">implementing a Vision Language Model from scratch</a>. Also discussed is the Google Search with LLMs, Token Merging for fast LLM inference, and <a href="https://huggingface.co/blog/maywell/llm-feature-transfer">creating chat models with a single click</a>.</li>
    <li><strong>Cutting-edge Conversations</strong>: A reading group is scheduled to discuss recent progress and share insights, furthering the exchange of knowledge in the AI space. To join the next session, please check out this <a href="https://discord.com/events/879548962464493619/1234913780048203856">link</a>.</li>
    <li><strong>AutoTrain Configs Introduced</strong>: AutoTrain now supports yaml config files simplifying the model training process, even for those new to machine learning. An announcement about this new feature has been <a href="https://twitter.com/abhi1thakur/status/1786368641388179797">tweeted</a>, and the Github repository with example configs can be accessed <a href="https://github.com/huggingface/autotrain-advanced">here</a>.</li>
</ul>

Links mentioned:


HuggingFace ▷ #general (225 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (12 messages🔥):

Links mentioned:


HuggingFace ▷ #cool-finds (11 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (19 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (45 messages🔥):

Links mentioned:


HuggingFace ▷ #computer-vision (42 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (12 messages🔥):

Link mentioned: Large Language Models Are State-of-the-Art Evaluators of Translation Quality: Tom Kocmi, Christian Federmann. Proceedings of the 24th Annual Conference of the European Association for Machine Translation. 2023.


HuggingFace ▷ #diffusion-discussions (17 messages🔥):

Links mentioned:


OpenInterpreter ▷ #general (212 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (104 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (15 messages🔥):


Eleuther ▷ #general (113 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (165 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (9 messages🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (7 messages):


Eleuther ▷ #lm-thunderdome (3 messages):

Link mentioned: Paper page - Prometheus 2: An Open Source Language Model Specialized in Evaluating Other Language Models: no description found


OpenRouter (Alex Atallah) ▷ #announcements (3 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (3 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (248 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #blog (7 messages):

Link mentioned: Introspective Agents: Performing Tasks With Reflection - LlamaIndex: no description found


LlamaIndex ▷ #general (226 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (4 messages):


OpenAccess AI Collective (axolotl) ▷ #general (33 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (8 messages🔥):

Link mentioned: Gradio configuration parameters by marijnfs · Pull Request #1591 · OpenAccess-AI-Collective/axolotl: Various parameters of Gradio were hardcoded (e.g. share=True, ip address, port, number of tokens, temperature) I made them configurable here. Additionally the default tokens were overwritten into t...


OpenAccess AI Collective (axolotl) ▷ #general-help (8 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #datasets (39 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (27 messages🔥):

Links mentioned:


Latent Space ▷ #ai-in-action-club (95 messages🔥🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #ai-companion (6 messages):


AI Stack Devs (Yoko Li) ▷ #team-up (2 messages):


AI Stack Devs (Yoko Li) ▷ #ai-town-discuss (15 messages🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #ai-town-dev (61 messages🔥🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #local-ai-stack (1 messages):

Link mentioned: GitHub - get-convex/llama-farm-chat: Use locally-hosted LLMs to power your cloud-hosted webapp: Use locally-hosted LLMs to power your cloud-hosted webapp - get-convex/llama-farm-chat


AI Stack Devs (Yoko Li) ▷ #paper-spam (1 messages):

Deforum Daily Papers: Papers will now be sent to <#1227492197541220394>


AI Stack Devs (Yoko Li) ▷ #ai-raspberry-pi (1 messages):

jakekies: ??


LAION ▷ #general (59 messages🔥🔥):

Links mentioned:


LAION ▷ #research (5 messages):


LangChain AI ▷ #general (45 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (6 messages):

Links mentioned:


LangChain AI ▷ #tutorials (2 messages):

Links mentioned:


tinygrad (George Hotz) ▷ #general (17 messages🔥):


tinygrad (George Hotz) ▷ #learn-tinygrad (12 messages🔥):

Link mentioned: tinygrad-notes/symbolic-mean.md at main · mesozoic-egg/tinygrad-notes: Tutorials on tinygrad. Contribute to mesozoic-egg/tinygrad-notes development by creating an account on GitHub.


Mozilla AI ▷ #llamafile (25 messages🔥):

Links mentioned:


DiscoResearch ▷ #mixtral_implementation (7 messages):


DiscoResearch ▷ #general (3 messages):

Links mentioned:


DiscoResearch ▷ #discolm_german (3 messages):


Interconnects (Nathan Lambert) ▷ #news (3 messages):

Link mentioned: ElevenLabs Is Building an Army of Voice Clones: A tiny start-up has made some of the most convincing AI voices. Are its creators ready for the chaos they’re unleashing?


Interconnects (Nathan Lambert) ▷ #ml-drama (2 messages):

Link mentioned: Prometheus 2: An Open Source Language Model Specialized in Evaluating Other Language Models: Proprietary LMs such as GPT-4 are often employed to assess the quality of responses from various LMs. However, concerns including transparency, controllability, and affordability strongly motivate the...


Interconnects (Nathan Lambert) ▷ #random (2 messages):


Interconnects (Nathan Lambert) ▷ #rl (4 messages):


LLM Perf Enthusiasts AI ▷ #prompting (7 messages):


Skunkworks AI ▷ #datasets (1 messages):


Skunkworks AI ▷ #off-topic (2 messages):


Datasette - LLM (@SimonW) ▷ #llm (3 messages):


Cohere ▷ #collab-opps (2 messages):

Link mentioned: Founder Institute: World’s largest pre-seed startup accelerator.: no description found


AI21 Labs (Jamba) ▷ #jamba (2 messages):


Alignment Lab AI ▷ #general-chat (1 messages):