Frozen AI News archive

Anime pfp anon eclipses $10k A::B prompting challenge

**Victor Taelin** issued a $10k challenge to GPT models, initially achieving only **10% success** with state-of-the-art models, but community efforts surpassed **90% success** within 48 hours, highlighting GPT capabilities and common skill gaps. In Reddit AI communities, **Command R Plus (104B)** is running quantized on **M2 Max hardware** via **Ollama** and **llama.cpp** forks, with **GGUF quantizations** released on Huggingface. Streaming text-to-video generation is now available through the **st2v** GitHub repo. **WD Tagger v3** was released for mass auto-captioning datasets with a WebUI. Lesser-known prompting techniques like self-tagging and generational frameworks produced thought-provoking outputs in OpenAI discussions, including experiments with self-evolving system prompts. Stable Diffusion users discussed image composition importance for training character LoRAs and best checkpoints for video game character generation. Discussions also covered scarcity of **5B parameter models** and open(ish) licenses for open source AI. Memes included jokes about ChatGPT and Gemini training data differences.

Canonical issue URL

4 days ago, Victor Taelin confidently tweeted a simple A::B challenge for GPTs and then offered a $10k contest to prove him wrong:

image.png

His initial attempts with all SOTA models got 10% success rates. Community submissions got 56%. It took another day for @futuristfrog to surpass 90%. The challenge lasted 48 hours in total. A fun lesson in GPT capability, and another reminder that failure to do something in 2024 AI pre AGI is often a simple skill issue.


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence. Comment crawling still not implemented but coming soon.

Technical Developments and Releases

Techniques and Prompting

Questions and Discussions

Memes and Humor

AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

AI and Robotics Research Developments

AI Agents and Robotics

LLM Developments and Capabilities

Retrieval Augmented Generation (RAG) Architectures

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries

1. Quantization and Optimization Breakthroughs for LLMs

2. Expanding Context Lengths and Attention Mechanisms

3. Open-Source AI Advancements and Community Engagement

4. Multimodal AI Advancements and Applications

5. Misc


PART 1: High level Discord summaries

Perplexity AI Discord

iOS Users Test Drive New Story Discovery: Perplexity AI is trialing an innovative story discovery format on iOS. Users are encouraged to provide feedback on their experiences through a designated channel, and can download the test app here.

AI Event Ends in Harmony: The Perplexity AI Discord event wrapped up with both eun08000 and codelicious sharing first place. Prize recipients will receive direct messages with details.

Claude 3 Opus - A Model Debate: On the server, the talk revolved around observed variations in Perplexity's implementation of the Claude 3 Opus model compared to others, particularly regarding tasks demanding creativity.

API Quirks and Queries: Users noted inconsistencies between Perplexity's API and web application, with the API showing more hallucinations; the API's default model diverges from the web version. The 'sonar-medium-online' model is suggested for API users to closely mimic the Sonar model accessible via the web app for non-Pro users.

Tech Enthusiasts Share and Learn: Users exchanged information on a variety of topics from how AI affects the music industry to Tesla's and Apple's latest tech innovations. Additionally, a case study featuring Perplexity AI highlighted a 40% speed increase in model training powered by Amazon Web Services, demonstrating Perplexity's efficient utilization of advanced machine learning infrastructure and techniques.


Nous Research AI Discord


Stability.ai (Stable Diffusion) Discord


Unsloth AI (Daniel Han) Discord

HTML Conversion Leaves Engineers Blue: AI engineers discussed the limitations of current language models like Chat GPT and Claude in accurately converting images to HTML, leading to lost color fidelity and rounded borders. A tongue-in-cheek proposal suggested the use of ASCII art as an alternative, stemming from its ability to elicit responses from AI models as shown in this Ars Technica article.

Aurora-M Lights Up Possibilities: An open-source multilingual model, Aurora-M, boasting 15.5 billion parameters, was introduced and caught the community's attention with its cross-lingual safety capabilities, further detailed in this paper. The findings show that safety alignment in one language can have a positive impact on other languages.

Jamba Juice or Mamba Sluice? Investment Opinions Clash: Engineers debated the investment into AI21 Labs' Jamba, especially given their recent fundraising of $155 million as reported by TechCrunch. The return on investment (ROI) of focused model fine-tuning was brought to light, presenting an optimistic view despite the model's upfront costs.

AI Fine-Tuning Perspectives Merge and Diverge: The community engaged in a robust exchange on fine-tuning approaches, such as unsupervised fine-tuning techniques mentioned like GGUF, and the benefits of Dynamic Positional Offsets (DPO). Specific strategies for fine-tuning and the application of techniques like LoRA in enhancing performance were discussed.

Private AI Hosting Hustle: Data privacy concerns have led members to host their AI projects on personal servers, with anecdotes of using platforms like Hircoir TTS independently. Some envisioned future plans include integrating advertisements to capitalize on the growing portfolio of models.


LM Studio Discord

Boost Your Model’s Performance: The LM Studio appears to leap ahead of alternatives like oogabooga and Faraday with a GUI that wins user preference for its higher quality outputs. Suggestions poured in for expansions, notably for file reading support and modalities such as text-to-image and text-to-voice; such features edge closer to what Devin already offers and are angled towards enhancing creativity and productivity.

Big Thinkers, Bigger Models: A technical crowd advocates the power play of handling heavyweight models such as the Command R+, tipping the scales at 104B, and recommending brawnier hardware like the Nvidia P40 for older yet hefty models. Discussions around VRAM spill into strategies for optimizing multi-GPU setups, hinting at the use of both RTX 4060 Ti and GTX 1070 to spread the computational load, and leveraging Tesla P40 GPUs despite potential outdated CUDA woes.

The Joy of Smoothly Running Models: On both ROCM and ROCm Preview Beta fronts, GPU support discussion was rife, including the use of AMD’s RX 5000 and 6000 series chips. Users flagged the "exit 42" errors on ROCm 0.2.19 Beta, rallying around debug builds for a solution, displaying a communal spirit in action. Meanwhile, whispers of Intel’s Advanced Matrix Extensions (AMX) stirred speculation on how LM Studio could tap into such formidable processing prowess.

Excavating Model Gems: A surge in shared resources and models came through announcements, including Starling-LM 7B, c4ai command r v01, and stable-code-instruct-3b, among others. Accessibility stands upfront with a collective push towards a community page on Hugging Face, where the latest GGUF quants shine, luring AI enthusiasts to experiment with the offerings such as Google's Gemma 1.1 2B, and stay alert for the upcoming 7B variant.

Sculpting the Vision Models Landscape: A member's inquisition about training LLMs to decipher stock market OHLC patterns, amidst praise for LM Studio’s utility in vision model implementations, ignites a spark in exploring how the intricate dance between technology and finance could be choreographed with AI's grace. The revelation of vision models on Hugging Face mirrors the community’s camera-ready attitude to snapshot and subsequently transpose this conceptual aesthetic into practical applications.


HuggingFace Discord

Gradio's API Recorder and Chatbot UI Fixes Gear Up for Release: Gradio version 4.26.0 introduces an API Recorder to translate interactions into code and addresses crucial bugs related to page load times and chatbot UI crashes. The update is detailed in the Gradio Changelog.

A Crescendo of Concern Over LLMs: Security concerns gain spotlight as 'Crescendo', a new method that challenges the ethical restraints of LLMs, and vulnerabilities in Cohere's Command-R-plus are exposed. Meanwhile, Mixture-of-Depths (Modes) proposal and llamaindex blogs offer innovative solutions for model efficiency and information retrieval.

NLP Community Finesse with SageMaker, Desire for PDF ChatGPT, and Sails Through Challenges: The community debates deploying models on SageMaker, customizing ChatGPT for PDFs, and shares fascination over Gemini 1.5's 10M context window. Solution seekers confront multi-GPU training hiccups and demand token count information when using Hugging Face libraries.

Thriving Repository of AI Contributions and Dialogues: HybridAGI's neuro-symbolic behavior programming on GitHub welcomes peer review, and the Hugging Face reading group archives its collective wisdom on GitHub. PaperReplica's open-source invitation and RAG-enabled llamaindex shine as beacons of collaborative learning and resource sharing.

Vision and Beyond: Dialogues in the computer vision channel touch on the utility of HuggingFace as a model repository, efficacy of different Transformer models (e.g., XCLIP), and address real-time challenges using tools like the HuggingFace 'datasets' library for parquet file manipulation. Meanwhile, an open call for resources to apply diffusion models to video enhancement signifies the domain's vibrant investigative spirit.


Modular (Mojo 🔥) Discord

Mojo Rising: A Dive into Special Functions and SICP Adaptation

MAX Aligns with AWS; Open Source Documentation Drive

Discord Dynamics: Python Interop and Contributing to Mojo's Growth

Var vs. Let - the Mojo Parameter Saga

Nightly Chronicles: From CPython Interop to Community Discussions

Blog Beats and Video Treats in Mojo's Creative Continuum


Eleuther Discord


OpenAI Discord


OpenRouter (Alex Atallah) Discord

Claude 3 Takes on Images: The Claude 3 models have been updated to multimodal, now supporting image input, requiring developers to modify existing codebases accordingly.

AI Goes Old School with Rock, Paper, Scissors: A new game at blust.ai, where players can challenge ChatGPT to a classic round of Rock, Paper, Scissors.

Frontends and Favorites Front and Center: Engineers discussed various OpenRouter API frontends like LibreChat, SillyTavern, and Jan.ai. Command-R+ has emerged as a favored model for coding tasks and interactions in Turkish, while concerns are raised about content censorship in models.

Performance Insights in Modeling: Conversations highlighted that Sonnet outstrips Opus in coding tasks, and Claude 3 is superior in PDF data extraction compared to Gemini Pro 1.5, which prompted some skepticism about its utility.

Model Efficacy Metrics Spark Debate: The community has voiced that model ranking based solely on usage statistics might not accurately reflect a model's worth, suggesting spending or retention as potential alternate measures.


LlamaIndex Discord

Revving Up RAG Applications: Marker-Inc-Korea introduced AutoRAG, an automated tool for tuning RAG pipelines to enhance performance, detailed and linked in their tweet. Meanwhile, create-llama was released to streamline the launch of full-stack RAG/agent applications, as announced in its tweet.

Tweaking Sales Pitches with AI: A new application using RAG to create personalized sales emails was featured in a recent webinar, ditching hard-coded templates with an LLM-powered approach, further info available in a tweet.

Deep Diving Into Documents: Andy Singal presented on multi-document agents that handle complex QA across numerous sources. The aim is to expand this functionality for more intricate inquiries, shared in a presentation tweet.

Metadata to the Rescue for Document Queries: To get page numbers and document references from multi-document queries, make sure to include this metadata before indexing, allowing retrieval of detailed references post-query.

Optimization Overhaul for Azure and Embedding Times: Participants noted issues with Azure's OpenAI not recognizing context and discussed using batching methods for faster embedding generation. Regarding challenges with ReAct agents and open-source models like "llama2" and "mistral", better router descriptions may improve model-routing performance.


OpenInterpreter Discord

Mistral Needs Muscle: Mistral 7B Instruct v0.2 has been acknowledged as high-performing, yet it demands substantial resources—expect to allocate at least 16GB of RAM and have some GPU support for smooth operation.

Challenges with Python Compatibility: There's a community consensus to stick with Python <=3.10 to avoid issues with TTS packages, with repeated suggestions to avoid using Python 3.11.4 for setups dependent on voice command recognition.

A Call for Better Documentation: Inquiries about local vision models and calls highlighting the need for more comprehensive examples and documentation in the Open Interpreter's cookbook reveal gaps that are yet to be filled.

Efficiency Over Expense with Local Models: The costliness of GPT-4 has prompted discussions around leveraging local models such as Hermes 7B and Haiku—less expensive yet slightly less refined alternatives offering privacy and lower operating costs.

Hardware Hang-Ups and Software Setbacks: The O1 community reported hardware issues, particularly with external push-button integration, and software setup challenges when installing on Windows, with tweaks including using chocolatey, virtualenv, and specific environment variables being part of the troubleshooting dialogue.

Relevant resources and conversations are threaded throughout the community, with direct engagement on issues being tracked on platforms like GitHub.


LangChain AI Discord


LAION Discord

Apple's AI Ambitions Under Scrutiny: Apple is criticized for the subpar performance of Metal Performance Shaders (MPS) and torch compile, even as recent merges aim to fix MPS issues in the PyTorch nightly branch. Community experiences with torch.compile vary, reflecting ongoing optimizations needed for Apple's platforms.

Copyright Conundrum: AI's use of copyrighted content for creating derivative works sparks legal debate, with consensus on the insufficiency of paraphrasing to avoid infringement. The community anticipates the need for substantial legal changes to accommodate new AI training data practices.

The Harmony of AI-Composed Music: Discussions about AI-generated music, involving companies like Suno and Nvidia, recognized rapid advancements but also forecasted potential legal spats with the music industry. Members also noted the less impressive progress in text-to-speech (TTS) technology compared to AI's leap in music generation.

AI Career Dynamics Shifting: The rise of freelance AI-related careers due to technological progress is noted, with resources like Bloomberry's analysis cited. Stability AI's CosXL model release sparks conversations about the efficacy of EDM schedules and offset noise in model training.

Novelties in AI Research Techniques: A new paper on transformers shows computational resource allocation can be dynamic, DARE's pruning technique for language models hints at preservable capabilities, and BrushNet introduces enhanced AI inpainting. Latent diffusion for text generation, referenced from a NeurIPS paper, indicates a potential shift in generative model techniques.


Latent Space Discord


OpenAccess AI Collective (axolotl) Discord


Interconnects (Nathan Lambert) Discord

Podcasting Gold: John Schulman to Possibly Feature on Show: Nathan Lambert is considering featuring John Schulman in a podcast, a move that stirred excitement among members. Moreover, a licensing change for text-generation-inference (TGI) to Apache 2 has spurred a significant increase in contributors to the open-source project.

Memes Channel Maintains Light-Heartedness: The memes channel included joking references to targetings without context, improvements in experiences, and confirmation of employment status, indicating a casual, light-hearted discourse among members.

Open AI Weights Debate Hits Engaged Nerve: The #reads channel had a vibrant discussion on the societal impacts of open foundation models, with a focus on safety thresholds, regulation feasibility, and AI's potential to manipulate societal processes. A shared visualization of Transformer attention mechanisms and speculation about future models that emphasize verification instead of generation were among the in-depth topics discussed.

Bridging the Knowledge Gaps with Visuals: The #sp2024-history-of-open-alignment channel discussed effective resources like lmsys and alpacaeval leaderboard to find state-of-the-art models. Additionally, an intent to visually categorize models for better comprehension was expressed, along with sharing a live document (Google Slides presentation) for an upcoming alignment talk and a guide (comprehensive spreadsheet) on open models by Xeophon.

A Note on AI Generated Music: Nathan noted the impressive quality of a new contender in AI music generation, posing a potential challenge to the Suno AI platform.


CUDA MODE Discord


tinygrad (George Hotz) Discord

Tinygrad Takes a Step Back: George Hotz has reverted the command queue in tinygrad and is opting to integrate the memory scheduler directly with the current scheduler model. This approach utilizes the multidevicegraph abstraction already in place, as discussed here.

TinyJIT Under the Microscope: The TinyJit tutorial has been released, although it may contain inaccuracies, particularly with the apply_graph_to_jit function, and users are encouraged to submit pull requests for corrections TinyJit Tutorial.

Tinygrad Learning Expanded: A collection of tutorials and guides for contributing to tinygrad are now available with a focus on topics like multi GPU training Multi GPU Training Guide.

Discord Roles Reflect Contribution: George Hotz redesigned roles within the tinygrad Discord to better reflect community engagement and contribution levels, reinforcing the value of collaboration and respect for others' time.

Unpacking MEC's Firmware Mystery: Discussions about MEC firmware's opcode architectures emerged with speculation on RISC-V and different instruction sets, revealing a potential cbz instruction and inclusive dialogue around the nuances of RISC-V ISA.


Mozilla AI Discord

Scan Reveals Llamafile's Wrongful Accusation: Versions of llamafile, including llamafile-0.6.2.exe and llamafile-0.7, were flagged as malware by antivirus software; utilizing appeal forms with the respective antivirus companies was suggested as a remedial step.

Run Llamafile Smoother in Kaggle: Users encountering issues when running llamafile on Kaggle found solace through an updated command that resolves CUDA compilation and compatible GPU architecture concerns, enabling efficient usage of llamafile-0.7.

RAG-LLM Gets Local Legs: A query about locally distributing RAG-LLM application without the burdens of Docker or Python was answered affirmatively, indicating the suitability of llamafile for such purposes, particularly beneficial for macOS audiences.

Taming the Memory Beast with an Argument: An out of memory error experienced by a user was rectified by adjusting the -ngl parameter, demonstrating the importance of fine-tuning arguments based on the specific capabilities of their NVIDIA GeForce GTX 1050 card.

Vulkan Integration Spurs Performance Gains: A proposition to bolster llamafile by integrating Vulkan support led to performance enhancements on an Intel-based laptop with an integrated GPU, yet this required the granular task of re-importing and amending the llama.cpp file.


DiscoResearch Discord


AI21 Labs (Jamba) Discord


Datasette - LLM (@SimonW) Discord


Skunkworks AI Discord


LLM Perf Enthusiasts AI Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Perplexity AI ▷ #announcements (1 messages):


Perplexity AI ▷ #general (1199 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (40 messages🔥):

Links mentioned:


Perplexity AI ▷ #pplx-api (40 messages🔥):

Links mentioned:


Nous Research AI ▷ #off-topic (15 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (49 messages🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (148 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #project-obsidian (5 messages):

Links mentioned:


Nous Research AI ▷ #rag-dataset (19 messages🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (567 messages🔥🔥🔥):

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (977 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (341 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (78 messages🔥🔥):

Link mentioned: Home: 2-5X faster 70% less memory QLoRA & LoRA finetuning - unslothai/unsloth


Unsloth AI (Daniel Han) ▷ #help (374 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (2 messages):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (148 messages🔥🔥):

Links mentioned:


LM Studio ▷ #💬-general (488 messages🔥🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (103 messages🔥🔥):

Links mentioned:


LM Studio ▷ #announcements (1 messages):

Link mentioned: Tweet from LM Studio (@LMStudioAI): If you've been around these parts for long enough, you might be missing @TheBlokeAI as much as we do 🥲. Us & @bartowski1182 decided to try to help fill the void. We're excited to share the n...


LM Studio ▷ #🧠-feedback (25 messages🔥):

Links mentioned:


LM Studio ▷ #📝-prompts-discussion-chat (2 messages):


LM Studio ▷ #🎛-hardware-discussion (39 messages🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (30 messages🔥):

Links mentioned:


LM Studio ▷ #autogen (17 messages🔥):

Link mentioned: [Bug]: [autogenstudio] agent llm send max_tokens: null · Issue #2050 · microsoft/autogen: Describe the bug When max_tokens parameter is None, the agent send a frame /v1/chat/completions with max_tokens: null. In this case the LLM don't understand and and stop after the second token. St...


LM Studio ▷ #langchain (4 messages):

Link mentioned: Switching from open ai api to local LLM: Small follow up post on our last one about building a rag agent with langchain and node


LM Studio ▷ #amd-rocm-tech-preview (97 messages🔥🔥):

Links mentioned:


LM Studio ▷ #model-announcements (3 messages):


HuggingFace ▷ #announcements (4 messages):

Link mentioned: Aurora-M: The First Open Source Biden-Harris Executive Order Red teamed Multilingual Language Model: no description found


HuggingFace ▷ #general (372 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (2 messages):

Links mentioned:


HuggingFace ▷ #cool-finds (11 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (44 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (10 messages🔥):

Link mentioned: GitHub - isamu-isozaki/huggingface-reading-group: This repository's goal is to precompile all past presentations of the Huggingface reading group: This repository's goal is to precompile all past presentations of the Huggingface reading group - isamu-isozaki/huggingface-reading-group


HuggingFace ▷ #computer-vision (12 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (24 messages🔥):

Links mentioned:


HuggingFace ▷ #diffusion-discussions (9 messages🔥):

Link mentioned: Stanford CS224N: Natural Language Processing with Deep Learning | 2023: Natural language processing (NLP) is a crucial part of artificial intelligence (AI), modeling how people share information. In recent years, deep learning ap...


HuggingFace ▷ #gradio-announcements (1 messages):


Modular (Mojo 🔥) ▷ #general (34 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (7 messages):


Modular (Mojo 🔥) ▷ #✍︱blog (2 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #ai (1 messages):

rxzfn: There is a moveable product like this, but using pcie


Modular (Mojo 🔥) ▷ #tech-news (2 messages):


Modular (Mojo 🔥) ▷ #🔥mojo (336 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (18 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-blogs-vids (7 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #nightly (71 messages🔥🔥):

Links mentioned:


Eleuther ▷ #general (80 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (313 messages🔥🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (1 messages):

Link mentioned: GitHub - ndif-team/nnsight: The nnsight package enables interpreting and manipulating the internals of deep learned models.: The nnsight package enables interpreting and manipulating the internals of deep learned models. - ndif-team/nnsight


Eleuther ▷ #lm-thunderdome (83 messages🔥🔥):

Links mentioned:


OpenAI ▷ #ai-discussions (220 messages🔥🔥):

Link mentioned: Wow Really GIF - Wow Really - Discover & Share GIFs: Click to view the GIF


OpenAI ▷ #gpt-4-discussions (72 messages🔥🔥):


OpenAI ▷ #prompt-engineering (57 messages🔥🔥):


OpenAI ▷ #api-discussions (57 messages🔥🔥):


OpenRouter (Alex Atallah) ▷ #announcements (4 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):

Link mentioned: Rock, Paper, Scissors Game by Blust.AI: Play Rock, Paper, Scissors against ChatGPT. It’s easy to play and a fun way to see if you can outsmart an AI.


OpenRouter (Alex Atallah) ▷ #general (322 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #blog (8 messages🔥):


LlamaIndex ▷ #general (254 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):


OpenInterpreter ▷ #general (170 messages🔥🔥):

Link mentioned: Join the Open Interpreter Discord Server!: A new way to use computers | 8147 members


OpenInterpreter ▷ #O1 (71 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #general (190 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (45 messages🔥):

Links mentioned:


LAION ▷ #general (157 messages🔥🔥):

Links mentioned:


LAION ▷ #research (23 messages🔥):

Links mentioned:


Latent Space ▷ #ai-general-chat (88 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (8 messages🔥):

Links mentioned:


Latent Space ▷ #ai-in-action-club (57 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (53 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (19 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (5 messages):


OpenAccess AI Collective (axolotl) ▷ #datasets (1 messages):

faldore: <@&1166009801583628349> porn spam


OpenAccess AI Collective (axolotl) ▷ #docs (3 messages):

Link mentioned: axolotl/examples/llama-2/lisa.yml at main · OpenAccess-AI-Collective/axolotl: Go ahead and axolotl questions. Contribute to OpenAccess-AI-Collective/axolotl development by creating an account on GitHub.


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (46 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (22 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (15 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #memes (9 messages🔥):


Interconnects (Nathan Lambert) ▷ #reads (55 messages🔥🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #sp2024-history-of-open-alignment (31 messages🔥):

Links mentioned:


CUDA MODE ▷ #general (15 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (7 messages):


CUDA MODE ▷ #cuda (1 messages):

Link mentioned: Tweet from Andrej Karpathy (@karpathy): Have you ever wanted to train LLMs in pure C without 245MB of PyTorch and 107MB of cPython? No? Well now you can! With llm.c: https://github.com/karpathy/llm.c To start, implements GPT-2 training on ...


CUDA MODE ▷ #torch (14 messages🔥):

Links mentioned:


CUDA MODE ▷ #announcements (2 messages):


CUDA MODE ▷ #algorithms (7 messages):

Links mentioned:


CUDA MODE ▷ #suggestions (1 messages):


CUDA MODE ▷ #beginner (1 messages):


CUDA MODE ▷ #youtube-recordings (1 messages):

itali4no: https://youtu.be/ws7angQYIxI?si=PcRy7siLQuFywpgp


CUDA MODE ▷ #jax (1 messages):


CUDA MODE ▷ #ring-attention (8 messages🔥):

Links mentioned:


CUDA MODE ▷ #off-topic (3 messages):


CUDA MODE ▷ #hqq (20 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton-viz (23 messages🔥):


tinygrad (George Hotz) ▷ #general (59 messages🔥🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (6 messages):

Links mentioned:


Mozilla AI ▷ #llamafile (26 messages🔥):

Links mentioned:


DiscoResearch ▷ #general (9 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (5 messages):

Link mentioned: C4AI Command R Plus - a Hugging Face Space by CohereForAI: no description found


AI21 Labs (Jamba) ▷ #jamba (9 messages🔥):


Datasette - LLM (@SimonW) ▷ #ai (2 messages):

Link mentioned: Run a Private RAG ChatGPT on QNAP NAS: QNAP NAS platforms have the most unique and capable hardware designes in the category. We added a GPU to one and tested the AI capabilities.


Datasette - LLM (@SimonW) ▷ #llm (3 messages):

Links mentioned:


Skunkworks AI ▷ #general (1 messages):


Skunkworks AI ▷ #off-topic (2 messages):

Link mentioned: Instructor, Generating Structure from LLMs: Instructor makes it easy to reliably get structured data like JSON from Large Language Models (LLMs) like GPT-3.5, GPT-4, GPT-4-Vision, including open source...


LLM Perf Enthusiasts AI ▷ #claude (2 messages):

Link mentioned: Tweet from Joschka Braun (@JoschkaBraun): I benchmarked @AnthropicAI's new tool use beta API on the Berkeley function calling benchmark. Haiku beats GPT-4 Turbo in half of the scenarios. Results in 🧵 A huge thanks to @shishirpatil_, @fa...