Frozen AI News archive

Mixtral 8x22B Instruct sparks efficiency memes

**Mistral** released an instruct-tuned version of their **Mixtral 8x22B** model, notable for using only **39B active parameters** during inference, outperforming larger models and supporting **5 languages** with **64k context window** and math/code capabilities. The model is available on **Hugging Face** under an **Apache 2.0 license** for local use. **Google** plans to invest over **$100 billion** in AI, with other giants like **Microsoft**, **Intel**, and **SoftBank** also making large investments. The UK criminalized non-consensual deepfake porn, raising enforcement debates. A former **Nvidia** employee claims Nvidia's AI chip lead is unmatchable this decade. AI companions could become a **$1 billion** market. AI has surpassed humans on several basic tasks but lags on complex ones. **Zyphra** introduced **Zamba**, a novel 7B parameter hybrid model outperforming **LLaMA-2 7B** and **OLMo-7B** with less training data, trained on 128 H100 GPUs over 30 days. **GroundX** API advances retrieval-augmented generation accuracy.

Canonical issue URL

As is their established pattern, Mistral followed up their magnet link with a blogpost, and an instruct-tuned version of their 8x22B model:

image.png

the image ended up sparking some friendly competition between Databricks, Google, and AI21, all of which merely emphasized that Mixtral created a new tradeoff between active params and MMLU performance:

image.png

Of course, what is unsaid that the active params count doesnt linearly correlate with cost to run dense models, and that singular focus on MMLU isn't ideal for less scrupulous competitors.


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/Singularity. Comment crawling works now but has lots to improve!

AI Investments & Advancements

AI Assistants & Applications

AI Models & Architectures


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

Mixtral 8x22B Instruct Model Release

RAG (Retrieval-Augmented Generation) Advancements

Snowflake Arctic Embed Models

Misc


AI Discord Recap

A summary of Summaries of Summaries

Stable Diffusion 3 and Stable Diffusion 3 Turbo Launches:

Unsloth AI Developments:

WizardLM-2 Unveiling and Subsequent Takedown:

Other notable topics include:


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord

Stable Diffusion 3 Turbo Charges the Scene: Stability AI has introduced Stable Diffusion 3 and Stable Diffusion 3 Turbo, now available on their Developer Platform API, with claims of the fastest and most reliable performance, supported by Fireworks AI. Interested parties can get started with SD3 at Stable Diffusion 3 & Developer API, and an open generative AI approach is promised with plans for the model weights to be available for self-hosting for members.

Refining Visually Intuitive Generative AI: The SDXL finetunes have made the use of SDXL refiners nearly obsolete, as they are now prevalent in Civitai downloads, suggesting a trend towards integrated finetunes over separate refiner modules, reflecting a community-driven optimization.

Model Merging Explored: There is lively discussion on model merging tactics within ComfyUI, grappling with complex mechanisms such as V-prediction and epsilon, highlighting the community's experimentation with these methods to achieve enhanced outcomes, yet acknowledging that correct implementations are crucial to prevent unpredictable results.

Navigating Diffusers Library Limitations: A conversation emerged around the limitations and dependencies in the diffusers pipeline, with a focus on Stable Video Diffusion Pipeline challenges. Despite these challenges, some users are optimizing usage by running models independently post-download, bypassing certain Hugging Face library constraints.

Awaiting SD3’s Membership Model Details: The community is keenly waiting for Stability AI to provide clarifications on Stable Diffusion 3 licensing for personal versus commercial use, especially in light of the new membership model revealed for accessing self-hosted weights.


Unsloth AI (Daniel Han) Discord

GPT-4 Gains Over GPT-3.5: The new iteration of GPT, GPT-4, is regarded as a fine-tuned enhancement over GPT-3.5, though specifics on performance metrics or features were not provided.

Mistral7B Shines in Multilingualism: Members conferred about the multilingual capabilities of the Mistral7B model, recommending the inclusion of diverse language data in training sets, particularly French, to improve performance.

Unsloth AI Gets Help from Fans: There’s a tangibly positive response from the community towards Unsloth AI, with users keen to help with documentation, expansion, and even considering donations. The Mixtral 8x22B model's release under Apache 2.0 was met with excitement for its promise in multilingual fluency and handling of extensive context windows.

Chroma Goes Go: The Chroma project leaps forward with an edge version written in Go, which utilizes SQLite and WASM for browser-based applications, now available on GitHub.

Mobile AI Deployment Discussed: The complexity of deploying AI models on mobile devices surfaced, noting challenges such as the absence of CUDA and the infeasibility of running standard Deep Learning Python codes on such platforms.


LM Studio Discord

AI Assistance for NeoScript Programming: A user looking for help with NeoScript programming expressed challenges in configuring AI models. Microsoft's new release, WaveCoder Ultra 6.7b, excels in code translation and could be a strong candidate for this task.

Solving AI's Echo Chamber: To combat repetitive AI responses, particularly in Dolphin 2 Mistral, members discussed strategies such as fine-tuning models and leveraging multi-turn conversation frameworks outlined in Azure's article.

Introducing the WizardLM-2 League: The debut of WizardLM-2 models sparked discussions about performance. Compatibility with existing tools, including the importance of using GGUF quants and version 0.2.19 or newer for proper functionality, was emphasized.

Tech Wizards at Play: One user successfully enabled direct communication between four 3090 GPUs, improving model performance by bypassing CPU/RAM. There was also chatter about the challenges of signing Windows executables, with a hint that the Windows versions are indeed signed with an Authenticode cert.

Quantization Conundrum and Model Preferences: Mixed reviews on quantization levels, from Q8 to Q6K, pointed to a preference for models with higher quantization levels when VRAM is sufficient. For large models, such as WizardLM-2-8x22B, GPUs like the 4090 with 24GB VRAM may be inadequate.


Nous Research AI Discord


Perplexity AI Discord

Robots Debating Their Roots: Engineers exchanged insights on the performance nuances of AI models including GPT-4 and Claude 3 Opus, with a shared sentiment that GPT-4 may exhibit "lazy" tendencies in real-world applications. The open-source Mixtral's 8x22B model is highlighted for its impressive capabilities, sparking debates on model efficacy.

Stumped by Stubborn Software Issues: A conversation was noted about achieving consistency between the web client and the API, with specific attention to parameters like temperature settings. Engineers are also discussing the benefits of including a rate limit counter in the API response for better management and transparency.

The Vanishing Messages Mystery: Concern was voiced over changes in the Perplexity API's payment method management, particularly the opacity surrounding the remaining message counts for pro users. This focus on transparency indicates professionals need clarity to manage resources efficiently.

A Tale of Truncated Tokens: Technical dialogue included challenges faced when engaging models with large context sizes, like a 42k token prompt, and the tendency for models to summarize rather than dive deep into lengthy documents. This could be pivotal as engineers optimize models to process complex prompts fully.

The Search for Smarter Searches: Members also discussed using site:URL search operators for more targeted information retrieval. Additionally, there is a call for better communication regarding rate limits in the API, including the possibility of a 429 response.


LAION Discord


OpenRouter (Alex Atallah) Discord

OpenRouter Welcomes WizardLM Raptors: OpenRouter announced the release of WizardLM-2 7B and a price drop for WizardLM-2 8x22B to $0.65/M tokens. The WizardLM-2 8x22B Nitro boasts over 100 transactions per second post its database restart.

Latency Labyrinth Resolved: Latency issues on various models such as Mistral 7B Instruct and Mixtral 8x7B Instruct were attributed to cloud provider DDoS protection, with updates concerning the resolution found in the associated discussion thread.

Calling All Frontend Mavericks: A member seeks web development assistance for an AI-based frontend project for OpenRouter, specifically emphasizing role-playing novel mode and conversation style systems. Ability to distinguish AI-generated text from user input is also requested.

AI Model Morality and Multilingual Mastery: Vigorous exchanges regarding both censorship protocols for NSFW content and the imperative for enhancing models' multilingual performance took place. Members looked forward to direct endpoints and new provider integrations for an anticipated AI model release.

Bitrate Bits and Quality Quibbles: Users showed a clear preference for a minimum of 5 bits per word (bpw) for model quantization, noting that reductions below this threshold notably compromise quality. Discussions underscored the trade-offs between efficient operation and maintaining high fidelity in AI outputs.


Modular (Mojo 🔥) Discord


CUDA MODE Discord

PyTorch Resource Debate: While discussing if "Deep Learning with PyTorch" is a relevant resource despite being 4 years old, members noted that the PyTorch core has remained stable, though significant updates have occurred in the compiler and distributed systems. A member shared a teaser for an upcoming edition of the book, which would include coverage of transformers and Large Language Models.

CUDA Custom GEMM Sparking Interest: The conversation involved improving GEMM performance in CUDA, with one member providing a new implementation that outperformed PyTorch's function on specific benchmarks, sharing their code on GitHub. However, another highlighted JIT compilation issues with torch.compile. The group also discussed optimal block size parameters, referencing a related code example on Gist.

Next-Gen Video Analysis & Robotics Gains Screenshare: Members shared links about Augmend's video processing features, which combine OCR and image segmentation, previewed on wip.augmend.us, and the full service to be hosted on augmend.com. Another highlight was Boston Dynamics' unveiling of a fully electric robot named Atlas intending for real-world applications, showcased in their All New Atlas | Boston Dynamics video.

Bridging the CUDA Toolkit Knowledge Gap: In the #beginner channel, members discussed issues related to using the CUDA toolkit on WSL, with one user facing problems running the ncu profiler. The community provided troubleshooting steps and stressed the importance of setting the correct CUDA path in environment variables. There was also an advisory that Windows 11 might be necessary for effective CUDA profiling on WSL 2, with one user providing a guide on the subject.

Quantization Dilemmas and Solutions in Air: A thorough chat occurred on the topic of quantization axes in GPT models with a highlight on the complexities when using axis=0. Participants suggested quantizing Q, K, and V separately with references to Triton kernels and an autograd optimization method for boosting speed and performance. Their debate continued with discussions of 2/3 bits quantization practicality and was supplemented with implementation details and benchmarks on GitHub.

Optimizing ML Model Performance: A GitHub notebook for extending PyTorch with CUDA Python garnered attention for speed enhancements but with a need for more optimization to fully tap into tensor core capabilities, as shared in the notebook's link. Additionally, there were mentions of optimizing the softmax function and block sizes for cache utilization, with insights shared through a GitHub pull request.


OpenAI Discord

Multiplayer GPT Headed for the Gaming Galaxy: Engineers discussed the potential of integrating GPT-Vision and camera inputs for a real-time gaming assistant to tackle multiple-choice games. The possibility of utilizing Azure or virtual machines to handle intensive computational tasks was raised, alongside leveraging TensorFlow or OpenCV for system management.

AI Versus Human Conundrum Continues: A philosophical debate emerged concerning the differences between AI and human cognition, discussing the prospects of AI acquiring human-like reasoning and emotions, and the role of quantum computing in this evolution.

The Quest for Knowledge Enhancements: Members sought information on how to prepare a knowledge base for custom GPT applications and questioned the arrival of the Whisper v3 API. The noted limitations such as GPT-4's token memory span being speculated to have shrunk triggered calls for improved clarity on API capabilities.

Creative Minds Favor Claude and Gemini: When tackling literature reviews and fictional works, AI aficionados recommended using models like Claude and Gemini 1.5. These tools were favored for their prowess in handling literary tasks and creative writing respectively.

Discord Channel Dynamics: Two channels, prompt-engineering and api-discussions, experienced a notable decrease in activity, with participants attributing the quiet to possible over-moderation and a recent string of timeouts, including a specific 5-month timeout case involving assistance to another user.


LlamaIndex Discord


Eleuther Discord


HuggingFace Discord


OpenAccess AI Collective (axolotl) Discord

Idefics2 Brings Multimodal Flair: The new multimodal model Idefics2 has been introduced, capable of processing both text and images with improved OCR and visual reasoning skills. It is offered in both base and fine-tuned forms and is under the Apache 2.0 license.

RTX 5090 Speculation Stokes Anticipation: NVidia is rumored to be considering an expedited release of the RTX 5090, potentially at Computex 2024, to stay ahead of AMD's advances, sparking discussions on hardware suitability for cutting-edge AI models.

Model Training Finetuning: Engineers shared insights on model training configurations, focusing on the 'train_on_input' parameter in loss calculation, and suggested using "TinyLlama-1.1B-Chat-v1.0" for fine-tuning small models for efficient experimentation.

Phorm AI Becomes Go-To Resource: Members referred to Phorm AI for various inquiries, including epoch-wise saving techniques and data preparation for models like TinyLlama for tasks like text-to-color code predictions.

Spam Flood Triggers Alerts: Multiple channels within the community were targeted by spam messages promoting OnlyFans content, attempting to divert attention from the AI-centric conversations and technical discourse.


Latent Space Discord

LLM Ranking Resource Revealed: A comprehensive website, LLM Explorer, has been shared, showcasing a plethora of open-source language models, each assessed through ELO scores, HuggingFace leaderboard ranks, and task-specific accuracy metrics, serving as a valuable resource for model comparison and selection.

AI+Human Symphony in the Gig Economy: The launch of Payman AI, a platform facilitating AI agents to remunerate humans for tasks beyond AI capabilities, has sparked interest; the concept promotes a cooperative ecosystem between AI and human talents in domains like design and legal services.

Supabase Embraces AI Inference: Supabase introduces a simple API for running AI inferences within its edge functions, allowing AI models such as gte-small to be employed directly in databases, as detailed in their announcement.

Buzz Around "Llama 3" and OpenAI API Moves: The AI community is abuzz about the mysterious "Llama 3" speculated to debut at a London hackathon, and OpenAI's Assistants API enhancements are drawing attention in light of a potential GPT-5 release, stirring debates about possible impacts on AI startups and platforms.

BloombergGPT Paper Club Session Goes Zoom: The LLM Paper Club invites engineers to a Zoom session on BloombergGPT, due to prior challenges with Discord screensharing, and the discussion has pivoted to Zoom for a better sharing experience. Participants can register for the event here, and further reminders to join the discussions are being circulated within the community.


OpenInterpreter Discord


Interconnects (Nathan Lambert) Discord

Big Win for qwen-1.5-0.5B: The qwen-1.5-0.5B model's winrate soared from 4% to 32% against heavyweights like AlpacaEval using generation in chunks. This approach, along with a 300M reward model, may be a game-changer in output searching.

How To Win Friends and Influence AIs: The recently unveiled Mixtral 8x22B, a polyglot SMoE model, is sharing the limelight owing to its impressive capabilities and the Apache 2.0 open license. Meanwhile, the rise of OLMo 1.7 7B indicates a notable stride in language model science with a robust performance leap on the MMLU benchmark.

Replicating Chinchilla: An Anomaly: Discrepancies in replicating the Chinchilla scaling paper by Hoffmann et al. have cast doubts around the paper's findings. The community's reaction ranged from confusion to concern, signaling an escalating drama around the challenge of scaling law verification.

Lighthearted Anticipation and Rumination: With playful banter on potential showdowns in olmo vs llama, community members show humor in competition. Moreover, Nathan Lambert teases the guild with a forecast of content deluge, signaling a possibly intense week of knowledge sharing.

Model Madness or Jocularity?: A side comment in an underpopulated channel by Nathan mentioned a potential tease involving WizardLM 2 as a troll, showing a blend of humor and light-heartedness amidst technical discussions.


Cohere Discord


LangChain AI Discord

AI Documentation Gets Facelift: In an effort to improve usability, contributors to the LangChain documentation are revamping its structure, introducing categories like 'tutorial', 'how to guides', and 'conceptual guide'. A member shared the LangChain introduction page, emphasizing LangChain's components such as building blocks, LangSmith, and LangServe, which aid in the development and deployment of applications with large language models.

Building with LangChain — An Expressive Endeavor?: Within the #general channel, a member sought advice on YC startup applications while drawing parallels to Extensiv, leading to the mention of several entities like Unsloth, Mistral AI, and Lumini. Simultaneously, challenges with LangServe integration when combined with Nemo Guardrails were highlighted due to Nemo's transformation of output structures.

Forge Ahead with New AI Tools and Services: GalaxyAI's debut of an API service with complimentary access to GPT-4 and GPT-3.5-turbo stirred up interest, showcased at Galaxy AI. Similarly, OppyDev’s fusion of an IDE and a chat client received attention, advocating an improved coding platform accessible at OppyDev AI. Meanwhile, Rubiks.ai appealed to tech enthusiasts to beta test their search engine and assistant at Rubiks.ai using code RUBIX.

AI Pioneers Share Educational Resources and Seek Collaboration: A member from #tutorials posted a YouTube tutorial on granting AI agents with long-term memory, igniting a discussion why 'langgraph' wasn't employed. Furthermore, a participant expressed eagerness to collaborate on new projects, inviting others to connect through direct messaging.

Diverse Dialogues on Data and Optimization: In a lively exchange, strategies for optimizing RAG (Retrieval-Augmented Generation) with large documents were evaluated, including document splitting. Members also dialogued over the best methods to manipulate CSV files with Langchain, suggesting improvements for chatbots and data processing.


DiscoResearch Discord


tinygrad (George Hotz) Discord

Int8 Integration in Tinygrad: Tinygrad has been confirmed to support INT8 computations, with recognition that such data type support often depends more on hardware capabilities rather than the software design itself.

Graph Nirvana with Tiny-tools: For enhanced graph visualizations in Tinygrad, users can visit Tiny-tools Graph Visualization to create slicker graphs than the basic GRAPH=1 setting.

Pytorch-Lightning's Hardware Adaptability: Discussions about Pytorch-Lightning touched on its hardware-agnostic capabilities, with practical applications noted on hardware like the 7900xtx. Discover Pytorch-Lightning on GitHub.

Tinygrad Meets Metal: Community members are exploring the generation of Metal compute shaders with tinygrad, discussing how to run simple Metal programs without Xcode and the possibility of applying this to meshnet models.

Model Manipulation and Efficiency in Tinygrad: A member's proposal for a fast, probabilistically complete Node.equals() prompted discussions on efficiency, while George Hotz explained layer device allocation, and users were directed toward tinygrad/shape/shapetracker.py or view.py for zero-cost tensor manipulations like broadcast and reshape.


Skunkworks AI Discord


Datasette - LLM (@SimonW) Discord


Alignment Lab AI Discord


Mozilla AI Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ▷ #announcements (1 messages):

Link mentioned: Stable Diffusion 3 API Now Available — Stability AI: We are pleased to announce the availability of Stable Diffusion 3 and Stable Diffusion 3 Turbo on the Stability AI Developer Platform API. 


Stability.ai (Stable Diffusion) ▷ #general-chat (1039 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (383 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (27 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (275 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (46 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (15 messages🔥):


LM Studio ▷ #💬-general (175 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (96 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (4 messages):


LM Studio ▷ #📝-prompts-discussion-chat (17 messages🔥):


LM Studio ▷ #🎛-hardware-discussion (18 messages🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (31 messages🔥):


LM Studio ▷ #amd-rocm-tech-preview (19 messages🔥):

Links mentioned:


LM Studio ▷ #model-announcements (3 messages):

Link mentioned: lmstudio-community/wavecoder-ultra-6.7b-GGUF · Hugging Face: no description found


Nous Research AI ▷ #off-topic (17 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (7 messages):

Links mentioned:


Nous Research AI ▷ #general (159 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (7 messages):


Nous Research AI ▷ #rag-dataset (10 messages🔥):

Link mentioned: The Normal Blog - Infinite Context LLMs: Going Beyond RAG with Extended Minds: In this blog we discuss how the transformer architecture naturally extends over external memories, and share empirical results which leverage this capability to succeed where RAG has struggled. These ...


Nous Research AI ▷ #world-sim (159 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #general (286 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (9 messages🔥):


Perplexity AI ▷ #pplx-api (4 messages):


LAION ▷ #general (285 messages🔥🔥):

Links mentioned:


LAION ▷ #research (13 messages🔥):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #announcements (5 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (4 messages):


OpenRouter (Alex Atallah) ▷ #general (271 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #general (67 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (1 messages):

ModularBot: From Modular: https://twitter.com/Modular/status/1780676643176231240


Modular (Mojo 🔥) ▷ #ai (2 messages):

Link mentioned: Unlock AI Agent real power?! Long term memory & Self improving: How to build Long term memory & Self improving ability into your AI Agent?Use AI Slide deck builder Gamma for free: https://gamma.app/?utm_source=youtube&utm...


Modular (Mojo 🔥) ▷ #🔥mojo (136 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (10 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-blogs-vids (1 messages):

Link mentioned: Maxim Zaks - Is Mojo just a hype?: no description found


Modular (Mojo 🔥) ▷ #📰︱newsletter (1 messages):

Zapier: Modverse Weekly - Issue 30 https://www.modular.com/newsletters/modverse-weekly-30


Modular (Mojo 🔥) ▷ #🏎engine (1 messages):

There was only one message provided with no mention of any discussion points, topics, or links to summarize. If you would like a summary of a more extensive conversation or a specific topic within the 🏎engine channel, please provide the relevant messages.


Modular (Mojo 🔥) ▷ #nightly (21 messages🔥):

Links mentioned:


CUDA MODE ▷ #general (11 messages🔥):

Link mentioned: Deep Learning with PyTorch, Second Edition: Everything you need to create neural networks with PyTorch, including Large Language and diffusion models.</b>

Deep Learning with PyTorch, Second Edition</i> updates the bestselling ori...


CUDA MODE ▷ #cuda (20 messages🔥):

Links mentioned:


CUDA MODE ▷ #torch (2 messages):


CUDA MODE ▷ #cool-links (2 messages):

Link mentioned: All New Atlas | Boston Dynamics: We are unveiling the next generation of humanoid robots—a fully electric Atlas robot designed for real-world applications. The new Atlas builds on decades of...


CUDA MODE ▷ #beginner (43 messages🔥):

Links mentioned:


CUDA MODE ▷ #youtube-recordings (1 messages):

marksaroufim: https://www.youtube.com/watch?v=DdTsX6DQk24


CUDA MODE ▷ #ring-attention (5 messages):


CUDA MODE ▷ #hqq (36 messages🔥):

Links mentioned:


CUDA MODE ▷ #llmdotc (76 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #massively-parallel-crew (14 messages🔥):


OpenAI ▷ #ai-discussions (167 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (7 messages):


OpenAI ▷ #prompt-engineering (5 messages):


OpenAI ▷ #api-discussions (5 messages):


LlamaIndex ▷ #blog (3 messages):

Link mentioned: MistralAI Cookbook - LlamaIndex: no description found


LlamaIndex ▷ #general (164 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (2 messages):


Eleuther ▷ #general (58 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (78 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (5 messages):


Eleuther ▷ #lm-thunderdome (21 messages🔥):

Link mentioned: Tweet from Aran Komatsuzaki (@arankomatsuzaki): Compression Represents Intelligence Linearly LLMs' intelligence – reflected by average benchmark scores – almost linearly correlates with their ability to compress external text corpora repo: ht...


Eleuther ▷ #multimodal-general (1 messages):

Links mentioned:


HuggingFace ▷ #announcements (10 messages🔥):

Links mentioned:


HuggingFace ▷ #general (85 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (3 messages):

Link mentioned: ML Blog - Improve ChatGPT with Knowledge Graphs: Leveraging knowledge graphs for LLMs using LangChain


HuggingFace ▷ #cool-finds (7 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (19 messages🔥):

Links mentioned:


HuggingFace ▷ #computer-vision (11 messages🔥):

Link mentioned: GitHub - lukas-blecher/LaTeX-OCR: pix2tex: Using a ViT to convert images of equations into LaTeX code.: pix2tex: Using a ViT to convert images of equations into LaTeX code. - lukas-blecher/LaTeX-OCR


HuggingFace ▷ #NLP (17 messages🔥):

Links mentioned:


HuggingFace ▷ #diffusion-discussions (8 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #general (44 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (2 messages):


OpenAccess AI Collective (axolotl) ▷ #other-llms (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #manticore (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #general-help (13 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #datasets (3 messages):


OpenAccess AI Collective (axolotl) ▷ #rlhf (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #hippogriff (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #minotaur (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #bots (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #community-showcase (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #runpod-help (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #deployment-help (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #docs (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #shearedmistral (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #replicate-help (1 messages):

aquash1553: @everyone Best OnlyFans Leaks & Teen Content 🍑 🔞 discord.gg/s3xygirlsss


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (36 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (5 messages):

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


Latent Space ▷ #ai-general-chat (68 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

Link mentioned: LLM Paper Club (BloombergGPT / TimeGPT paper) · Zoom · Luma: This week @yikes will be covering BloombergGPT: https://arxiv.org/abs/2303.17564 Also submit and vote for our next paper:…


Latent Space ▷ #llm-paper-club-west (19 messages🔥):

Link mentioned: Join our Cloud HD Video Meeting: Zoom is the leader in modern enterprise video communications, with an easy, reliable cloud platform for video and audio conferencing, chat, and webinars across mobile, desktop, and room systems. Zoom ...


OpenInterpreter ▷ #general (59 messages🔥🔥):

Link mentioned: The Worst Product I've Ever Reviewed... For Now: The Humane AI pin is... bad. Almost no one should buy it. Yet.MKBHD Merch: http://shop.MKBHD.comTech I'm using right now: https://www.amazon.com/shop/MKBHDIn...


OpenInterpreter ▷ #O1 (17 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ideas-and-feedback (11 messages🔥):


Interconnects (Nathan Lambert) ▷ #news (17 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (9 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (1 messages):

natolambert: shittiest leaderboard winner lol


Interconnects (Nathan Lambert) ▷ #random (23 messages🔥):


Interconnects (Nathan Lambert) ▷ #reads (3 messages):

Link mentioned: Beavis and Butt-Head - SNL: A NewsNation livestream event on AI is derailed by two audience members (Ryan Gosling, Mikey Day).Saturday Night Live. Stream now on Peacock: https://pck.tv/...


Interconnects (Nathan Lambert) ▷ #sp2024-history-of-open-alignment (1 messages):

natolambert: should I wizardLM 2 as a troll lol


Cohere ▷ #general (54 messages🔥):

Links mentioned:


Cohere ▷ #project-sharing (3 messages):

Links mentioned:


LangChain AI ▷ #announcements (1 messages):

Link mentioned: Introduction | 🦜️🔗 LangChain: LangChain is a framework for developing applications powered by large language models (LLMs).


LangChain AI ▷ #general (38 messages🔥):

Links mentioned:


LangChain AI ▷ #langserve (1 messages):


LangChain AI ▷ #share-your-work (4 messages):

Links mentioned:


LangChain AI ▷ #tutorials (5 messages):

Link mentioned: Unlock AI Agent real power?! Long term memory & Self improving: How to build Long term memory & Self improving ability into your AI Agent?Use AI Slide deck builder Gamma for free: https://gamma.app/?utm_source=youtube&utm...


DiscoResearch ▷ #mixtral_implementation (12 messages🔥):

Link mentioned: mistral-community/Mixtral-8x22B-v0.1 · [AUTOMATED] Model Memory Requirements: no description found


DiscoResearch ▷ #general (8 messages🔥):

Links mentioned:


DiscoResearch ▷ #benchmark_dev (1 messages):

Link mentioned: Reddit - Dive into anything: no description found


tinygrad (George Hotz) ▷ #general (9 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (9 messages🔥):


Skunkworks AI ▷ #off-topic (4 messages):

Links mentioned:


Datasette - LLM (@SimonW) ▷ #llm (3 messages):

Links mentioned:


Alignment Lab AI ▷ #oo (2 messages):


Mozilla AI ▷ #llamafile (2 messages):