Frozen AI News archive

Llama-3-70b is GPT-4-level Open Model

**Meta** has released **Llama 3**, their most capable open large language model with **8B and 70B parameter versions** supporting **8K context length** and outperforming previous models including **Llama 2** and **Mistral 7B**. **Groq** serves the **Llama 3 70B** model at **500-800 tokens/second**, making it the fastest GPT-4-level token source. Discussions highlight AI scaling challenges with **Elon Musk** stating that training **Grok 3** will require **100,000 Nvidia H100 GPUs**, and **AWS** planning to acquire **20,000 B200 GPUs** for a **27 trillion parameter model**. Microsoft unveiled **VASA-1** for lifelike talking face generation, while **Stable Diffusion 3** and its extensions received mixed impressions. Concerns about AI energy usage and political bias in AI were also discussed.

Canonical issue URL

With a sample size of 1600 votes, the early results from Lmsys were even better than reported benchmarks suggested, which is rare these days:

image.png

This is the first open model to beat Opus, which itself was the first model to briefly beat GPT4 Turbo. Of course this may drift over time, but things bode very well for Llama-3-400b when it drops.

Already Groq is serving the 70b model at 500-800 tok/s, which makes Llama 3 the hands down fastest GPT-4-level token source period.

With recent replication results on Chinchilla coming under some scrutiny (don't miss Susan Zhang banger, acknowledged by Chinchilla coauthor), Llama 2 and 3 (and Mistral, to a less open extent) have pretty conclusively consigned Chinchilla laws to the dustbin of history.


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/Singularity. Comment crawling works now but has lots to improve!

Meta's Llama 3 Release and Capabilities

Image/Video AI Progress and Stable Diffusion 3

AI Scaling Challenges and Compute Requirements

AI Safety, Bias and Societal Impact Discussions

AI Memes and Humor


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

Meta Llama 3 Release

Open Source AI Landscape

Ethical and Societal Implications


AI Discord Recap

A summary of Summaries of Summaries

Meta's Llama 3 Release Sparks Excitement and Debate

Mixtral Raises the Bar for Open-Source AI

Efficient Inference and Model Compression Gain Traction

Open-Source Tooling and Applications Flourish

Emerging Research Highlights


PART 1: High level Discord summaries

Perplexity AI Discord


Unsloth AI (Daniel Han) Discord


LM Studio Discord


Nous Research AI Discord

A Call for Multi-GPU Support: There are struggles with achieving efficient long context inference for models like Jamba using multi-GPU setups; deepspeed and accelerate documentation lack guidance on the matter.

Ripe for an Invite: TheBloke's Discord server resolved its inaccessible invite issue, with the new link now available: Discord Invite.

Reports Go Commando: The /report command has been introduced for effectively reporting rule violators within the server.

Llama 3 Ignites Benchmarking Blaze: Llama 3 is being rigorously benchmarked and compared to Mistral among users, with its performance and AI chat templates under the lens. Concerns about model limitations, such as the 8k token context limit, and restrictive licensing were prominent.

Pickle Cautions and AI Index: Dialogues on compromised systems via insecure pickle files and non-robust GPT models featured in the conversation. The AI community was directed to the AI Index Report for 2023 for insights on the year's development.

Cross-Model Queries and Support Calls: Queries included the search for effective prompt formats for Hermes-based models, anticipated release of llama-3-Hermes-Pro, and whether axolotl supports simultaneous multi-model training. The support for long context inferences on GPU clusters using models like jamba is under development, as seen in the vLLM project's GitHub pull request.

VLM on Petite Processors: A project aiming to deploy VLM (Vision Language Models) on Raspberry Pis for educational use hints at the ever-growing versatility in AI deployment platforms.

Data Dilemmas and Dimensionality Debates: Open-source models' need for fine-tuning and issues with data diversity, including the curse of dimensionality, have been topics of agreement. Moreover, strategies for creating effective RAG databases ranged from single large to multiple specialized databases.

Simulation Joins AI Giants: A fervent discussion has taken place centered around the integration of generative AI like Llama 3 and Meta.ai with world-sim, exploring the creation of rich, AI-powered narratives.


CUDA MODE Discord

Matrix Multiplication Mastery: Engineers debated optimal strategies for tiling matrix multiplication in odd-sized scenarios, proposing padding or boundary-specific code to improve efficiency. They highlighted the balance between major part calculations and special edge case handling.

CUDA Kernels Under the Microscope: Discussions on FP16 matrix multiplication (matmul) errors surfaced, suggesting the superior error handling of simt_hgemv compared to typical fp16 accumulation approaches. The group also examined dequantization in quantized matmuls, sequential versus offset memory access, and the value of vectorized operations like __hfma2, __hmul2, and __hadd2.

On the Shoulders of Giants: Members explored integrating custom CUDA and Triton kernels with torch.compile, sharing a Custom CUDA extensions example and directing to a comprehensive C++ Custom Operators manual.

CUDA Quest for Knowledge: There was an exchange on CUDA learning resources with the suggestion to learn it before purchasing hardware, and recommending a YouTube playlist for the theory and a GitHub CUDA guide for practice.

Leveraging CUDA for LLM Optimization: The community successfully reduced a CUDA model training loop from 960ms to 77ms using NVIDIA Nsight Compute for optimizations, highlighting the specific improvements and considering multi-GPU approaches for further enhancements. Details on the loop optimization can be found in a pull request.

Training Garb for Engineers: Discussions for CUDA Mode events necessitated coordination regarding recording duties, sparking conversations on suitable workflows and tools for capturing and potentially editing the sessions, in addition to managing event permissions and scheduling.


OpenAccess AI Collective (axolotl) Discord


Stability.ai (Stable Diffusion) Discord

Mark Your Calendars for SD3 Weights: Discussions indicate excitement for the upcoming May 10th release of Stable Diffusion 3 local weights, with members anticipating new capabilities and enhancements.

Censorship or Prudence?: Conversations surfaced concerns regarding the Stable Diffusion API, which might produce blurred outputs for certain prompts, signaling a disparity in content control between local versions and API usage.

GPU Picking Made Simpler: AI practitioners highlighted the cost-effectiveness of the RTX 3090 for AI tasks, weighing its advantages over pricier options like the RTX 4080 or 4090, factoring in VRAM and computational efficiency.

Artistic Mastery in AI: Dialogue in the community has been geared towards fine-tuning content generation, with members exchanging advice on creating specific image types, such as half-face portrayals, and controlling the nuances of the resulting AI-generated art.

AI Assistance Network: Resources like a detailed Comfy UI tutorial have been shared for community learning, and users are both seeking and providing tips on handling technical errors, including img2img IndexError and strategies for detecting hidden watermarks in AI imagery.


Latent Space Discord

Rocking the Discord Server with AI: A member explored the idea of summarizing a dense Discord server on systems engineering using Claude 3 Haiku and an AI news bot; they also shared an invitational link.

Meta's Might in Machine Learning: Meta introduced Llama 3, with conversations buzzing around its 8B and 70B model iterations outclassing SOTA performance, a forthcoming 400B+ model, and comparison to GPT-4. Participants noted Llama 3's superior inference speed, especially on Groq Cloud.

Macs and Llamas, an Inference Odyssey: Debates flared up about running large models like Llama 3 on Macs, with some members suggesting creative workarounds by combining local Linux boxes with Macs for optimized performance.

Hunt for the Ultimate LLM Blueprint: In search of efficiency, community members shared litellm, a promising resource to adapt over 100 LLMs with consistent input/output formats, simplifying the initiation of such projects.

Podcast Wave Hits the Community: Latent Space aired a new podcast episode featuring Jason Liu, with community members showing great anticipation and sharing the announcement Twitter link.

Engage, Record, and Prompt: The LLM Paper Club held discussions on the relevance of tokenizers and embeddings, announced the recording of sessions for YouTube upload, and examined model architectures like ULMFiT's LSTM. In-the-know participants confirmed PPO's auxiliary objectives and engaged in jest about the so-called 'prompting epoch.'

AI Evaluations and Innovations: The AI In Action Club pondered the pros and cons of using Discord versus Zoom, shared insights into LLM Evaluation, tackled unidentified noise during sessions, and shared strategies for abstractive summarization evaluation. Links to Eugene Yan's articles were circulated, underscoring the importance of reliability in AI evaluations.


Eleuther Discord

Best-fit Packing: Less Truncation, More Performance: A new Best-fit Packing method reduces truncation in large language model training, aiming for optimal document packing into sequences, according to a recent paper.

Unpacking the Softmax Bottleneck: Small language models underperform due to saturation linked with the softmax bottleneck, with challenges for models under 1000 hidden dimensions, as discussed in a recent study.

Scaling Laws Remain Chinchillated: Conversations in the scaling-laws channel have concluded that the Chinchilla token count per parameter stays consistent and that there might be more benefit in adding parameters over accumulating more data.

DeepMind Dives into Sparse Autoencoders: DeepMind's mechanistic interpretability team outlined advancements in Sparse Autoencoders (SAEs) and provided insights on interpretability challenges and techniques in a forum post, along with a relevant tweet.

Tackling lm-evaluation-harness Challenges: Efforts to contribute to the lm-evaluation-harness project have been hampered by the complexity of configurations and the need for a cleaner implementation method, with shared insights into the potential for multilingual benchmarking via PRs.


LAION Discord


HuggingFace Discord

Spanning Languages and Models: A Summary of Discourse

Technical Exchange Flourishing: AI Engineers exchange knowledge on everything from deep reinforcement learning (DRL) in object detection to GPU issues in Gradio and the perplexing 'cursorop' error in TensorFlow. Discussions are also oriented towards 3D vision datasets and solutions for consistent backgrounds in inpainting with Lora. An open call was made to explore Counterfactual-Inception research on GitHub.


OpenRouter (Alex Atallah) Discord


OpenAI Discord

Turbo Challenged by Claude: Users have reported slow performance with gpt-4-turbo-2024-04-09, finding it slower than its predecessor, GPT-4-0125-preview. Inquiries were made about faster versions, and some have integrated Claude to compensate for speed issues, yet with mixed results.

AI Grapples with PDFs: Conversations zeroed in on the inefficiency of PDFs as a data input format for AI, with community members advising the use of plain text or structured formats like JSON, while also noting XML is not currently supported for files.

Performance Anxiety Over ChatGPT: Members expressed concerns over the declining performance of ChatGPT, sparking debate over possible reasons which ranged from strategic responses to legal challenges to deliberate performance downgrades.

Engineering More Effective Prompts: There was a community effort to confirm and update the prompt engineering best practices, as recommended in the OpenAI guide with discussions pointing to real issues in prompt consistency and failure to adhere to instructions.

Integrating AI with Blockchain: A blockchain developer called for collaboration on projects combining AI with blockchain, suggesting an interaction between advanced prompt engineering and decentralized technologies.


Interconnects (Nathan Lambert) Discord


Modular (Mojo 🔥) Discord


Cohere Discord


LlamaIndex Discord

Retrieval Augmented Generations Right at Our Fingertips: Engineers at Elastic have released a blog post demonstrating the construction of a Retrieval Augmented Generation (RAG) application using Elasticsearch and LlamaIndex, an integration of open tools including @ollama and @MistralAI.

Llama 3 Gets a Handy Cookbook: The LlamaIndex team has provided early support for Llama 3, the latest model from Meta, through a "cookbook" detailing usage from simple prompts to entire RAG pipelines. The guide can be fetched from this Twitter update.

Setting Up Shop Locally with Llama 3: For those looking to run Llama 3 models in a local environment, Ollama has shared a notebook update that includes simple command changes. The update can be applied by altering "llama2" to "llama3" as detailed here.

Puzzle & Dashboards: Pinecone and LLM Daily Struggles: Amidst technical exchanges, there was curiosity about how Google's Vertex AI handles typos in signs like "timbalands", as seen on their demo site, and ongoing dialogues surrounding the creation of an interactive dashboard for generating recipes from input ingredients.

Ready, Set, Track LlamaIndex's Progress: Interest around tracking the development of LlamaIndex spiked among engineers following confirmation that LlamaIndex has secured funding, a nod to the project's growth and anticipated advancements in the space.


DiscoResearch Discord

Mixtral's Multilingual Might: The Mixtral model mix of English and German showcases its language prowess, though evaluations are imminent. Technical challenges, including shape errors and OOM issues, hint at the complexity of training large models, while the efficacy of parameters such as "router_aux_loss_coef" in Mixtral's config remains a point of debate.

Meta's Llama Lightning Strikes: Meta's Llama 3 enters the fray, touting multilingual capabilities but with discernible performance discrepancies in non-English languages. Access to the new tokenizer is anticipated, and critiques focus on downstream usage restrictions of model outputs, sparking a discussion on the confluence of open source and proprietary constraints.

German Language Models Under Microscope: Initial tests suggest Llama3 DiscoLM German lags behind Mixtral in German proficiency, with notable grammar issues and incorrect token handling, despite a Gradio demo availability. Questions regarding the Llama3's dataset alignment and tokenizer configurations arise, and comparisons with Meta's 8B models show performance gaps that beg investigation.


OpenInterpreter Discord

ESP32 Demands WiFi for Linguistic Acumen: An engineer pointed out that ESP32 requires a WiFi connection to integrate with language models, emphasizing the necessity of network connectivity for operational functionality.

Ollama 3's Performance Receives Engineer's Applause: In the guild, there was a buzz about the performance of Ollama 3, with engineers experimenting with the 8b model and probing into enhancements for the text-to-speech (TTS) and speech-to-text (STT) models for accelerated response times.

OpenInterpreter Toolkit Trials and Tribulations: Users shared challenges with OpenInterpreter, ranging from file creation issues using CLI that wraps output with echo to BadRequestError during audio transmission attempts with M5Atom.

Fine-Tuning Local Language Mastery: Guild members discussed how to set up OS mode locally with OpenInterpreter, providing a Colab notebook for guidance and exchanged insights on refining models like Mixtral or LLama with concise datasets for nimble learning.

Exploring Meta_llama3_8b: A member shared a link to Hugging Face where fellow engineers can interact with the Meta_llama3_8b model, indicating a resource for hands-on experimentation and evaluation within the community.


LangChain AI Discord


Alignment Lab AI Discord


Mozilla AI Discord

Llama 3 8b Takes the Stage: The llamafile-0.7 update now supports Llama 3 8b models using the -m <model path> parameter, as discussed by richinseattle; however, there's a token issue with the instruct format highlighted alongside a Reddit discussion.

Patch on the Horizon: A pending update to llamafile promises to fix compatibility issues with Llama 3 Instruct, which is detailed in this GitHub pull request.

Quantum Leap in Llama Size: jartine announced the imminent release of a quantized version of llama 8b on Llamafile, indicating advancements for the efficiency-seeking community.

Meta Llama Weights Unbound: jartine shared the Meta Llama 3 8B Instruct executable weights for community testing on Hugging Face, noting that there are a few kinks to work out, including a broken stop token.

Model Mayhem Under Management: Community efforts in testing Llama 3 8b models yielded optimistic results, with a fix for the stop token issue in Llama 3 70b communicated by jartine; minor bugs are to be anticipated.


Skunkworks AI Discord

Databricks Goes GPU: Databricks has released a public preview of model serving, enhancing performance for Large Language Models (LLMs) with zero-config GPU optimization but may increase costs.

Ease of LLM Fine-Tuning: A new guide explains fine-tuning LLMs using LoRA adapters, Flash Attention, and tools like DeepSpeed, available at modal.com, offering strategies for efficient weight adjustments in models.

Affordable Serverless Solutions: An affordable serverless hosting guide using GPUs is available on GitHub, which could potentially lower expenses for developers - check the modal-examples repo.

Mixtral 8x22B Raises the Bar: The Mixtral 8x22B is a new model employing a sparse Mixture-of-Experts, detailed in a YouTube video, setting high standards for AI efficiency and performance.

Introducing Meta Llama 3: Facebook's Llama 3 adds to the roster of cutting-edge LLMs, open-sourced for advancing language technologies, with more information available on Meta AI's blog and a promoting YouTube video.


LLM Perf Enthusiasts AI Discord


Datasette - LLM (@SimonW) Discord

Karpathy's Llama 3 Lasso: Andrej Karpathy's tweet raised discussions on the potential of compact models, noting an 8B parameter model trained on a 15T dataset as an example of possibly undertraining common LLMs by factors of 100-1000X, pointing engineers towards the notion of longer training cycles for smaller models.

Small Models, Big Expectations: Reactions to Karpathy's insights echo among members who express enthusiasm for the deployment of small yet efficient models like Llama 3, indicating a community ready to embrace optimal resource utilization in developing smaller, mightier LLMs.

Plugin Installation Snags: A member's ModuleNotFoundError while installing a llm plugin led to the revelation that conflicting installations from both brew and pipx might be at the root. A clean reinstall ended the ordeal, hinting at the necessity of vigilant environment management.

Concurrent Confusion Calls for Cleanup: The cross-over installation points from brew and pipx led a user astray, sparking reminders within the community to check which version of a tool is being executed with which llm to dodge similar issues in the future.

LLM Fun Facts: Amidst the technical back-and-forth, a shared use case for llm provided a light-hearted moment, presenting a practical, engaging application of the technology for members to explore.


tinygrad (George Hotz) Discord


AI21 Labs (Jamba) Discord

Long Context Inference Woes with Jamba: A Jamba user is struggling with long context inferences on a 2x A100 cluster and is seeking troubleshooting code for the distributed system's issue. There has been no follow-up discussion or provided solutions to the problem yet.


PART 2: Detailed by-Channel summaries and links

Perplexity AI ▷ #general (1059 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (14 messages🔥):

Links mentioned:


Perplexity AI ▷ #pplx-api (11 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (1147 messages🔥🔥🔥):

Links mentioned:

We're upgrading Meta AI with our new state-of-the-art Llama 3 AI model, which we're open sourcing. With this new model, we believe Meta AI is now the most intelligent AI assistant that you can freely use.

We're making Meta AI easier to use by integrating it into the search boxes at the top of WhatsApp, Instagram, Facebook, and Messenger. We also built a website, meta.ai, for you to use on web.

We also built some unique creation features, like the ability to animate photos. Meta AI now generates high quality images so fast that it creates and updates them in real-time as you're typing. It'll also generate a playback video of your creation process.

Enjoy Meta AI and you can follow our new @meta.ai IG for more updates.": 157K likes, 9,028 comments - zuckApril 18, 2024 on : "Big AI news today. We're releasing the new version of Meta AI, our assistant that you can ask any question across our apps and glasses....Fail to load a tokenizer (CroissantLLM) · Issue #330 · unslothai/unsloth: Trying to run the colab using a small model: from unsloth import FastLanguageModel import torch max_seq_length = 2048 # Gemma sadly only supports max 8192 for now dtype = None # None for auto detec...Mark Zuckerberg - Llama 3, $10B Models, Caesar Augustus, & 1 GW Datacenters: Zuck on:- Llama 3- open sourcing towards AGI - custom silicon, synthetic data, & energy constraints on scaling- Caesar Augustus, intelligence explosion, biow...[Usage]: Llama 3 8B Instruct Inference · Issue #4180 · vllm-project/vllm: Your current environment Using the latest version of vLLM on 2 L4 GPUs. How would you like to use vllm I was trying to utilize vLLM to deploy meta-llama/Meta-Llama-3-8B-Instruct model and use OpenA...HuggingChat: Making the community's best AI chat models available to everyone.‘Her’ AI, Almost Here? Llama 3, Vasa-1, and Altman ‘Plugging Into Everything You Want To Do’: Llama 3, Vasa-1, and a host of new interviews and updates, AI news comes a bit like London buses. I’ll spend a couple minutes covering the last-minute Llama ...no title found: no description foundAdaptive Text Watermark for Large Language Models: no description foundGoogle Colaboratory: no description foundTweet from Andrej Karpathy (@karpathy): Congrats to @AIatMeta on Llama 3 release!! 🎉 https://ai.meta.com/blog/meta-llama-3/ Notes: Releasing 8B and 70B (both base and finetuned) models, strong-performing in their model class (but we'l...LLAMA-3 🦙: EASIET WAY To FINE-TUNE ON YOUR DATA 🙌: Learn how to fine-tune the latest llama3 on your own data with Unsloth. 🦾 Discord: https://discord.com/invite/t4eYQRUcXB☕ Buy me a Coffee: https://ko-fi.com...How to Fine Tune Llama 3 for Better Instruction Following?: 🚀 In today's video, I'm thrilled to guide you through the intricate process of fine-tuning the LLaMA 3 model for optimal instruction following! From setting...meta-llama/Meta-Llama-3-8B-Instruct · Fix chat template to add generation prompt only if the option is selected: no description found


Unsloth AI (Daniel Han) ▷ #announcements (1 messages):

Link mentioned: Google Colaboratory: no description found


Unsloth AI (Daniel Han) ▷ #random (6 messages):

Link mentioned: ‘Her’ AI, Almost Here? Llama 3, Vasa-1, and Altman ‘Plugging Into Everything You Want To Do’: Llama 3, Vasa-1, and a host of new interviews and updates, AI news comes a bit like London buses. I’ll spend a couple minutes covering the last-minute Llama ...


Unsloth AI (Daniel Han) ▷ #help (341 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (4 messages):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (3 messages):


LM Studio ▷ #💬-general (661 messages🔥🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (617 messages🔥🔥🔥):

Links mentioned:


LM Studio ▷ #announcements (1 messages):

Links mentioned:


LM Studio ▷ #🧠-feedback (5 messages):


LM Studio ▷ #📝-prompts-discussion-chat (4 messages):


LM Studio ▷ #🎛-hardware-discussion (16 messages🔥):


LM Studio ▷ #🧪-beta-releases-chat (5 messages):

Link mentioned: lmstudio-community/Meta-Llama-3-8B-Instruct-GGUF · Hugging Face: no description found


LM Studio ▷ #autogen (7 messages):


LM Studio ▷ #rivet (1 messages):


LM Studio ▷ #avx-beta (2 messages):


LM Studio ▷ #amd-rocm-tech-preview (25 messages🔥):

Links mentioned:


LM Studio ▷ #model-announcements (1 messages):


Nous Research AI ▷ #ctx-length-research (2 messages):


Nous Research AI ▷ #off-topic (18 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (33 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (807 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #rules (1 messages):


Nous Research AI ▷ #ask-about-llms (12 messages🔥):

Link mentioned: [Model] Jamba support by mzusman · Pull Request #4115 · vllm-project/vllm: Add Jamba support to vLLM, This PR comprises two parts: the Jamba modeling file and the Mamba memory handling. Since Jamba is a hybrid model (which alternates between mamba and transformer layers),...


Nous Research AI ▷ #project-obsidian (1 messages):


Nous Research AI ▷ #rag-dataset (24 messages🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (446 messages🔥🔥🔥):

Links mentioned:


CUDA MODE ▷ #general (28 messages🔥):

Links mentioned:


CUDA MODE ▷ #cuda (30 messages🔥):


CUDA MODE ▷ #torch (8 messages🔥):

Links mentioned:


CUDA MODE ▷ #beginner (14 messages🔥):

Links mentioned:


CUDA MODE ▷ #pmpp-book (4 messages):


CUDA MODE ▷ #hqq (81 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #llmdotc (552 messages🔥🔥🔥):

Links mentioned:


CUDA MODE ▷ #massively-parallel-crew (10 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #general (489 messages🔥🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (13 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (43 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #runpod-help (3 messages):


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (14 messages🔥):

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (471 messages🔥🔥🔥):

Links mentioned:


Latent Space ▷ #ai-general-chat (229 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (3 messages):


Latent Space ▷ #llm-paper-club-west (66 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-in-action-club (69 messages🔥🔥):

Links mentioned:


Eleuther ▷ #general (115 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (149 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (38 messages🔥):


Eleuther ▷ #interpretability-general (2 messages):

Link mentioned: Progress Update #1 from the GDM Mech Interp Team: Summary — AI Alignment Forum: Introduction This is a progress update from the Google DeepMind mechanistic interpretability team, inspired by the Anthropic team’s excellent monthly…


Eleuther ▷ #lm-thunderdome (14 messages🔥):

Links mentioned:


LAION ▷ #general (255 messages🔥🔥):

Links mentioned:


LAION ▷ #research (15 messages🔥):

Link mentioned: no title found: no description found


HuggingFace ▷ #announcements (11 messages🔥):

Links mentioned:


HuggingFace ▷ #general (166 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (4 messages):

Link mentioned: Meta Releases LLaMA 3: Deep Dive & Demo: Today, 18 April 2024, is something special! In this video, In this video I'm covering the release of @meta's LLaMA 3. This model is the third iteration of th...


HuggingFace ▷ #cool-finds (9 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (12 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (1 messages):

Link mentioned: GitHub - IVY-LVLM/Counterfactual-Inception: Contribute to IVY-LVLM/Counterfactual-Inception development by creating an account on GitHub.


HuggingFace ▷ #computer-vision (2 messages):


HuggingFace ▷ #NLP (3 messages):


HuggingFace ▷ #diffusion-discussions (4 messages):


OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

Link mentioned: Mixtral 8x22B by mistralai | OpenRouter: Mixtral 8x22B is a large-scale language model from Mistral AI. It consists of 8 experts, each 22 billion parameters, with each token using 2 experts at a time. It was released via [X](https://twitter...


OpenRouter (Alex Atallah) ▷ #general (198 messages🔥🔥):

Links mentioned:


OpenAI ▷ #ai-discussions (106 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (22 messages🔥):


OpenAI ▷ #prompt-engineering (30 messages🔥):


OpenAI ▷ #api-discussions (30 messages🔥):


Interconnects (Nathan Lambert) ▷ #ideas-and-feedback (3 messages):

Link mentioned: Don't throw away your value model! Generating more preferable text with Value-Guided Monte-Carlo Tree Search decoding: Inference-time search algorithms such as Monte-Carlo Tree Search (MCTS) may seem unnecessary when generating natural language text based on state-of-the-art reinforcement learning such as Proximal Pol...


Interconnects (Nathan Lambert) ▷ #news (142 messages🔥🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #sp2024-history-of-open-alignment (16 messages🔥):


Interconnects (Nathan Lambert) ▷ #posts (2 messages):


Modular (Mojo 🔥) ▷ #general (19 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (2 messages):


Modular (Mojo 🔥) ▷ #ai (2 messages):

Link mentioned: Meta Releases LLaMA 3: Deep Dive & Demo: Today, 18 April 2024, is something special! In this video, In this video I'm covering the release of @meta's LLaMA 3. This model is the third iteration of th...


Modular (Mojo 🔥) ▷ #🔥mojo (118 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (1 messages):

arnaud6135: thank you, I'll read it right away 😄


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (3 messages):


Modular (Mojo 🔥) ▷ #nightly (15 messages🔥):


Cohere ▷ #general (157 messages🔥🔥):

Links mentioned:


Cohere ▷ #project-sharing (3 messages):


LlamaIndex ▷ #blog (5 messages):

Links mentioned:


LlamaIndex ▷ #general (118 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (2 messages):


DiscoResearch ▷ #mixtral_implementation (8 messages🔥):

Link mentioned: maxidl/Mixtral-8x22B-v0.1-Instruct-sft-en-de · Hugging Face: no description found


DiscoResearch ▷ #general (21 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (26 messages🔥):

Links mentioned:


OpenInterpreter ▷ #general (18 messages🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (17 messages🔥):


OpenInterpreter ▷ #ai-content (1 messages):

kieguin: https://huggingface.co/spaces/ysharma/Chat_with_Meta_llama3_8b


LangChain AI ▷ #general (28 messages🔥):

Links mentioned:


LangChain AI ▷ #langserve (2 messages):


LangChain AI ▷ #langchain-templates (1 messages):


LangChain AI ▷ #share-your-work (3 messages):

Links mentioned:


Alignment Lab AI ▷ #ai-and-ml-discussion (3 messages):

Links mentioned:


Alignment Lab AI ▷ #programming-help (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #looking-for-collabs (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #landmark-dev (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #oo (6 messages):

Links mentioned:


Alignment Lab AI ▷ #landmark-evaluation (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #open-orca-community-chat (3 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #leaderboard (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #looking-for-workers (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #looking-for-work (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #join-in (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #fasteval-dev (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #qa (2 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Mozilla AI ▷ #llamafile (19 messages🔥):

Links mentioned:


Skunkworks AI ▷ #finetuning (6 messages):

Links mentioned:


Skunkworks AI ▷ #off-topic (2 messages):

Links mentioned:


LLM Perf Enthusiasts AI ▷ #general (6 messages):

Link mentioned: Falling Falling Down Stairs GIF - Falling Falling Down Stairs Stairs - Discover & Share GIFs: Click to view the GIF


Datasette - LLM (@SimonW) ▷ #ai (3 messages):


Datasette - LLM (@SimonW) ▷ #llm (3 messages):


tinygrad (George Hotz) ▷ #general (2 messages):

Link mentioned: Meta Llama 3: Build the future of AI with Meta Llama 3. Now available with both 8B and 70B pretrained and instruction-tuned versions to support a wide range of applications.


AI21 Labs (Jamba) ▷ #jamba (1 messages):