Frozen AI News archive

Not much happened today

**Ilya Sutskever** steps down as Chief Scientist at **OpenAI** after nearly a decade, with **Jakub Pachocki** named as his successor. **Google DeepMind** announces **Gemini 1.5 Pro** and **Gemini 1.5 Flash** models featuring 2 million token context and improved multimodal capabilities, alongside demos of **Project Astra** AI assistant, **Imagen 3** text-to-image model, and **Veo** generative video model. **GPT-4o** tops the VHELM leaderboard and outperforms competitors on LMSYS Chatbot Arena. **Reka Core** multimodal model with 128K context and **Alibaba's Qwen1.5-110B** open-source model are released. **Salesforce** shares an online RLHF recipe.

Canonical issue URL

AI News for 5/14/2024-5/15/2024. We checked 7 subreddits, 384 Twitters and 30 Discords (427 channels, and 6455 messages) for you. Estimated reading time saved (at 200wpm): 686 minutes.

'Twas the night after I/O, when all through AI
Not a startup was posting, not even on LI
The UBI research was studied by e/accs with care
In hopes that AGI soon would be there

You can wish Ilya and Jan and Evan well (is there something to the departure timeline?), read about GPT4o's incredible multi-Needlestack performance, or watch John Schulman or Sama's latest interviews, if you're team OpenAI, or you can congratulate Mike Krieger on joining Anthropic, or you can read all the Google I/O roundups that came after us (it seems we underrated PaliGemma initially).


Table of Contents

[TOC]


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

Ilya Sutskever Leaving OpenAI

Google I/O AI Announcements

AI Model Releases and Benchmarks

Multimodal AI and Video Models

Memes and Humor


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

OpenAI Leadership Changes and Internal Dynamics

GPT-4o Capabilities and Limitations

Google I/O AI Announcements

Open Source Alternatives and Concerns

Implications and Societal Impact

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries. We are concluding that Claude still remains the best summarizer model so we are dropping the GPT4T and 4o comparisons.

  1. Unveiling of New AI Models and Capabilities:

    • Google introduced several new AI models at Google I/O, including Veo for high-quality video generation, Imagen 3 for improved text-to-image capabilities, and Gemma 2, a 27B parameter model. Source
    • OpenAI's GPT-4o was revealed as the top model on LMSYS's Chatbot Arena leaderboard under a secret name before its launch. Source
    • Nous Research released Hermes 2 Θ, an experimental model merging Hermes 2 Pro and Llama-3 Instruct, outperforming previous models on benchmarks while retaining function calling capabilities. Source
  2. Advances in Multimodal AI and Unified Models:

    • Discussions centered around the challenges and potential of multimodal models, with members exploring unified models like ImageBind that bind information across multiple modalities using joint embeddings. Source
    • Google's Gemini 1.5 Flash and Gemini 1.5 Pro were introduced, offering multimodal capabilities for visual understanding, classification, summarization, and content creation from various inputs. Source
    • Members discussed the potential of integrating multimodal models directly into smartphones and edge devices for low latency and enhanced multimodal functionalities. Source
  3. Optimization and Efficiency Efforts for LLMs:

    • Techniques like Gemini's context caching and llama.cpp's prompt caching were discussed as ways to make LLM workflows more efficient and cost-effective by reducing token usage for long prompts. Source
    • Members explored strategies to improve the L2 cache hit rate for better performance, referencing resources like the Triton Matrix Multiplication tutorial on block-level multiplication and pointer arithmetic. Source
    • Discussions revolved around optimizing tensor allocations and caching when using torch.compile, with recommendations to replace dynamic allocations with pre-allocated tensors and leverage static caching to reduce overhead. Source
  4. Debates on LLM Evaluation and Industry Dynamics:

    • A blog post highlighted the closed nature of current LLM evaluation practices, dominated by academic benchmarks and private A/B testing, calling for broader accessibility in evaluations. Source
    • Members discussed Anthropic's transition towards becoming a product company, OpenAI's potential foray into search with a key hire, and the strategic need for AI companies to offer end-user products rather than just APIs or services. Source
    • The departure of Ilya Sutskever from OpenAI sparked discussions about potential reshuffles within the company, with Sam Altman and others commenting on the transition. Source

PART 1: High level Discord summaries

OpenAI Discord

GPT-4o Faces Creative Block: GPT-4o's faster response time compared to GPT-4 comes with a trade-off in creative writing tasks, where it often regurgitates rough drafts instead of enhancing them with intelligent revising. This has caused some discontent among users trying to leverage the AI for writing enhancements.

Models Go Musical with Auditory Descriptions: Community members have creatively used prompts asking GPT-4 and GPT-4o to describe instrumental songs, like "The XX Intro" and "Tears in Rain", to gauge the models' descriptive abilities when it comes to auditory perceptions. Results from these prompts may provide insights into each model's interpretive skills.

An Imaginative Challenge in Imagery Generation: A user encountered difficulties with both GPT-4 and GPT-4o when requesting specific side views for platform game development – the AI tended to supply unwanted isometric perspectives and extraneous details, indicating a gap in perspective understanding and context adherence.

Access and Features on a Staggered Path: The rollout of GPT-4o features has been sequential, with functionality like Voice Mode and multimodal capabilities first reaching API partners, then Plus users. This gradual deployment has led to confusion and access issues for some members.

Custom Models Spark Collaboration: The discussion regarding the integration of custom GPTs with GPT-4o and feedback sessions on custom models like OptimusGPT highlight the community's eagerness for improvement and collaboration. Users have been advised to migrate their custom GPTs to GPT-4o within the forthcoming weeks for enhanced performance.

Simplicity in AI Voice Assistance: A member highlighted the Plug & Play AI Voice Assistant, which can be set up quickly and invited the community to try it out. Despite the repeated announcements, the focus was on the assistant's ease of setup and the value of user feedback on its efficacy.


Unsloth AI (Daniel Han) Discord

Machine Learns Mythos: An AI geared towards Cthulhu worship was developed by a guild member using Unsloth to create and fine-tune data, resulting in the creation of TinyLlama and Mistral 7B Cthulhu models with resources available on Huggingface.

Sailing the Quantization Seas: Conversations tackled challenges in quantization and model merging, with members sharing tips like manually upcasting to 16-bit before merging and using notebooks to facilitate conversion processes, illustrating the complex terrain of optimizing AI models for better performance.

Global Model Outreach: Unsloth was recognized in an AI News feature for its strides in open-source AI development, and community members rallied in support of a proposal to showcase Unsloth at an upcoming New York City open-source data pipeline meetup.

GPT-4 Lifeguard On Duty: Assistance was offered to a guild member struggling with Trigonometry problems, evidencing the community's rapid response in providing resources like ChatGPT and Claude for academic aid.

AI Summarization Scrutinized: Potential conflicts with European data privacy laws were flagged concerning the use of AI to summarize Discord interactions, which signals the ongoing vigilance required to balance technological innovation with legal compliance.


Perplexity AI Discord


Nous Research AI Discord

Bold Breakthrough in LLM Performance: The newly launched Hermes 2 Θ, outshining Hermes 2 Pro and Llama-3 Instruct, boasts superior performance in benchmarks while maintaining the ability to call functions, as announced in the announcements.

Discord Meets Innovation: A tool exploiting a bug in Discord allows embedding AV1 videos larger than 500MB, which can also be shared on platforms like Twitter, as discussed in off-topic.

GPT-4's Mixed Reviews: Despite GPT-4's prowess in data science tasks, general channel discussions reveal its underperformance in complex tasks and a tendency to lose context, hinting at trade-offs between speed and accuracy.

Nordic AI Language Model Unleashed: interesting-links showcases Viking 7B, a leading-edge multilingual LLM designed for Nordic languages by Silo AI and University of Turku's TurkuNLP, enhancing language AI accessibility.

AI Skepticism and Enthusiasm Intertwined: General sentiment across various channels such as general and ask-about-llms remains mixed with enthusiasm over new models like Hermes 2 Θ, yet sceptical on multimodal capabilities and the barriers faced when building LLMs from scratch.


LM Studio Discord

Getting Vocal with LLMs: Talk of integrating voice interaction with local large language models (LLMs) highlighted the use of tools like AnythingLLM. The community discussed resource-intensive solutions involving Whisper.cpp and Coqui TTS, albeit with complexities and suboptimal experiences.

Beefing Up the Hardware Arms: Debates swung around hardware preferences for AI models, pitching a 3060Ti GPU against dual 16-core Xeon V4 CPUs. Enthusiasts mooted over VRAM's pivotal role, with a bias towards Nvidia cards for top-tier AI performance. The mention of a 4060 sparked interest for its prospective gains.

PrivateGPT vs. AnythingLLM - A Document Query Duel: The competition between PrivateGPT and AnythingLLM for querying documents with LLMs incited a technical analysis. Discussions underlined setup intricacies and user-friendly aspects of each platform.

MacOS First Strikes A Sour Note: A Mac-tier debate surfaced with grievances regarding app release priorities, primarily the MacOS-first strategy from OpenAI. This spun into a dialogue on the complexities and divergences in MacOS versus Windows app development.

Battle of the Giants in the Model Arena: From uncensored local LLM recommendations, notably Dolphin 2.8 Mistral 7B v0.2, to the nuances of quantization and model performance, the community dissected various AI paradigms. Aforementioned also was Command R models' comparison and GPU-related enigmas.

Hacking the Hardware Frontier: ROGUE RX6600, not typically supported by AMD for the ROCM build, gamely runs in Koboldcpp, while official llama.cpp binaries restrict usage due to GPU ID verification processes. Users flagged user-interface (UI) complexities within LM Studio settings.

Gleaning GPU Gems: Tips on GPU resource optimization with Windows Task Manager sallied forth, with quirky recommendations like disabling hardware acceleration to enhance resource visibility. However, struggles continue with configuring CUDA on select laptops, leading to persisting model loading errors in LM Studio.

Old Guard vs. New Recruits in GPU Tussle: Tesla M40's disappointing showdown with GeForce 1060 on LLM tasks and the touted VRAM speed's importance got limelight. Financial constraints loomed over users, with low-end PCs finding refuge in modest local models and APUs revealing no performance perks over CPUs in llama.cpp.

Beta Build Blues: In beta territory, ruminations on multimodal feature parity shared space with reports of LM Studio's launch issues due to lacking AVX2 support. A user's exasperation with a non-launching LM Studio was quelled by identifying that the AGX instruction set was paramount for operation.

The Developer's Digest: Intel’s overture for Intel GPU support using SYCL for llama.cpp broadened the horizon for LM Studio. Conversations flourished around DL model adaptation, the quest for AGI, and community calls to keep dev chatter tethered to LM Studio's APIs and software construction.


HuggingFace Discord

GPT-4o Stealthy Champion: OpenAI's GPT-4o was confirmed to be the top model under a secret name in the LMSYS's Chatbot Arena, boasting undisclosed performance feats.

Datasets and Models Leverage Enhanced: A team released a 700,000-sample Vietnamese dataset for open-source language modeling, while AutoTrain extended its toolkit with Object Detection functionality, and Diarizers emerged as a new library for fine-tuning speaker diarization systems with multilingual support on Hugging Face's Hub.

AI-Powered Story Crafters: A reading group engaged in a comprehensive review of AI story generation, with discussion pivoting towards refining the GROVE framework paper and community members sharing endeavors and learnings via Medium.

Visual Data to Revenue Insights: Inquiry in the #computer-vision channel sparked a discussion about the feasibility of training a model that converts images to sales data outputs; the original poster provided a related dataset link for reference.

Enhancing Chatbots with LangChain: In the #NLP channel, a member sought to improve chatbot conversations using LangChain, with suggestions directing to an initial starter example for using local LLM and embedding models.


Stability.ai (Stable Diffusion) Discord


OpenRouter (Alex Atallah) Discord


Modular (Mojo 🔥) Discord

Mojo Integrates MLIR: Engineers discussed Mojo's ability to execute MLIR code with minor syntax adjustments, contributing to Mojo's versatility and access to lower-level features.

Strategies for Mojo Mastery: A variety of resources for learning Mojo were recommended, including the Mojo SDK manual and the Mandelbrot notebook, with the community highlighting the language's advantages like cross-vendor GPU code portability.

Python Convenience Without Python: The community is exploring alternatives to Python dependencies within the Mojo toolchain, indicating a drive for a more language-agnostic ecosystem. Follow the progress on the feature request on GitHub.

C/C++ and Python Interop with Mojo Abuzz: There's active discussion on calling C/C++ libraries using ffi and dealing with Python interoperability issues, reflecting a keen interest in Mojo's interlanguage capabilities. Engineers are sharing insights on the mechanics, evidenced by the shared tweetorial and problem-solving threads.

Modular’s Multimedia Mojo Hustle: Modular provided updates and tutorials via new videos on Mojo nightly builds and MAX Graph API, as well as via a MAX Graph API blog tutorial. Additionally, two tweets teasing updates and a community meeting were noted, although details remained unspecified.


Eleuther Discord

Mimetic Initialization Shows Promise: Introducing mimetic initialization to Transformers yields significant accuracy improvements on datasets like CIFAR-10 and ImageNet, per a shared paper. This technique mimics weight patterns from pre-trained models, signaling potential for more efficient training.

Dataset Diversification with Sakuga-42M: The new Sakuga-42M dataset was unveiled, containing 42 million keyframes of cartoon animations and aiming to reduce biases of models trained on natural images. An arXiv link to the dataset provides the gateway for further exploration.

Hypernetworks Pique Interest for Initialization: Discussions emerged around employing hypernetworks for weight initialization, suggesting the possibility of symbolic regression for crafting innovative initialization techniques.

Leveraging Dot Products in Neural Networks: A lively discussion endorsed the effectiveness of dot products in neural networks, with a member linking to an article that examines their connection with Fourier transforms and implications for cognitive processing.

Enhancing Multiple Choice Analysis: Debates flared around optimizing the processing of multiple-choice questions in models, highlighting the lm-evaluation-harness' approach to manage requests per answer and considering an output export feature for accuracy analysis, referencing GitHub code.


CUDA MODE Discord


LlamaIndex Discord


Interconnects (Nathan Lambert) Discord

Nathan Lambert Stirs AI Talks: Nathan Lambert critiqued OpenAI's user-centric approach, expressing this view in a tweet and addressed Google's generative video advances at Google I/O as impressive, but noted some announcements like Gemini 1.5 Ultra were overlooked.

Google Unveils Gemma 2: Google announced Gemma 2, a 27-billion parameter model, at Google I/O, with updates to their AI suite including Gemini 1.5 Pro and Flash, as reported by TechCrunch.

Tokenizer Tweaks Trouble Engineers: Discussions surfaced over whether OpenAI re-pretrains with a new tokenizer or extends their current tokenizer for an LLM, alongside sharing a novel concept of Zero-Shot Tokenizer Transfer (ZeTT) discussed in an arXiv paper.

Convergence in Neural Networks Observed: Emerging research suggests neural networks, across modalities, are converging to a common statistical model of reality, as proposed in a paper and supported by Phillip Isola's mention.

AI Evaluation and Industry Shifts Highlighted: A shared blog post underscored the closed nature of current LLM evaluation practices, while discussion touched on Anthropic's move towards becoming a product company, OpenAI's notable hire hinting a possible foray into search, and the strategic need for AI companies to offer products informed by a tweet and an article.


LangChain AI Discord

AI is Waiting... Literally: Users express frustration with the slow response times of LangChain agents, taking 2-3 minutes to handle large inputs and invoke tools, and they look for prompt resolution tips. Active discussions revolved around the use of python-socketio to stream LLM responses, as participants exchanged code snippets and troubleshooting advice.

Wake Up, Server, Wake Up!: For users of hosted Langserve, intermittent issues with server inactivity and rate limiting errors are leading to unpredictable service availability. Queries are raised about whether upgrading to a Pro plan would alleviate some of these headaches and how to access more extensive logs.

Snowflake Costs in Focus with AI Optimization: An innovative Snowflake Cost Monitoring tool integrating LangChain's capabilities with Snowflake and OpenAI was demoed, aiming to streamline data visualization and analysis. The work-in-progress tool's features are showcased in a Loom video presentation.

Monetizing AI, Java Style: A Langserve user is experimenting with the py4j library to facilitate micropayment functionalities for AI interactions through a JVM, targeting crypto SDK integrations. The setup aims to innovate micropayment structures by tracking prompt/response token counts and adding a profit margin to the OpenAI API keypair usage.

Database Dilemmas and Embedding Efficiency: Threads run with discussions on embedding transfers between vector databases like pgvector and Qdrant. Members shared strategies for parallel transfer and optimizing retrieval speed, backing their points with references like the Supabase blog on Matryoshka Embeddings. Moreover, clarifications were sought on the deprecation of LLMChain in favor of RunnableSequence for MultiQueryRetriever, amid notes of API alignment holdups.


LAION Discord


OpenInterpreter Discord


Datasette - LLM (@SimonW) Discord


Latent Space Discord


OpenAccess AI Collective (axolotl) Discord


Cohere Discord


tinygrad (George Hotz) Discord


AI Stack Devs (Yoko Li) Discord


DiscoResearch Discord

Token Troubles and Triumphs: Engineers squabble over the lack of data on vocab_size vs. tokens/byte for German, highlighting a gap in the tokenizer dataset which favors language mixture.

The Ungreedy Tokenizer Arrives: A new tool for the tokenization trade, TokenMonster project, an "Ungreedy subword tokenizer and vocabulary trainer", receives a bright spotlight for its utility in Python, Go, & Javascript.

A Saucy AI Demo Goes Viral: GPT-4o's recent demo faced a round of chuckles and raised brows for being suggestively humorous, as noted in a humorous tweet by a user.

Vocab Shockwave Hits Twitter: The "o200k_base" vocab for GPT-4o triggers reactions of surprise and possibly dismay among the tech community, as conveyed by a tweet.

Ilya's Farewell from OpenAI: Ilya Sutskever parts ways with OpenAI, sparking conversations on the future of the research giant, an exit announced via Twitter.


Skunkworks AI Discord


MLOps @Chipro Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

OpenAI ▷ #ai-discussions (954 messages🔥🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (178 messages🔥🔥):


OpenAI ▷ #prompt-engineering (128 messages🔥🔥):

Link mentioned: ChatGPT can now access the live Internet. Can the API?: Given the news announcement I am wondering if the API now has that same access to the Internet. Thanks in advance!


OpenAI ▷ #api-discussions (128 messages🔥🔥):

Link mentioned: ChatGPT can now access the live Internet. Can the API?: Given the news announcement I am wondering if the API now has that same access to the Internet. Thanks in advance!


OpenAI ▷ #api-projects (4 messages):


Unsloth AI (Daniel Han) ▷ #general (453 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (37 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (229 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (4 messages):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #community-collaboration (4 messages):


Perplexity AI ▷ #general (646 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (14 messages🔥):


Perplexity AI ▷ #pplx-api (11 messages🔥):


Nous Research AI ▷ #off-topic (9 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (12 messages🔥):

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

Links mentioned:


Nous Research AI ▷ #general (342 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (40 messages🔥):

Links mentioned:


Nous Research AI ▷ #project-obsidian (2 messages):

Link mentioned: google/paligemma-3b-pt-224 · Hugging Face: no description found


Nous Research AI ▷ #bittensor-finetune-subnet (2 messages):


Nous Research AI ▷ #world-sim (22 messages🔥):


LM Studio ▷ #💬-general (176 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (109 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (10 messages🔥):


LM Studio ▷ #⚙-configs-discussion (3 messages):


LM Studio ▷ #🎛-hardware-discussion (13 messages🔥):


LM Studio ▷ #🧪-beta-releases-chat (5 messages):


LM Studio ▷ #🛠-dev-chat (104 messages🔥🔥):


HuggingFace ▷ #announcements (3 messages):

Links mentioned:


HuggingFace ▷ #general (306 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (3 messages):


HuggingFace ▷ #cool-finds (9 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (12 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (11 messages🔥):

Links mentioned:


HuggingFace ▷ #computer-vision (8 messages🔥):

Link mentioned: tonyassi/sales1 · Datasets at Hugging Face: no description found


HuggingFace ▷ #NLP (10 messages🔥):

Link mentioned: Starter Tutorial (Local Models) - LlamaIndex: no description found


HuggingFace ▷ #diffusion-discussions (11 messages🔥):

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (282 messages🔥🔥):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #announcements (5 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):

obiefernandez: I signed up but it's not clear what the unique value proposition is


OpenRouter (Alex Atallah) ▷ #general (200 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #general (41 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (2 messages):


Modular (Mojo 🔥) ▷ #📺︱youtube (3 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #✍︱blog (1 messages):

Zapier: Modular: MAX Graph API Tutorial https://www.modular.com/blog/max-graph-api-tutorial


Modular (Mojo 🔥) ▷ #announcements (1 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #🔥mojo (120 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #nightly (17 messages🔥):

Links mentioned:


Eleuther ▷ #general (23 messages🔥):


Eleuther ▷ #research (79 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (45 messages🔥):

Link mentioned: Understanding the Covariance Structure of Convolutional Filters: Neural network weights are typically initialized at random from univariate distributions, controlling just the variance of individual weights even in highly-structured operations like convolutions. Re...


Eleuther ▷ #interpretability-general (1 messages):

ocg6377: I might also be interested in helping, depending on what's needed


Eleuther ▷ #lm-thunderdome (5 messages):

Link mentioned: lm-evaluation-harness/lm_eval/models/utils.py at a9eaaf46f1e246e5ce090e37f2f99fe1cfe5a919 · EleutherAI/lm-evaluation-harness: A framework for few-shot evaluation of language models. - EleutherAI/lm-evaluation-harness


CUDA MODE ▷ #general (3 messages):


CUDA MODE ▷ #cuda (19 messages🔥):

Links mentioned:


CUDA MODE ▷ #torch (39 messages🔥):

Links mentioned:


CUDA MODE ▷ #beginner (3 messages):

Link mentioned: Lecture 3: Getting Started With CUDA for Python Programmers: Recording on Jeremy's YouTube https://www.youtube.com/watch?v=nOxKexn3iBoSupplementary Content: https://github.com/cuda-mode/lecture2/tree/main/lecture3Speak...


CUDA MODE ▷ #pmpp-book (4 messages):


CUDA MODE ▷ #off-topic (3 messages):


CUDA MODE ▷ #triton-puzzles (1 messages):

Links mentioned:


CUDA MODE ▷ #llmdotc (65 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #blog (4 messages):


LlamaIndex ▷ #general (130 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):

pier1337: What’s the state of the art for RAGs in May?


Interconnects (Nathan Lambert) ▷ #news (25 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #other-papers (10 messages🔥):

Link mentioned: Tweet from Phillip Isola (@phillip_isola): We survey evidence from the literature, then provide several new results including: As LLMs get bigger and better, they learn representations that are more and more similar to those learned by visi...


Interconnects (Nathan Lambert) ▷ #ml-questions (42 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (24 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #posts (3 messages):


LangChain AI ▷ #general (73 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #langserve (2 messages):


LangChain AI ▷ #share-your-work (2 messages):

Link mentioned: Crystal Cost Demo: In this video, I give a quick demo of Crystal Cost, an AI-powered streamlit app that simplifies data monitoring on data warehouses. Crystal Cost uses natural language processing and agents to query da...


LAION ▷ #general (51 messages🔥):

Links mentioned:


LAION ▷ #research (16 messages🔥):

Links mentioned:


OpenInterpreter ▷ #general (35 messages🔥):

Link mentioned: Thank You Sticker Thanks Sticker GIF - Thank You Sticker Thanks Sticker Line Sticker - Discover & Share GIFs: Click to view the GIF


OpenInterpreter ▷ #O1 (23 messages🔥):


OpenInterpreter ▷ #ai-content (2 messages):

Link mentioned: Jack Mielke's AI podcast notes from #176 Linus Torvalds (Creator of Linux): Checkout the AI podcast notes created using Snipd


Datasette - LLM (@SimonW) ▷ #ai (58 messages🔥🔥):

Links mentioned:


Datasette - LLM (@SimonW) ▷ #llm (1 messages):


Latent Space ▷ #ai-general-chat (56 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

Link mentioned: LLM Paper Club (Eugene on Evals) · Zoom · Luma: Eugene is walking us thru ALL the evals: https://eugeneyan.com/writing/evals/ Also submit and vote for our next paper:…


OpenAccess AI Collective (axolotl) ▷ #general (33 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (6 messages):


OpenAccess AI Collective (axolotl) ▷ #general-help (2 messages):


OpenAccess AI Collective (axolotl) ▷ #datasets (3 messages):

Link mentioned: TIGER-Lab/MMLU-Pro · Datasets at Hugging Face: no description found


OpenAccess AI Collective (axolotl) ▷ #runpod-help (1 messages):


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (8 messages🔥):

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


Cohere ▷ #general (16 messages🔥):


Cohere ▷ #project-sharing (2 messages):


tinygrad (George Hotz) ▷ #general (2 messages):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (14 messages🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #ai-town-discuss (5 messages):

Link mentioned: AI Town on HuggingFace - a Hugging Face Space by radames: no description found


AI Stack Devs (Yoko Li) ▷ #ai-town-dev (4 messages):

Link mentioned: AI Town on HuggingFace - a Hugging Face Space by radames: no description found


AI Stack Devs (Yoko Li) ▷ #ai-raspberry-pi (1 messages):

tommy1901: just gonna posting some cool stuff here


DiscoResearch ▷ #general (9 messages🔥):

Links mentioned:


Skunkworks AI ▷ #announcements (1 messages):


MLOps @Chipro ▷ #general-ml (1 messages):

Link mentioned: The Past, Present, and Future of AI Hardware - SingleLunch: no description found