Frozen AI News archive

OpenAI's PR Campaign?

**OpenAI** faces user data deletion backlash over its new partnership with StackOverflow amid GDPR complaints and US newspaper lawsuits, while addressing election year concerns with efforts like the Media Manager tool for content opt-in/out by 2025 and source link attribution. **Microsoft** develops a top-secret airgapped GPT-4 AI service for US intelligence agencies. OpenAI releases the Model Spec outlining responsible AI content generation policies, including NSFW content handling and profanity use, emphasizing clear distinctions between bugs and design decisions. **Google DeepMind** announces **AlphaFold 3**, a state-of-the-art model predicting molecular structures with high accuracy, showcasing cross-domain AI techniques. New research on **xLSTM** proposes scaling LSTMs to billions of parameters, competing with transformers in performance and scaling. Microsoft introduces **vAttention**, a dynamic memory management method for efficient large language model serving without PagedAttention.

Canonical issue URL

In a time when StackOverflow users are deleting their data in response to the new OpenAI partnership (with SO responding poorly), with GDPR complaints and US newspaper lawsuits and the NYT accusing it of scraping 1m hours of YouTube, and a general state of anxiety in a big election year (something OpenAI has explicitly addressed), there seems to be a recent pushback this week to highlight OpenAI's efforts to be a trustworthy institution:

image.png

As @sama says: "We will listen, debate, and adapt this over time, but i think it will be very useful to be clear when something is a bug vs. a decision.". Per Joanne Jang:

image.png

The whole model spec is worth reading and seems very thoughtfully designed.


Table of Contents

[TOC]


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

AI Models and Architectures

Scaling and Efficiency

Open Source Models

Benchmarks and Evaluation

Ethics and Safety


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

Advances in AI Models and Hardware

Emerging AI Applications and Developer Tools

AI Ethics, Regulation and Societal Impact


AI Discord Recap

A summary of Summaries of Summaries

  1. New AI Model Releases and Comparisons: The AI community saw a flurry of new models released, including Idefics2 8B Chatty, CodeGemma 1.1, DeepSeek-V2 with 236B parameters, IBM Granite code models, and Llama 3 Gradient 4194k with extended context length. Discussions compared their performance, use cases, and the implications of larger models like DeepSeek-V2. HuggingFace Announcement

  2. Optimization Techniques for AI Models and Code: Extensive discussions revolved around optimizing AI models, code, and workflows, covering techniques like quantization, LoRA for diffusion model fine-tuning, torch.compile challenges with dynamic batch shapes, CUDA kernel fusion and performance analysis, and string concatenation optimization strategies in Mojo. Examples include CUDA kernel fusion PR, LoRA endorsement, and Mojo StringBuilder.

  3. Large Language Model (LLM) Applications and Tools: Several tools and applications leveraging LLMs were showcased, such as LlamaIndex for building agentic systems, Langchain for streamlining customer support, Athena for autonomous data workflows, OpenDevin as an open-source AI coding assistant, and Quickscope for automated game testing in Unity. Relevant examples include the LlamaIndex course, Langchain's LangGraph article, and Quickscope's launch.

  4. Ethical AI Development and Guidance: Initiatives focused on responsible AI practices, such as OpenAI's Model Spec draft for guiding model behavior, AlphaFold 3 implementation in PyTorch for biomolecular structure prediction, and discussions on creating inclusive language datasets for German AI assistants. Examples include the OpenAI Model Spec document, AlphaFold 3 PyTorch implementation, and gender-inclusive German resources.


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord


Nous Research AI Discord


OpenAI Discord


Unsloth AI (Daniel Han) Discord


LM Studio Discord


Perplexity AI Discord


HuggingFace Discord

AI's Grand Slam: New Models Take the Field: The AI field has introduced a slew of new models, including Idefics2 8B Chatty, CodeGemma 1.1 focused on coding tasks, and the gargantuan DeepSeek-V2 with 236B parameters. For code-specific needs, there's IBM Granite, and for enhanced context windows, we've got Llama 3 Gradient 4194k.

Sharpening the AI Saw: AI enthusiasts tackled diverse integration challenges, grappling with the implementation and efficacy of models like LangChain with DSpy and LayoutLMv3, and debated their practical utility against stalwarts like BERT. They delved into using Gradio Templates to prototype AI demos. Some sought knowledge on using CPU-efficient models for teaching purposes by trying out ollama and llama cpp python. Meanwhile, others looked into predictive open-source AI tools for streamlining repetitive tasks.

AI Illuminates Dark Data Corners: In the realm of AI datasets, there's a focus on improving transparency, epitomized by a YouTube tutorial on converting datasets from parquet to CSV using Polars. Furthermore, a succinct analogy for Multimodal AI was presented in a two-minute YouTube video, shedding light on the capabilities of models like Med-Gemini.

Tools of the Trade - Enhancing Developer Arsenal: In the quest for automating routine processes, a member shared an article about using Langchain’s LangGraph to augment customer support. When it comes to diffusion models, the advice is coalescing around using LoRA as the go-to method for fine-tuning tasks. Meanwhile, the visual crowd embraced the new adlike library and celebrated the enhancements to HuggingFace's object detection guides, adding mAP tracking.

Building Tomorrow's Research Ecosystem: The creative community teems with innovations like EurekAI, which promises a more organized research methodology, and Rubik's AI, seeking beta testers to refine its research assistant platform. An interesting experiment by Udio AI highlighted a fresh tune generated via AI, while BIND opens doors for utilizing protein-language models in drug discovery, offering a progressive GitHub resource at Chokyotager/BIND.


Eleuther Discord


Modular (Mojo 🔥) Discord


CUDA MODE Discord

Dynamic Batching Blues: AI engineers discussed the tribulations of using torch.compile for dynamic batch shapes, which causes excessive recompilations and impacts performance. While padding to static shapes can mitigate issues, full support for dynamic shapes, especially with nested tensors, awaits integration.

Triton's fp8 and Community Repo: Triton now includes support for fp8, as per updates on the official GitHub referring to a fused attention example. There's a community push to centralize Triton resources; a new community-driven Triton-index repository aims to catalog released kernels, and there's talk of curating a dataset specifically for Triton kernels, reflecting a drive for collaborative development.

CUDA Quest for GPU Proficiency: A multi-threaded conversation shone light on the optimization journey in CUDA, revealing a merged pull request to fuse residual and layernorm forward in CUDA, analyses of kernel performance metrics, and the quest to manage communication overheads in distributed training for optimal utilization of GPU architecture.

Optimization vs. Orientation for NHWC Tensors: The performance conundrum for tensor normalization orientation surfaced, leaving engineers pondering whether permuting tensors from NHWC to NCHW is more efficient than using NHWC-specific algorithms on GPUs, despite the risk of access pattern inefficiencies.

Apple M4 Steals the Spotlight: In hardware news, Apple heralded its M4 chip, designed to uplift the iPad Pro. Meanwhile, an AI engineer highlighted the capability of "panther lake" to deliver 175 TOPS, underscoring the rapid advancements and competition in chip performance.


OpenRouter (Alex Atallah) Discord


LAION Discord


OpenInterpreter Discord


LangChain AI Discord


LlamaIndex Discord

AI Education Leveling Up: LlamaIndex and deeplearning.ai announce a new course on creating agentic RAG systems, endorsed by AI expert Andrew Y. Ng. Engineers can learn about advanced concepts like routing, tool use, and sophisticated multi-step reasoning. Enroll here.

Scheduled Learning Opportunity: An upcoming LlamaIndex webinar spotlights OpenDevin, an open-source project by Cognition Labs designed to function as an autonomous AI engineer. The webinar is set for Thursday at 9am PT and is creating buzz for its potential to streamline coding and engineering tasks. Reserve your seat now.

Latest Tech from LlamaIndex: An update to LlamaIndex introduces the StructuredPlanningAgent, enhancing agents' task management by breaking them into smaller, more manageable sub-tasks. This development supports a range of agent workers, potentially boosting efficiency in tools like ReAct and Function Calling. Discover the influence of this tech.

Peering into Agent Observations: Engineers explore the feasibility of extracting detailed observation data from ReAct Agents and the utilization of local PDF parsing through PyMuPDF. Methods to improve the specificity and relevance of LLM (Large Language Model) responses, and the optimization of retrieval systems using reranking models, prompted a thorough technical exchange.

Towards Cooperative AI: A vibrant idea exchange occurred around multi-agent systems, envisioning a future with seamless agent collaboration and complex task execution. The concept nods to solutions like crewai and autogen, with additional focus on the capability of agents to create snapshots and rewind actions for enhanced operation.


OpenAccess AI Collective (axolotl) Discord

Layer Activation Unexpectedness: Discussions identified an anomaly where one layer in a model exhibited higher values than others, raising concerns and curiosity about the implications for neural network behavior and optimizer strategies.

LLM Training Data Discrepancies and Human Data Influence: It was noted that ChatQA is trained on a distinct mixture of data, contrasting with the GPT-4/Claude dataset used for most models, and the use of LIMA RP human data was highlighted for its potential to increase model training specificity.

Releasing RefuelLLM-2 to the Wild: RefuelLLM-2 has been open-sourced, boasting prowess in handling "unsexy data tasks,” with model weights available on Hugging Face and details shared via Twitter.

Practical Quantization Questions and GPU Quagmires: Queries were raised about creating a language-specific LLM and training with quantization on standard laptops, as well difficulties encountered with Cuda out of memory errors when using a config file for the phi3 mini 4K/128K FFT on 8 A100 GPUs, prompting a search for a working config example.

wandb Woes and Gradient Gamble: Members sought advice on Weights & Biases (wandb) configuration options and investigated strategies to handle the exploding gradient norm problem as well as considering trade-offs between 4-bit and 8-bit loading for efficiency versus model performance.


Interconnects (Nathan Lambert) Discord

LSTMs Strike Back: A recent paper discussed LSTMs scaled to billions of parameters, with potential LSTM enhancements like exponential gating and matrix memory to challenge Transformer dominance. There were concerns about flawed comparisons and a lack of hyperparameter tuning in the research.

AI's Behavioral Blueprint Unveiled: OpenAI's Model Spec draft was announced, designed to navigate model behavior in their API and implement reinforcement learning from human feedback (RLHF), as documented in Model Spec (2024/05/08).

Chatbot Reputation Under Microscope: A conversation emerged about how chatgpt2-chatbot could negatively impact LMsys' credibility, suggesting the system is overtaxed and unable to refuse requests. Licensing issues were also raised concerning chatbotarena's data releases without permissions from LLM providers.

Gemini 1.5 Pro Hits a High Note: Gemini 1.5 Pro was praised for its ability to transcribe podcast chapters accurately, incorporating timestamps despite some errors.

Awaiting the Snail's Wisdom: Community members showed anticipation and support for a seemingly important entity or event referred to as "snail," with posts suggesting a mix of awaiting news and summoning involvement from certain ranks.


Latent Space Discord


tinygrad (George Hotz) Discord

Tinygrad Tech Talk: Reshaping and Education: Discussions about Tinygrad's documentation on tensor reshaping sparked criticism for being too abstract, leading to a collaborative effort to demystify the concept through a community-created explanatory document. Advanced reshape optimizations, potentially using compile-time index calculations, were considered to enhance performance.

Tinygrad's BITCAST Clarified: There's active work on understanding and improving the BITCAST operation in tinygrad, as seen in a GitHub pull request, aiming to simplify certain operations and remove the need for arguments like "bitcast=false".

ML Concepts Demystified: A user conveyed the difficulty of deciphering machine learning terminology, specifically when simple concepts are buried under math jargon. This aligned with calls within the community for clearer and more approachable learning materials.

Tinygrad's No-Nonsense Policy: @georgehotz reinforced community guidelines, reminding members that the forum is not meant for beginner-level queries and that valuable time should not be taken for granted.

Engineering Discussions Advance in Sorting UOp Queries: The intricacies of Tinygrad’s operations, such as whether symbolic.DivNode should accept node operands, were debated, potentially signaling a future update to improve recursive handling within operations like symbolic.arange.


Cohere Discord


DiscoResearch Discord


Mozilla AI Discord


Datasette - LLM (@SimonW) Discord


Alignment Lab AI Discord


AI Stack Devs (Yoko Li) Discord


LLM Perf Enthusiasts AI Discord

GPT-4-turbo Hunt in Azure: An engineer is on the lookout for GPT-4-turbo 0429 availability in Azure regions, specifically mentioning operational issues with Sweden's Azure services.


The Skunkworks AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ▷ #general-chat (737 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ctx-length-research (13 messages🔥):


Nous Research AI ▷ #off-topic (4 messages):

Link mentioned: GitHub - secondlaw-ai/skyrim: 🌎 🤝 AI weather models united: 🌎 🤝 AI weather models united. Contribute to secondlaw-ai/skyrim development by creating an account on GitHub.


Nous Research AI ▷ #interesting-links (13 messages🔥):

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

Link mentioned: worldsim: no description found


Nous Research AI ▷ #general (345 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (37 messages🔥):

Links mentioned:


Nous Research AI ▷ #project-obsidian (1 messages):


Nous Research AI ▷ #bittensor-finetune-subnet (11 messages🔥):


Nous Research AI ▷ #rag-dataset (4 messages):


Nous Research AI ▷ #world-sim (93 messages🔥🔥):

Link mentioned: New Conversation - Eigengrau Rain: no description found


OpenAI ▷ #annnouncements (2 messages):


OpenAI ▷ #ai-discussions (305 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (7 messages):


OpenAI ▷ #prompt-engineering (30 messages🔥):


OpenAI ▷ #api-discussions (30 messages🔥):


Unsloth AI (Daniel Han) ▷ #general (108 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (13 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (194 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (18 messages🔥):

Links mentioned:


LM Studio ▷ #💬-general (112 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (21 messages🔥):

Link mentioned: T5: no description found


LM Studio ▷ #🧠-feedback (12 messages🔥):


LM Studio ▷ #📝-prompts-discussion-chat (2 messages):


LM Studio ▷ #⚙-configs-discussion (38 messages🔥):

Link mentioned: llava 1.5 invalid output after first inference (llamacpp server) · Issue #7060 · ggerganov/llama.cpp: I use this server config: "host": "0.0.0.0", "port": 8085, "api_key": "api_key", "models": [ { "model": "models/phi3_mini_mod...


LM Studio ▷ #🎛-hardware-discussion (68 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (41 messages🔥):


LM Studio ▷ #autogen (1 messages):


LM Studio ▷ #langchain (2 messages):


LM Studio ▷ #amd-rocm-tech-preview (3 messages):


LM Studio ▷ #crew-ai (1 messages):


LM Studio ▷ #🛠-dev-chat (18 messages🔥):

Links mentioned:


Perplexity AI ▷ #general (222 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (29 messages🔥):


Perplexity AI ▷ #pplx-api (11 messages🔥):

Link mentioned: Supported Models: no description found


HuggingFace ▷ #announcements (1 messages):


HuggingFace ▷ #general (198 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (3 messages):

Links mentioned:


HuggingFace ▷ #cool-finds (5 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (15 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (1 messages):

Since there's only a single message provided without any further context or discussion, the summary would only reflect that message:

There are no links or comments to add to this summary. If more messages or discussions were provided, they could be summarized accordingly.


HuggingFace ▷ #computer-vision (9 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (10 messages🔥):


HuggingFace ▷ #diffusion-discussions (9 messages🔥):


Eleuther ▷ #general (63 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (131 messages🔥🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (32 messages🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (10 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #general (90 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (1 messages):

ModularBot: From Modular: https://twitter.com/Modular/status/1788281021085225170


Modular (Mojo 🔥) ▷ #✍︱blog (1 messages):

Link mentioned: Modular: Developer Voices: Deep Dive with Chris Lattner on Mojo: We are building a next-generation AI developer platform for the world. Check out our latest post: Developer Voices: Deep Dive with Chris Lattner on Mojo


Modular (Mojo 🔥) ▷ #🔥mojo (67 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (14 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-blogs-vids (2 messages):


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (15 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #nightly (45 messages🔥):

Links mentioned:


CUDA MODE ▷ #general (23 messages🔥):

Link mentioned: Diffusion Inference Optimization: no description found


CUDA MODE ▷ #triton (10 messages🔥):

Links mentioned:


CUDA MODE ▷ #torch (2 messages):


CUDA MODE ▷ #algorithms (1 messages):

andreaskoepf: xLSTM paper is out: https://arxiv.org/abs/2405.04517


CUDA MODE ▷ #cool-links (2 messages):

Link mentioned: Tutorial: Matrix Transpose in CUTLASS: The goal of this tutorial is to elicit the concepts and techniques involving memory copy when programming on NVIDIA® GPUs using CUTLASS and its core backend library CuTe. Specifically, we will stud…


CUDA MODE ▷ #beginner (3 messages):


CUDA MODE ▷ #jax (1 messages):

Link mentioned: Multi chip performance in JAX: The larger the models we use get the more it becomes necessary to be able to perform training of machine learning models over multiple chips. In this blog post we will explain how to efficiently use G...


CUDA MODE ▷ #off-topic (4 messages):

Link mentioned: Apple introduces M4 chip: Apple today announced M4, the latest Apple-designed silicon chip delivering phenomenal performance to the all-new iPad Pro.


CUDA MODE ▷ #irl-meetup (5 messages):

Link mentioned: ICLR 2024 -The Twelfth International Conference on Learning Representations Whova Web Portal: May 7 – 11, 2024, Messeplatz 1, Postfach 277, A-1021 Wien


CUDA MODE ▷ #llmdotc (126 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #oneapi (3 messages):

Links mentioned:

  PyTorch

: no description found


OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):


OpenRouter (Alex Atallah) ▷ #general (115 messages🔥🔥):

Links mentioned:


LAION ▷ #general (106 messages🔥🔥):

Links mentioned:


LAION ▷ #research (6 messages):

Link mentioned: GitHub - lllyasviel/IC-Light: More relighting!: More relighting! Contribute to lllyasviel/IC-Light development by creating an account on GitHub.


OpenInterpreter ▷ #general (11 messages🔥):

Link mentioned: OpenPipe: Fine-Tuning for Developers: Convert expensive LLM prompts into fast, cheap fine-tuned models.


OpenInterpreter ▷ #O1 (83 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (10 messages🔥):

Link mentioned: GitHub - szczyglis-dev/py-gpt: Desktop AI Assistant powered by GPT-4, GPT-4 Vision, GPT-3.5, DALL-E 3, Langchain, Llama-index, chat, vision, voice control, image generation and analysis, autonomous agents, code and command execution, file upload and download, speech synthesis and recognition, access to Web, memory, prompt presets, plugins, assistants & more. Linux, Windows, Mac.: Desktop AI Assistant powered by GPT-4, GPT-4 Vision, GPT-3.5, DALL-E 3, Langchain, Llama-index, chat, vision, voice control, image generation and analysis, autonomous agents, code and command execu...


LangChain AI ▷ #general (49 messages🔥):

Link mentioned: Rubik's AI - AI research assistant & Search Engine: no description found


LangChain AI ▷ #langserve (13 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (5 messages):

Links mentioned:


LangChain AI ▷ #tutorials (1 messages):

ntelo007: What is the benefit of this?


LlamaIndex ▷ #announcements (2 messages):

Links mentioned:


LlamaIndex ▷ #blog (2 messages):

<ul>
  <li><strong>OpenDevin, the Autonomous AI Engineer**: [OpenDevin](https://twitter.com/llama_index/status/1787858033412063716) is an open-source autonomous AI engineer from <strong>@cognition_labs</strong>, capable of executing complex engineering tasks and collaborating on software projects.</li>
  <li><strong>StructuredPlanningAgent Enhances LlamaIndex**: The latest LlamaIndex update includes the <strong>StructuredPlanningAgent</strong> which assists agents in planning by breaking down tasks into sub-tasks, making them easier to execute. It supports various agent workers like ReAct and Function Calling. [Find out more](https://twitter.com/llama_index/status/1787971603936199118).</li>
</ul>

LlamaIndex ▷ #general (50 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (16 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #other-llms (1 messages):


OpenAccess AI Collective (axolotl) ▷ #general-help (8 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #datasets (2 messages):


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (1 messages):


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (19 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (5 messages):

Link mentioned: xLSTM: Extended Long Short-Term Memory: In the 1990s, the constant error carousel and gating were introduced as the central ideas of the Long Short-Term Memory (LSTM). Since then, LSTMs have stood the test of time and contributed to numerou...


Interconnects (Nathan Lambert) ▷ #random (24 messages🔥):

Link mentioned: Tweet from ハードはんぺん (@U8JDq51Thjo1IHM): I’m-also-a-good-gpt2-chatbot I’m-a-good-gpt2-chatbot ?? Quoting Jimmy Apples 🍎/acc (@apples_jimmy) @sama funny guy arnt you. Gpt2 back on lmsys arena.


Interconnects (Nathan Lambert) ▷ #reads (4 messages):

Link mentioned: Model Spec (2024/05/08): no description found


Interconnects (Nathan Lambert) ▷ #posts (4 messages):


Latent Space ▷ #ai-general-chat (26 messages🔥):

Links mentioned:


Latent Space ▷ #llm-paper-club-west (2 messages):


tinygrad (George Hotz) ▷ #general (14 messages🔥):

Link mentioned: UOps.BITCAST by chenyuxyz · Pull Request #3747 · tinygrad/tinygrad: implicitly fixed no const folding for bitcast


tinygrad (George Hotz) ▷ #learn-tinygrad (11 messages🔥):

Links mentioned:


Cohere ▷ #general (12 messages🔥):


Cohere ▷ #project-sharing (2 messages):

Link mentioned: no title found: no description found


Cohere ▷ #collab-opps (1 messages):

Link mentioned: Notion – The all-in-one workspace for your notes, tasks, wikis, and databases.: A new tool that blends your everyday work apps into one. It's the all-in-one workspace for you and your team


DiscoResearch ▷ #general (3 messages):


DiscoResearch ▷ #discolm_german (6 messages):

Links mentioned:


Mozilla AI ▷ #llamafile (5 messages):

Link mentioned: GitHub - Mozilla-Ocho/llamafile: Distribute and run LLMs with a single file.: Distribute and run LLMs with a single file. Contribute to Mozilla-Ocho/llamafile development by creating an account on GitHub.


Datasette - LLM (@SimonW) ▷ #ai (1 messages):

Link mentioned: Reddit - Dive into anything: no description found


Datasette - LLM (@SimonW) ▷ #llm (3 messages):

Link mentioned: Design and implement parameterization mechanism · Issue #4 · simonw/llm-evals-plugin: Initial thoughts here: #1 (comment) I want a parameterization mechanism, so you can run the same eval against multiple examples at once. Those examples can be stored directly in the YAML or can be ...


Alignment Lab AI ▷ #ai-and-ml-discussion (1 messages):

Links mentioned:


Alignment Lab AI ▷ #general-chat (2 messages):


AI Stack Devs (Yoko Li) ▷ #app-showcase (1 messages):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #team-up (1 messages):

jakekies: I want to join


LLM Perf Enthusiasts AI ▷ #openai (1 messages):