Frozen AI News archive

Jamba: Mixture of Architectures dethrones Mixtral

**AI21 labs** released **Jamba**, a **52B parameter MoE model** with **256K context length** and open weights under Apache 2.0 license, optimized for single A100 GPU performance. It features a unique blocks-and-layers architecture combining transformer and MoE layers, competing with models like **Mixtral**. Meanwhile, **Databricks** introduced **DBRX**, a **36B active parameter MoE model** trained on **12T tokens**, noted as a new standard for open LLMs. In image generation, advancements include **Animatediff** for video-quality image generation and **FastSD CPU v1.0.0 beta 28** enabling ultra-fast image generation on CPUs. Other innovations involve style-content separation using **B-LoRA** and improvements in high-resolution image upscaling with **SUPIR**.

Canonical issue URL

It's a banner week for MoE models, with DBRX yesterday and Qwen releasing a small MoE model today. However we have to give the top spot to yet another monster model release...

The recently $200m richer AI21 labs released Jamba today (blog, HF, tweet, thread from in person presentation). The headline details are:

It is notable both for its performance in its weight class (we'll come back to what "weight class" now means):

image.png

and for its throughput + memory requirements in long context scenarios:

image.png

re: weight class: It seems every design decision was taken to maximize the performance gained from a single A100:

"As depicted in the diagram below, AI21’s Jamba architecture features a blocks-and-layers approach that allows Jamba to successfully integrate the two architectures. Each Jamba block contains either an attention or a Mamba layer, followed by a multi-layer perceptron (MLP), producing an overall ratio of one Transformer layer out of every eight total layers.

The second feature is the utilization of MoE to increase the total number of model parameters while streamlining the number of active parameters used at inference—resulting in higher model capacity without a matching increase in compute requirements. To maximize the model’s quality and throughput on a single 80GB GPU, we optimized the number of MoE layers and experts used, leaving enough memory available for common inference workloads.

In a step ahead of Together's preceding StripedHyena, Jamba juices up the classic Mamba architecture with transformer and MoE layers:

image.png

They released a base model, but it comes ready with Huggingface PEFT support. This actually looks like a genuine Mixtral competitor, and that's only good things for the open AI community.


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence. Comment crawling still not implemented but coming soon.

Large Language Models

Stable Diffusion & Image Generation

AI Assistants & Agents

AI Hardware & Performance

Memes & Humor

AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

TO BE COMPLETED


PART 0: Summary of Summaries of Summaries


PART 1: High level Discord summaries

LM Studio Discord


Unsloth AI (Daniel Han) Discord

Breaking the Ice with Unsloth AI: Engineers have embraced tips and tricks for using Unsloth's template system, with the community finding practical benefits like reduced model output anomalies. Regular updates (2-3 times weekly) ensure that performance improvements continue, while installation instructions optimize setup times on Kaggle.

Gaming Mingle Amidst Coding Jungle: Technical exchanges were accompanied by lighter conversations, including game developer talks and shared gaming experiences—in particular, constructing a demo app using AI assistance, bridging entertainment with machine learning.

Layering it on Thick: Unsloth AI discussions have calved off into deeper explorations, including leveraging optimizer adjustments in resuming fine-tuning from checkpoints and proper chat template integration for various LLMs. The community also spotlighted key resources—Github repositories, Colab notebooks, and educational YouTube videos—for fine-tuning LLMs.

Modeling Showcase Spotlight: The community proudly presented adaptations, like converting the Lora Adapter for Tinyllama, and shared details of the Mischat model, which was fine-tuned using Unsloth's methodologies. A member introduced an AI digest on their Substack blog, summarizing recent AI developments.

Quantum Leap in Quantization: AI enthusiasts investigated specialized techniques such as LoRA training conversation, embedding quantization for faster retrieval, and the emerging QMoE compression framework. A newly introduced LISA strategy that streamlines fine-tuning across layers attracted significant attention for its memory efficiency.


Nous Research AI Discord


Stability.ai (Stable Diffusion) Discord

Links from the discussion included resources and tools:


Perplexity AI Discord

DBRX Breaks Through to Perplexity Labs: Databricks' DBRX language model has made waves by outclassing GPT-3.5 and proving competitive with Gemini 1.0 Pro, with favorable performance in math and coding benchmarks, which can be explored at Perplexity Labs.

The Developer's Dilemma: Perplexity vs. Claud: Engineers have debated whether Perplexity Pro or Claud Pro better suits their workflow, with a bias toward Perplexity for its transparency. Various model strengths like Claude 3 Opus were scrutinized, while Databricks' DBRX was singled out for its impressive math and coding capabilities.

Perplexity API Performs a Speedrun: The sonar-medium-online model showcased an unexpected speed increase, reaching or even surpassing that of sonar-small-online with higher quality output. Yet, inconsistencies surfaced with API responses compared to the Perplexity web interface, like the failure to retrieve "Olivia Schough spouse" data, prompting discussion on whether extra parameters could correct this.

Sharing Insights and Laughs: Community interaction included debunking a supposed Sora text-to-video model as a rickroll, emphasizing the importance of shareability for threads, and exploring varied search queries on Perplexity AI, ranging from coherent C3 models to French translations for "Perplexityai."

Vision Support Still in the Dark: Despite inquiries, Vision support for the API remains absent, as indicated by humorous responses about the current lack of even citations, suggesting no immediate plans for inclusion.


Latent Space Discord

Claude Takes the Terraform Crown: In the IaC domain, Claude has outshined its peers in generating Terraform scripts, with a comparison blog post on TerraTeam's website spotlighting its superior performance. The meticulous comparison can be accessed at TerraTeam's blog.

DBRX-Instruct Flexes Its Parameters: Databricks has stepped into the spotlight with DBRX-Instruct, a 132 billion parameter Mixture of Experts model that underwent a costly ($10M) and lengthy (2 months) training on 3072 NVIDIA H100 GPUs. Insights about DBRX-Instruct are split between Vitaliy Chiley's tweet and Wired's article.

Licensing Logistics Linger for DBRX: The community scrutinized DBRX's licensing terms, with members strategizing on how to best engage with the model within its usage boundaries. Key insight came from the shared legal concerns and strategies, including Amgadoz's spotlight on Databricks' open model license.

TechCrunch Questions DBRX's Market Muscle: A discussion was sparked by TechCrunch's critical analysis of Databricks' $10M DBRX investment, contrasting it against the already established OpenAI's GPT series. TechCrunch challenged the competitive edge provided by such investments, and a full read is recommended at TechCrunch.

Emotionally Intelligent Chatbots Get High Fives: Hume AI captured attention with its emotionally perceptive chatbot, adept at analyzing and responding to emotions. This disruptive emotional detection capability prompted a mix of excitement and practical use case discussions among members, including 420gunna sharing the Hume AI demo and a related CEO interview.

Mamba Slithers into the Spotlight: In discussions, the Mamba model was singled out for its innovation in the Transformer space, addressing efficiency woes effectively. The potent conversation revolved around Mamba's prowess and architectural decisions aimed at enhancing computational efficiency.

Fine-Tuning Finesse: The topic of fine-tuning Whisper, OpenAI's automatic speech recognition model, was dissected, with consensus that it's advisable when dealing with scarce language resources or specialized terminology in audio.

Cosine Similarity Crosstalk: The group engaged in technical tete-a-tetes over the use of cosine similarity in embeddings, casting doubts on its effectiveness as a semantic similarity measure. The pivot of the discussion was the paper titled "Is Cosine-Similarity of Embeddings Really About Similarity?", which members used as a reference point.

Screen Sharing Snafus: Technical trials with Discord's screen sharing triggered community troubleshooting, including workaround sharing and a collective call for Discord to enhance this feature. Members shared practical solutions to address the ongoing screen sharing issues.


Eleuther Discord


OpenAI Discord


HuggingFace Discord

Stable Diffusion Steps up for Solo Performances: Discussions on Stable Diffusion focused on generating new images from a list, but the existing pipeline handles single images. For personalized text-to-image models, DreamBooth emerged as a favorite, while the Marigold depth estimation pipeline is set for integration with new modalities like LCM.

AI Engineers Seek Smarter NLP Navigation: Engineers sought roadmaps for mastering NLP in 2024, with recommendations including "The Little Book of Deep Learning" and Karpathy's "Zero to Hero" playlist. Others explored session-based recommendation systems, questioning the efficacy of models like GRU4Rec and Bert4Rec, while loading errors with 'facebook/bart-large-cnn' prompted calls for help. Suggestions for managing the infinitely generative behavior of LLMs included Supervised Fine Tuning (SFT) and tweaks to repetition penalties.

Accelerating GPU Gains with MPS and Sagemaker: macOS users gained an advantage with MPS support now part of key training scripts, while a discussion on AWS SageMaker highlighted NVIDIA Triton and TensorRT-LLM for benchmarking GPU-utilizing model latency, cost, and throughput.

Innovations and Resources in the Computer Vision Sphere: Amidst efforts to utilize stitched images for training models, individuals also wrestled with fine-tuning DETR-ResNet-50 on specific datasets and investigated zero-shot classifier tuning for beginners. There was also an SOS for non-gradio_client testing methods for instruct pix2pix demos, with the community eager to recommend alternatives and resources.

DL Models in the Spotlight: The NLP community is examining papers on personalizing text-to-image synthesis to conform closely to text prompts. The RealCustom paper discusses balancing subject resemblance with textual control, and another study addresses text alignment in personalized images, as referenced on arXiv.


OpenInterpreter Discord


Modular (Mojo 🔥) Discord

Bug Squashing in VSCode Debugging: A GitHub-reported VSCode debugging issue with the Mojo plugin was resolved using a recommended workaround that proved successful on a MacBook.

Mojo and MAX Updates Make Headlines: The Mojo language style guide is now available, as is moplex, a new complex number library on GitHub. MAX 24.2 updates include the adoption of List over DynamicVector as referenced in the changelog.

Learning Resources Stand Out: A free chapter from Rust for Rustaceans was recommended for understanding Rust's lifetime management, while Modular's latest tweets garner attention without spawning further dialogue.

Open Source Embrace Boosts Mojo's Modularity: Modular has open-sourced the Mojo standard library under Apache 2, with nightly builds accessible, and MAX 24.2 introduces improved support for dynamic input shapes as demonstrated in their blog.

API Discrepancies and Enhancements Discussed: Users discussed inconsistencies between Mojo and Python APIs concerning TensorSpec, directing others to MAX Engine runtime documentation and MAX's example repository for clarity.

Open Source and Nightly Builds Beckon Collaboration: Developers are invited to jump on the open-source bandwagon with the Modular open-source initiative, which includes Mojo standard library updates and new features lined up in their latest changelog, while MAX platform's evolution with v24.2 offers new capabilities, particularly in dynamic shapes.


OpenRouter (Alex Atallah) Discord

Cheer for cheerful_dragon_48465: A username, cheerful_dragon_48465, received praise for being amusing, and Alex Atallah signaled an upcoming announcement that will highlight a notable user contribution.

Midnight Rose Clamors for Clarity: The Midnight Rose model was unresponsive without errors, leading to confusion among users before the OpenRouter team resolved the issue, yet the underlying problem remains unsolved.

Size Matters in Tokens: Users discussed the discrepancies in context sizes for Gemini models, which are measured in characters, not tokens, causing confusion, and acknowledged the need for better clarification on the topic.

Testing Troubles with Gemini Pro 1.5: Users facing Error 503 with Gemini Pro 1.5 were informed that the issues arose because the model was still in the testing phase, indicating a gap between OpenRouter's service expectations and reality.

The Ethereum Payment Conundrum: OpenRouter's shift to requiring payments through the ETH network via Coinbase Commerce, and the subsequent discussion on incentives for US bank transfers, highlighted the evolving landscape of crypto payments in the AI sector.


CUDA MODE Discord


LlamaIndex Discord


OpenAccess AI Collective (axolotl) Discord

Databricks Drops DBRX: Databricks introduced DBRX Base and DBRX Instruct, boasting 132B total parameters, outshining LLaMA2-70B and other models, with an open model license and insights provided in their technical blog.

Axolotl Devs Debugging: The Axolotl AI Collective has rectified a trainer.py batch size bug and discussed technical issues like transformer incompatibilities, DeepSpeed and PyTorch binary problems, and large model loading challenges with qlora+fsdp.

Innovating Jamba and LISA: AI21 Labs revealed Jamba, an architecture capable of handling 256k token context on A100 80GB GPUs, while discussion on LISA's superiority over LoRA in instruction following tasks occurs, referencing PRs #701 and #711 in the LMFlow repository.

Performance with bf16: A lively debate took place around using bf16 precision for both training and optimization, citing torchtune team's findings on memory efficiency and stability akin to fp32, sparking interest in its broader implementation.

Resource Hunt for Fine-Tuning Finesse: Community members seek comprehensive educational materials for fine-tuning or training open source models, indicating a preference for varied formats like blogs, articles, and videos, aiming for a strong foundation before diving into axolotl.


LAION Discord


tinygrad (George Hotz) Discord

Tinygrad Tightens the Screws: Dynamic discussions about tinygrad reveal attempts to close the performance gap with PyTorch, through heuristics for operations like gemv and gemm and direct manipulation of GPU kernels. Insights include kernel fusion challenges, potential view merging optimizations, and community-driven documentation efforts.

NVIDIA Claims the Crown in MLPerf: Recent MLPerf Inference v4.0 results sparked conversation, noting how NVIDIA continues to lead in performance metrics, with Qualcomm showing strong results and Habana’s Gaudi2 demonstrating its lack of design for inference tasks.

SYCL Stepping Up to CUDA: A tweet highlighted SYCL as a promising alternative to NVIDIA’s CUDA, stirring anticipation for wider industry adoption and a break from current monopolistic trends in AI hardware.

API Allegiances and Industry Impact: Members weighed in on OpenCL’s diminished utilization and the potential of Vulkan for achieving uniform hardware acceleration interfaces, debating their respective roles in the larger ecosystem.

View Merging on the Horizon: The discussions also probed the refinement of tinygrad's ShapeTracker to potentially consolidate views, considering the importance of tensor transformation histories and backpropagation functionality when contemplating structural changes.


LangChain AI Discord

OpenGPTs Discussion Welcomes Engineers: A new channel for OpenGPTs project on GitHub has been introduced encouraging contributions and dialogue amongst the community.

JavaScript Chatbots versus Document Fetchers: AI engineers explore building dynamic chatbots with JavaScript, diverging from static document retrieval. For guidance, a Colab notebook has been shared.

Deploying with Custom Domains Hiccup: Deploying FastAPI RAG apps with LangChain on custom domains like github.io is sparking curiosity; yet, documentation discrepancies on LangChain Pinecone integration pose challenges that await resolution.

LangSmith Traces AI's Steps: Using LangChain's LangSmith for tracing AI actions employs environment variables such as LANGCHAIN_TRACING_V2, which offers granular logging capabilities.

Tutorial Unlocks PDF to JSON Conversion Mysteries: A new YouTube tutorial breaks down the conversion of PDFs to JSON using LangChain's Output Parsers and GPT, simplifying a once complex task. The community's insights are requested to enhance such educational content.


Interconnects (Nathan Lambert) Discord


DiscoResearch Discord

DBRX Instruct Makes a Grand Entrance: Databricks unveiled a new 132 billion parameter sparse MoE model, DBRX Instruct, trained on a staggering 12 trillion tokens, boasting prowess in few-turn dialogues, alongside releasing DBRX Base under an open license, furnished with insights in their blog post.

DBRX's Inner Workings Decoded: DBRX distinguishes itself with a merged attention mechanism, distinct normalization technique, and a unique tokenization method that has been refined through various bug fixes, with its technical intricacies documented on GitHub.

Hands-On with DBRX Instruct: AI enthusiasts can now experiment with DBRX Instruct through an interactive Hugging Face space, complete with a system prompt for tailoring response styles.

Mixtral's Multilingual Muscles Flexed for Free: Mixtral's translation API can be tapped into without charge via groq, subject to rate limits, and open for community-driven experimentation.

Occi 7B Outshines in Translation Quality: Users have noted the exceptional translation fidelity of Occi 7B via the occiglot/occiglot-7b-de-en-instruct model and have embarked on a quest to gauge the translation caliber across services like DisCoLM, GPT-4, Deepl, and Azure Translate, showcasing their efforts on Hugging Face.


Alignment Lab AI Discord


LLM Perf Enthusiasts AI Discord


Skunkworks AI Discord


The Datasette - LLM (@SimonW) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

LM Studio ▷ #💬-general (335 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (72 messages🔥🔥):

Links mentioned:


LM Studio ▷ #announcements (2 messages):

Links mentioned:


LM Studio ▷ #🧠-feedback (1 messages):


LM Studio ▷ #🎛-hardware-discussion (109 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (96 messages🔥🔥):

Links mentioned:


LM Studio ▷ #langchain (1 messages):

Unfortunately, there's insufficient context and content to extract topics, discussion points, links, or blog posts of interest from the provided message. The single message fragment you've provided does not contain enough information for a summary. Please provide more messages for a detailed summary.


LM Studio ▷ #amd-rocm-tech-preview (92 messages🔥🔥):

Links mentioned:


LM Studio ▷ #crew-ai (4 messages):


Unsloth AI (Daniel Han) ▷ #general (293 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (21 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (202 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (7 messages):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (25 messages🔥):

Links mentioned:


Nous Research AI ▷ #ctx-length-research (6 messages):


Nous Research AI ▷ #off-topic (15 messages🔥):

Link mentioned: Cohere int8 & binary Embeddings: Cohere int8 & binary Embeddings - Scale Your Vector Database to Large Datasets#ai #llm #ml #deeplearning #neuralnetworks #largelanguagemodels #artificialinte...


Nous Research AI ▷ #interesting-links (11 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (285 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (55 messages🔥🔥):

Link mentioned: Tweet from Cody Blakeney (@code_star): It’s finally here 🎉🥳 In case you missed us, MosaicML/ Databricks is back at it, with a new best in class open weight LLM named DBRX. An MoE with 132B total parameters and 32B active 32k context len...


Nous Research AI ▷ #project-obsidian (3 messages):


Nous Research AI ▷ #rag-dataset (51 messages🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (106 messages🔥🔥):

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (436 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):


Perplexity AI ▷ #general (326 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (14 messages🔥):


Perplexity AI ▷ #pplx-api (15 messages🔥):


Latent Space ▷ #ai-general-chat (110 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (3 messages):

Link mentioned: Latent Space (Paper Club & Other Events) · Events Calendar: View and subscribe to events from Latent Space (Paper Club & Other Events) on Luma. Latent.Space events. PLEASE CLICK THE RSS LOGO JUST ABOVE THE CALENDAR ON THE RIGHT TO ADD TO YOUR CAL. "Ad...


Latent Space ▷ #llm-paper-club-west (183 messages🔥🔥):

Links mentioned:


Eleuther ▷ #general (174 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (35 messages🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (5 messages):

Link mentioned: grok-1/run.py at main · xai-org/grok-1: Grok open release. Contribute to xai-org/grok-1 development by creating an account on GitHub.


Eleuther ▷ #interpretability-general (39 messages🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (5 messages):


OpenAI ▷ #ai-discussions (113 messages🔥🔥):


OpenAI ▷ #gpt-4-discussions (14 messages🔥):


OpenAI ▷ #prompt-engineering (30 messages🔥):


OpenAI ▷ #api-discussions (30 messages🔥):


HuggingFace ▷ #general (103 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (2 messages):

Link mentioned: Groking Groq: A Deep Dive on Deep Learning: To "Grok" is to learn something deeply- as if you're drinking it in. AI has a way of requiring that you Grok a number of seemingly unrelated topics; making i...


HuggingFace ▷ #cool-finds (16 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (9 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (3 messages):

Links mentioned:


HuggingFace ▷ #core-announcements (1 messages):

Link mentioned: Build software better, together: GitHub is where people build software. More than 100 million people use GitHub to discover, fork, and contribute to over 420 million projects.


HuggingFace ▷ #computer-vision (15 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (12 messages🔥):


HuggingFace ▷ #diffusion-discussions (14 messages🔥):

Links mentioned:


OpenInterpreter ▷ #general (67 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (86 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (2 messages):

Link mentioned: Open Interpreter - Advanced Experimentation: ➤ Twitter - https://twitter.com/techfrenaj➤ Twitch - https://www.twitch.tv/techfren➤ Discord - https://discord.com/invite/z5VVSGssCw➤ TikTok - https://www....


Modular (Mojo 🔥) ▷ #general (75 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (5 messages):


Modular (Mojo 🔥) ▷ #✍︱blog (4 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #announcements (1 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #🔥mojo (54 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (3 messages):


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (1 messages):


Modular (Mojo 🔥) ▷ #🏎engine (6 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (4 messages):


OpenRouter (Alex Atallah) ▷ #general (144 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #general (40 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (8 messages🔥):

Links mentioned:


CUDA MODE ▷ #cuda (6 messages):

Links mentioned:


CUDA MODE ▷ #torch (2 messages):


CUDA MODE ▷ #jobs (5 messages):

Link mentioned: CAREERS AT NVIDIA: no description found


CUDA MODE ▷ #beginner (6 messages):

Link mentioned: Install WSL: Install Windows Subsystem for Linux with the command, wsl --install. Use a Bash terminal on your Windows machine run by your preferred Linux distribution - Ubuntu, Debian, SUSE, Kali, Fedora, Pengwin,...


CUDA MODE ▷ #ring-attention (15 messages🔥):

Links mentioned:


CUDA MODE ▷ #off-topic (2 messages):

Links mentioned:


CUDA MODE ▷ #triton-puzzles (54 messages🔥):

Links mentioned:


LlamaIndex ▷ #blog (3 messages):

Link mentioned: RSVP to LLM x Law Hackathon @Stanford #3 | Partiful: As artificial intelligence (AI) continues to revolutionize industries across the globe, the legal sector is no exception. LLMs, a foundation model capable of understanding and generating natural langu...


LlamaIndex ▷ #general (120 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (3 messages):

Link mentioned: Elevate Responses: RAG with LlamaIndex & MongoDB: no description found


OpenAccess AI Collective (axolotl) ▷ #general (85 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (24 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (3 messages):


LAION ▷ #general (75 messages🔥🔥):

Links mentioned:


LAION ▷ #research (31 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #general (43 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (60 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #announcements (1 messages):

Link mentioned: GitHub - langchain-ai/opengpts: Contribute to langchain-ai/opengpts development by creating an account on GitHub.


LangChain AI ▷ #general (72 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (3 messages):

Links mentioned:


LangChain AI ▷ #tutorials (1 messages):

Link mentioned: How to convert a PDF to JSON using LangChain Output Parsers and GPT: This video tutorial demonstrates how to convert a PDF to JSON using LangChain's Output Parsers and GPT.A task like this used to be complicated but can now be...


Interconnects (Nathan Lambert) ▷ #news (30 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (18 messages🔥):

Links mentioned:


DiscoResearch ▷ #general (8 messages🔥):

Links mentioned:


DiscoResearch ▷ #discolm_german (10 messages🔥):

Links mentioned:


Alignment Lab AI ▷ #general-chat (7 messages):

Link mentioned: Introducing DBRX: A New State-of-the-Art Open LLM | Databricks: no description found


LLM Perf Enthusiasts AI ▷ #irl (4 messages):

Link mentioned: RSVP to Coffee + Cowork | Partiful: Hi everyone! The Exa team is excited to host a pop-up coffeeshop and co-work in our home office this Saturday! Feel free to stop by for some very fancy coffee/matcha + breakfast, or bring a laptop an...


Skunkworks AI ▷ #general (1 messages):


Skunkworks AI ▷ #off-topic (1 messages):

pradeep1148: https://www.youtube.com/watch?v=LWz2QaSRl2Y