Frozen AI News archive

Zero to GPT in 1 Year

**GPT-4 Turbo** reclaimed the top leaderboard spot with significant improvements in coding, multilingual, and English-only tasks, now rolled out in paid **ChatGPT**. Despite this, **Claude Opus** remains superior in creativity and intelligence. **Mistral AI** released powerful open-source models like **Mixtral-8x22B** and **Zephyr 141B** suited for fine-tuning. **LangChain** enhanced tool integration across models, and **Hugging Face** introduced Transformer.js for running transformers in browsers. Medical domain-focused **Medical mT5** was shared as an open-source multilingual text-to-text model. The community also highlighted research on LLMs as regressors and shared practical advice on OCR/PDF data modeling from **Vik Paruchuri**'s journey.

Canonical issue URL

As many would expect, the April GPT4T release retook the top spot on LMsys and it is now rolled out in paid ChatGPT and with a new lightweight reproducible evals repo. We've said before that OpenAI will have to prioritize rolling out new models in ChatGPT to reignite growth.

All in all, a quiet before the presumable storm of the coming Llama 3 launch. You could check out the Elicit essay/podcast or the Devin vs OpenDevin vs SWE-Agent livestream. However we give today's pride of place to Vik Paruchuri, who wrote about his journey from engineer to making great OCR/PDF data models in 1 year.

image.png

These fundamentals are likely much more valuable than keeping on top of day to day news and we like featuring quality advice like this where we can.


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence. Comment crawling works now but has lots to improve!

TO BE COMPLETED


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

GPT-4 and Claude Updates

Open-Source Models and Frameworks

Research and Techniques

Industry News and Opinions

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries

Some other noteworthy discussions include:


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord

Forge Ahead, Leave No A1111 Behind: Forge, a new fork of Automatic1111 boasting performance enhancements, is gathering acclaim. Enthusiasts can explore Forge without ditching A1111 and utilize ComfyUI models for a more efficient workflow.

Ella Falls Short in Anime Art: Experimentation with Ella for anime-style image generation ends in disappointment, failing to meet user expectations even with recommended checkpoints. Despite high hopes, the quality of anime images generated by Ella remains subpar and is considered unusable for the genre.

Stable Diffusion 3 Brings Hope and Doubt: The community is abuzz with a blend of anticipation and skepticism around Stable Diffusion 3 (SD3), particularly about its potential to overcome current model limitations like bokeh effects, color fidelity, and celebrity recognition.

Expanding Toolbox for Image Perfection: In discussion are several tools and extensions enhancing Stable Diffusion outputs, including BrushNet for outpainting and solutions improving depth-fm and geowizard for architecture, as well as a color correction extension.

Cascade Gains Notoriety for Fast Learning: Cascade stands out within the Stable Diffusion models for its swift learning capabilities and distinct characteristics, although it's noted for a steeper learning curve, affectionately deemed the "strange cousin of the SD family."


Cohere Discord

CORS Crashes Cohere Connections: Users encountered CORS policy errors preventing access to the Cohere dashboard, with issues arising from cross-origin JavaScript fetch requests from https://dashboard.cohere.com to https://production.api.os.cohere.ai.

Arguments Over Context Length: A passionate discussion unfolded regarding the effectiveness of extended context lengths in large language models (LLMs) versus Retrieval-Augmented Generation (RAG), debating computational costs and diminishing benefits of longer contexts.

Rerank 3's Pricing and Promotion: Rerank V3 has been announced with a pricing of $2 per 1k searches and an introductory promotional discount of 50%. For those seeking the prior version, Rerank V2 remains available at $1 per 1k searches.

Navigating Cohere's Fine-Tuning and Deployment: Questions arose about the possibilities of on-premise and platform-based fine-tuning of Cohere's LLMs, alongside deployment options on AWS Bedrock or similar on-premise scenarios.

Boosted Search with Rerank 3 Overview: Rerank 3 launches to enhance enterprise search, claiming a threefold increase in inference speed and support for over 100 languages with its extended 4k context. It integrates with Elastic's Inference API to improve enterprise search functionalities, with resources available such as a Cohere-Elastic integration guide and a practical notebook example.


Unsloth AI (Daniel Han) Discord

Ghost 7B Aces Multiple Languages: The new Ghost 7B model is generating buzz due to its prowess in reasoning and understanding of Vietnamese, and is eagerly anticipated by the AI community. It is highlighted as a more compact, multilinguistic alternative that could serve specialized knowledge needs.

Double Take on Fine-Tuning Challenges: Discussions surfaced regarding difficulties in fine-tuning NLP models, with a gap noted between promising training evaluations and disappointing practical inference performance. Particularly, a lack of accuracy in non-English NLP contexts has been a point of frustration among engineers.

Efficient Model Deployment Strategies Sought: Engineers are actively sharing strategies and resources to streamline the deployment of models like Mistral-7B post-training. Concerns over VRAM limits persist, prompting discourse on optimizing batch sizes and embedding contextual tokens to conserve memory.

Unsloth AI Champions Extended Context Windows: The Unsloth AI framework is commended for reducing memory usage by 30% and merely increasing time overhead by 1.9% while enabling fine-tuning with context windows as long as 228K as detailed on their blog. This represents a significant leap compared to the previous benchmarks, offering a new avenue for LLM development.

The Importance of Domain-Specific Data: There is a consensus on the need for more precise, domain-specific datasets, as generic data collection is insufficient for specialized models requiring detailed context. Best practices are still being debated, with many looking towards platforms like Hugging Face for advanced dataset solutions.


Nous Research AI Discord


CUDA MODE Discord


Perplexity AI Discord


LM Studio Discord

Quantization Quest Continues: The Mixtral-8x22B model is now quantized and available for download, yet it is not fine-tuned and may challenge systems that can't handle the 8x7b version. A model loading error can be resolved by upgrading to LM Studio 0.2.19 beta preview 3.

Navigating Through Large Model Dilemmas: Users shared experiences running large models on insufficient hardware, suggesting cloud solutions or hardware upgrades like the NVIDIA 4060ti 16GB. For those tackling time series data, a Temporal Fusion Transformer (TFT) was suggested as being well-suited for the task.

GPU vs. CPU: A Performance Puzzle: When running AI models, more system memory can help load larger LLMs, but full GPU inference with a card like the NVIDIA RTX A6000 is optimal for performance.

Emerging ROCm Enigma in Linux: Linux users curious about the amd-rocm-tech-preview support are left hanging, while those with compatible hardware like the 7800XT report coil whine during tasks. Meanwhile, building the gguf-split binary for Windows is a hurdle for testing on AMD hardware, requiring a look into GitHub discussions and pull requests for guidance.

BERT's Boundaries and Embedding Exploits: The Google BERT models are generally not directly usable with LM Studio without task-specific fine-tuning. For text embeddings utilizing LM Studio, larger parameter models like mxbai-large and GIST-large have been recommended over the standard BERT base model.

Please note that while this summary is comprehensive, specific channels may contain additional detailed discussions and links relevant to AI engineers.


Eleuther Discord

BERT's Bidirectional Brainache: Engineers raised the complexity of extending context windows for encoder models like BERT, referencing difficulty with bidirectional mechanisms and pointing to MosaicBERT which applies FlashAttention, with questions about its absence in popular libraries despite contributions.

Rethinking Transformers with Google's Mixture-of-Depths Model: Researchers are discussing Google's novel Mixture-of-Depths approach, which allocates computing differently in transformer-based models. Also catching attention is RULER's newly open-source yet initially empty repository here, aimed at revealing the real context size of long-context language models.

Scale the Data Mountain Wisely: A paper proposing that data curation is indispensable and cannot ignore computational constraints was shared. The discourse included a symbolic search for entropy-based methods in scaling laws and a reflection on foundational research principles.

Odd Behaviors in Large Language Models Puzzles Analysts: Members expressed intrigue over NeoX's embedding layer behavior, questioning if weight decay was omitted during training. They compared NeoX's output to other models and confirmed a distinct behavior, igniting curiosity about the technical specifics and implications.

Quantization Quest and Dataset Dilemmas: Community efforts include an attempt at 2-bit quantization to reduce VRAM usage for the Mixtral-8x22B model, while confusion arose around The Pile dataset's inconsistent sizing and the lack of extraction code for varied archive types.


OpenRouter (Alex Atallah) Discord


Modular (Mojo 🔥) Discord


LangChain AI Discord

Note: Links related to adult content promotions have been actively ignored in this summary as they are clearly not relevant to the technical and engineering discussions of the guild.


HuggingFace Discord

Tweet Alert: osanseviero Shares News: osanseviero tweeted, potentially hinting at new insights or updates; check it out here.

RAG Chatbot Employs Embedded Datasets: The RAG chatbot uses the not-lain/wikipedia-small-3000-embedded dataset to inform its responses, merging retrieval and generative AI for accurate information inferences.

RMBG1.4 Gains Popularity: The integration of RMBG1.4 with the transformers library has garnered significant interest, reflected in 230K downloads this month.

Marimo-Labs Innovates Model Interaction: Marimo-labs released a Python package allowing the creation of interactive playgrounds for Hugging Face models; a WASM-powered marimo application lets users query models with their tokens.

NLP Community Pursues Longer-Context Encoders: AI engineers discussed the pursuit of encoder-decoder models like BigBird and Longformer for handling longer text sequences around 10-15k tokens and shared strategies for training interruption and resumption with trainer.train()'s resume_from_checkpoint.

Vision and Diffusion Achievements: GPU process management is enhanced with nvitop, while developers tackle video restoration through augmentation and temporal considerations, referencing works like NAFNet, BSRGAN, Real-ESRGAN, and All-In-One-Deflicker. Meanwhile, insights into Google's multimodal search capabilities are sought for improved image and typo brand recognition, with interest in the underpinnings of AI-demos' identifying technology.


Latent Space Discord


LAION Discord

Draw Things Draws Criticism: Participants voiced their disappointment with Draw Things, pointing out its lack of a complete open source offering; the provided version omits crucial features including metal-flash-attention support.

Questionable Training Feats of TempestV0.1: Community members met the TempestV0.1 Initiative's claim of 3 million training steps with skepticism, questioning both that and the physical plausibility of its 6 million-image dataset occupying only 200GB.

Will LAION 5B Demo Reappear?: Regarding the Laion 5B web demo, there's uncertainty about its return, despite mentions of Christoph indicating a comeback with no given timeline or further information.

Alert on LAION Scams: Warnings circulated on scams such as cryptocurrency schemes misusing LAION's name, with recommendations to stay cautious and discussions about combating this with an announcement or automatic moderation enhancements.

Advancements in Diffusion and LRU Algorithms: The community is evaluating improved Least Recently Used (LRUs) algorithms on Long Range Arena benchmarks and discussing guidance-weight strategies to enhance diffusion models, with relevant research (research paper) and an active GitHub issue (GitHub issue) being applied to huggingface's diffusers.


LlamaIndex Discord


OpenInterpreter Discord

Trouble in Installation Town: Members reported problems installing Poetry and litellm—a successful fix for the former included running pip install poetry, whereas diagnosing litellm issues involved using interpreter --version and pip show litellm. Further troubleshooting pointed towards the necessity of Python installation and particular git commits for package restorations.

Patience, Grasshopper, for Future Tech Gadgets: Inquiries were made on the preorder and delivery of new devices, revealing that some tech gadgets are still in the prototyping phase with shipments expected in the summer months. The conversation highlighted typical delays faced by startups in manufacturing and encouraged patience from eager tech aficionados.

Transformers Redefined in JavaScript: The transformers.js GitHub repository, offering a JavaScript-based machine learning solution capable of running in the browser sans server, piqued the interest of AI engineers. Meanwhile, a cryptic mention of an AI model endpoint at https://api.aime.info popped up without additional detail or fanfare.

OpenAI Plays the Credits Game: OpenAI's shift to prepaid credits away from monthly billing, which includes a promotion for free credits with a deadline of April 24, 2024, sparked curiosity and a flurry of information exchanges among the members regarding the implications for various account types.

Events and Contributions Galore: Community event Novus invites were buzzing as engineers looked forward to networking without the fluff, while a successful session on using Open Interpreter as a library yielded a repository of Python templates for budding programmers.


OpenAccess AI Collective (axolotl) Discord

Discussing Strategies and Anticipations in AI Development:

Training Challenges and Model Modifications:

Ecosystem Expansion and Assistance:

Resourcefulness in Resource Constraints:

Curiosity Meets Data Acquisition:


OpenAI Discord

API Stumbles with AttributeErrors: An OpenAI API user encountered an AttributeError in the client.beta.messages.create method in Python, raising concerns about potential documentation being out of sync with library updates. The shared code snippet didn't yield a solution within the guild discussions.

Models in the Spotlight: Members shared varying experiences using AI models like Gemini 1.5 and Claude, touching on differences in context windows, memory, and code query handling. For C# development specifically in Unity, the gpt-4-turbo and Opus model were suggested for efficacy.

Efficiency Hurdles with GPT-4 Turbo: One member observed that the GPT-4-turbo model appeared less skilled at function calls, while another was unsure about accessing it; however, detailed examples or solutions were not provided.

Large Scale Text Editing with LLMs: Queries about editing large documents with GPT sparked a discussion on the potential need for third-party services to bypass the standard context window limitations.

Navigating the Prompt Engineering Galaxy: For those embarking on prompt engineering, Prompting Guide was recommended as a resource, while integrating Wolfram with GPT can be managed via Wolfram GPT link and the @mention feature within the platform.


DiscoResearch Discord

Big Win for Dense Models: The launch of Mistral-22B-V.01, a new 22B parameter dense model, marks a notable achievement as it transitions from being a compressed Mixture of Experts (MoE) to a dense form, establishing a precedent in the MoE to Dense model conversion arena.

Crosslingual Conundrums and Corpus Conversations: While engineers work on balancing English and German data in models like DiscoLM 70b, with plans for updated models, they cited the need for better German benchmarks. Occiglot-7B-DE-EN-Instruct showed promise, hinting that a mix of English and German training data could be efficacious.

Sifting Through SFT Strategies: The community shared insights on the potential benefits of integrating Supervised Fine-Tuning (SFT) data early in the pretraining phase, backed by research from StableLM and MiniCPM, to enhance model generalization and prevent overfitting.

Zephyr Soars with ORPO: Zephyr 141B-A35B, derived from Mixtral-8x22B and fine-tuned via a new algorithm named ORPO, was introduced and is available for exploration on the Hugging Face model hub.

MoE Merging Poses Challenges: The community's experiments with Mergekit to create custom MoE models through merging highlighted underwhelming performance, sparking an ongoing debate on the practicality of SFT on narrow domains versus conventional MoE models.


Interconnects (Nathan Lambert) Discord

Increment or Evolution?: Nathan Lambert sparked a debate regarding whether moving from Claude 2 to Claude 3 represents genuine progress or just an "INCREMENTAL" improvement, raising questions about the substance of AI version updates.

Building Better Models Brick by Brick: Members discussed the mixing of pretraining, Supervised Fine-Tuning (SFT), and RLHF, pointing out the respective techniques are often combined, although this practice is poorly documented. A member committed to providing insights on applying annealing techniques to this blend of methodologies.

Casual Congrats Turn Comical: A meme became an accidental expression of congratulations causing a moment of humor, while another conversation clarified that the server does not require acceptance for subscriptions.

Google's CodecLM Spotlight: The community examined Google's CodecLM, shared in a research paper, noting it as another take on the "learn-from-a-stronger-model" trend by using tailored synthetic data.

Intellectual Exchange on LLaMA: A link to "LLaMA: Open and Efficient Foundation Language Models" was posted, indicating an active discussion on the progress of open, efficient foundation language models with a publication date of February 27, 2023.


tinygrad (George Hotz) Discord


Skunkworks AI Discord


LLM Perf Enthusiasts AI Discord


Alignment Lab AI Discord

Cry for Code Help: A guild member has requested help with their code by seeking direct messages from knowledgeable peers.

Server Invites Scrutiny: Concerns were raised over the excessive sharing of Discord invites on the server, sparking discussions about their potential ban.

Vitals Check on Project OO2: A simple inquiry was made into the current status of the OO2 project, questioning its activity.


Datasette - LLM (@SimonW) Discord


Mozilla AI Discord


AI21 Labs (Jamba) Discord

Hunting for Jamba's Genesis: A community member expressed a desire to find the source code for Jamba but no URL or source location was provided.

Eager for Model Merging Mastery: A link to a GitHub repository, moe_merger, was shared that lays out a proposed methodology for model merging, although it's noted to be in the experimental phase.

Thumbs Up for Collaboration: Gratitude was shared by users for the resource on merging models, indicating a positive community response to the contribution.

Anticipation in the Air: There's a sense of anticipation among users for updates, likely regarding ongoing projects or discussions from previous messages.

Shared Wisdom on Standby: Users are sharing resources and expressing thanks, showcasing a collaborative environment where information and support are actively exchanged.


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ▷ #general-chat (846 messages🔥🔥🔥):

Links mentioned:


Cohere ▷ #general (522 messages🔥🔥🔥):

Links mentioned:


Cohere ▷ #announcements (1 messages):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (268 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (244 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (7 messages):

Link mentioned: ghost-x (Ghost X): no description found


Unsloth AI (Daniel Han) ▷ #suggestions (1 messages):

starsupernova: oh yes yes! i saw those tweets as well!


Nous Research AI ▷ #off-topic (15 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (8 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (369 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (25 messages🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (63 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #general (6 messages):

Links mentioned:


CUDA MODE ▷ #cuda (168 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #torch (16 messages🔥):

Links mentioned:


CUDA MODE ▷ #beginner (12 messages🔥):

Links mentioned:


CUDA MODE ▷ #pmpp-book (1 messages):


CUDA MODE ▷ #ring-attention (8 messages🔥):

Links mentioned:


CUDA MODE ▷ #off-topic (4 messages):

Links mentioned:


CUDA MODE ▷ #hqq (11 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton-viz (3 messages):


CUDA MODE ▷ #llmdotc (98 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #general (281 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (12 messages🔥):


Perplexity AI ▷ #pplx-api (6 messages):

Link mentioned: Feature Roadmap: no description found


LM Studio ▷ #💬-general (173 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (46 messages🔥):

Links mentioned:


LM Studio ▷ #📝-prompts-discussion-chat (2 messages):


LM Studio ▷ #🎛-hardware-discussion (23 messages🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (9 messages🔥):

Links mentioned:


LM Studio ▷ #amd-rocm-tech-preview (12 messages🔥):

Links mentioned:


Eleuther ▷ #general (96 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (93 messages🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (11 messages🔥):

Link mentioned: Tweet from Pratyush Maini (@pratyushmaini): 1/ 🥁Scaling Laws for Data Filtering 🥁 TLDR: Data Curation cannot be compute agnostic! In our #CVPR2024 paper, we develop the first scaling laws for heterogeneous & limited web data. w/@goyalsach...


Eleuther ▷ #interpretability-general (8 messages🔥):


Eleuther ▷ #lm-thunderdome (1 messages):

butanium: Someone in my lab was also wondering if those chat_template branches were usable or not


Eleuther ▷ #gpt-neox-dev (10 messages🔥):


OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):

.o.sarge.o.: Seems to be an issue when trying to purchase tokens. Here is an image


OpenRouter (Alex Atallah) ▷ #general (213 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #general (9 messages🔥):

Link mentioned: Backdrop Build: Together we build - Bring that crazy idea to life alongside hundreds of other amazing builders in just 4 weeks.


Modular (Mojo 🔥) ▷ #💬︱twitter (1 messages):

ModularBot: From Modular: https://twitter.com/Modular/status/1778482233957101869


Modular (Mojo 🔥) ▷ #✍︱blog (1 messages):

Link mentioned: Modular: Row-major vs. column-major matrices: a performance analysis in Mojo and NumPy: We are building a next-generation AI developer platform for the world. Check out our latest post: Row-major vs. column-major matrices: a performance analysis in Mojo and NumPy


Modular (Mojo 🔥) ▷ #🔥mojo (125 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (2 messages):

Link mentioned: mog/examples/readme/layout.mojo at main · thatstoasty/mog: Contribute to thatstoasty/mog development by creating an account on GitHub.


Modular (Mojo 🔥) ▷ #community-blogs-vids (1 messages):

Links mentioned:


LangChain AI ▷ #general (107 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #langserve (4 messages):

Link mentioned: Join the Teen Content 18+ 🍑🔞 Discord Server!: Check out the Teen Content 18+ 🍑🔞 community on Discord - hang out with 441 other members and enjoy free voice and text chat.


LangChain AI ▷ #langchain-templates (3 messages):

Link mentioned: Join the Teen Content 18+ 🍑🔞 Discord Server!: Check out the Teen Content 18+ 🍑🔞 community on Discord - hang out with 441 other members and enjoy free voice and text chat.


LangChain AI ▷ #share-your-work (8 messages🔥):

Links mentioned:


LangChain AI ▷ #tutorials (4 messages):

Link mentioned: Join the Teen Content 18+ 🍑🔞 Discord Server!: Check out the Teen Content 18+ 🍑🔞 community on Discord - hang out with 441 other members and enjoy free voice and text chat.


HuggingFace ▷ #announcements (9 messages🔥):

<ul>
  <li><strong>Osanseviero's Tweet Blast</strong>: A new tweet has been shared by osanseviero, exciting news or insights expected. Check out the tweet <a href="https://twitter.com/osanseviero/status/1778430866718421198">here</a>.</li>
  <li><strong>Highlighting the Highlights</strong>: Community Highlights #53 delivers diverse verified user content including a Portuguese introduction to Hugging Face, a fashion try-on space, and various intriguing GitHub repositories.</li>
  <li><strong>Embedded for Success</strong>: The RAG chatbot is powered by an embedded dataset via <a href="https://huggingface.co/datasets/not-lain/wikipedia-small-3000-embedded">not-lain/wikipedia-small-3000-embedded</a>, serving as a retrieval source for generating user-informed responses.</li>
  <li><strong>Retrieval and Generation Duo</strong>: Combining retrieval from an embedded dataset with generative AI, the RAG chatbot innovatively seeks to provide accurate information inferences.</li>
  <li><strong>Rocking Downloads with RMBG1.4</strong>: RMBG1.4, integrated with the transformers library, hits a new milestone with 230K downloads this month, indicating strong community interest and usage.</li>
</ul>

Links mentioned:


HuggingFace ▷ #general (64 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (1 messages):

Link mentioned: Contain Your Composure: On Podman-Compose, Code Cleanup, and Tiny Llamas: This video tutorial will walk you through the process of building microservices using Podman-Compose, Yet another Markdown language (YAML) files, Small Langu...


HuggingFace ▷ #cool-finds (6 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (8 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (1 messages):


HuggingFace ▷ #computer-vision (14 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (12 messages🔥):


HuggingFace ▷ #diffusion-discussions (5 messages):

Links mentioned:


Latent Space ▷ #ai-general-chat (86 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (3 messages):

Link mentioned: Supervise the Process of AI Research — with Jungwon Byun and Andreas Stuhlmüller of Elicit: Timestamps:00:00:00 Introductions00:07:45 How Johan and Andreas Joined Forces to Create Elicit00:10:26 Why Products are better than Research00:15:49 The Evol...


Latent Space ▷ #llm-paper-club-west (26 messages🔥):

Links mentioned:


LAION ▷ #general (93 messages🔥🔥):

Links mentioned:


LAION ▷ #announcements (1 messages):


LAION ▷ #research (19 messages🔥):

Links mentioned:


LAION ▷ #learning-ml (1 messages):


LlamaIndex ▷ #announcements (1 messages):


LlamaIndex ▷ #blog (4 messages):


LlamaIndex ▷ #general (101 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):


OpenInterpreter ▷ #general (80 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (24 messages🔥):


OpenInterpreter ▷ #ai-content (2 messages):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (54 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (11 messages🔥):

Link mentioned: Reddit - Dive into anything: no description found


OpenAccess AI Collective (axolotl) ▷ #datasets (3 messages):


OpenAccess AI Collective (axolotl) ▷ #community-showcase (3 messages):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (18 messages🔥):

Links mentioned:


OpenAI ▷ #ai-discussions (54 messages🔥):


OpenAI ▷ #gpt-4-discussions (8 messages🔥):


OpenAI ▷ #prompt-engineering (8 messages🔥):


OpenAI ▷ #api-discussions (8 messages🔥):


DiscoResearch ▷ #mixtral_implementation (13 messages🔥):

Links mentioned:


DiscoResearch ▷ #general (4 messages):

Link mentioned: open-llm-leaderboard/details_mistral-community__Mixtral-8x22B-v0.1 · Datasets at Hugging Face: no description found


DiscoResearch ▷ #discolm_german (22 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (12 messages🔥):

Link mentioned: Tweet from rohan anil (@arohan): Interesting! “Answer the following multiple choice question. The last line of your response should be of the following format: 'ANSWER: $LETTER' (without quotes) where LETTER is one of ABCD. ...


Interconnects (Nathan Lambert) ▷ #ml-questions (6 messages):


Interconnects (Nathan Lambert) ▷ #random (12 messages🔥):


Interconnects (Nathan Lambert) ▷ #reads (1 messages):


Interconnects (Nathan Lambert) ▷ #sp2024-history-of-open-alignment (1 messages):

Link mentioned: aligning open language models - a natolambert Collection: no description found


tinygrad (George Hotz) ▷ #general (18 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (6 messages):


Skunkworks AI ▷ #datasets (7 messages):

Links mentioned:


LLM Perf Enthusiasts AI ▷ #claude (3 messages):


LLM Perf Enthusiasts AI ▷ #openai (4 messages):


Alignment Lab AI ▷ #ai-and-ml-discussion (1 messages):

fredipy: <@748528982034612226>


Alignment Lab AI ▷ #general-chat (4 messages):


Alignment Lab AI ▷ #oo2 (1 messages):

aslawliet: Is the project still alive?


Datasette - LLM (@SimonW) ▷ #ai (4 messages):

Link mentioned: GitHub - stanford-oval/storm: An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.: An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations. - stanford-oval/storm


Datasette - LLM (@SimonW) ▷ #llm (1 messages):

Link mentioned: fix: macos zsh llm cmd hangs by nkkko · Pull Request #12 · simonw/llm-cmd: Fix for #11, tested on M1 MacOs (14.3.) in Terminal and Alacritty (zsh), now works fine.


Mozilla AI ▷ #announcements (1 messages):

<ul>
  <li><strong>Gradio UI for Figma Launches:</strong> Mozilla Innovations introduces <strong>Gradio UI for Figma</strong>, a library based on Hugging Face's Gradio, to facilitate rapid prototyping in the design phase. Access the toolkit on <a href="https://www.figma.com/@futureatmozilla">Figma here</a>.</li>
  <li><strong>Join the Gradio UI Discussion:</strong> A conversation thread about <strong>Gradio UI for Figma</strong> with Thomas Lodato from Mozilla’s Innovation Studio is available for those interested in discussing the tool further. Join in on Discord through <a href="https://discord.com/channels/1089876418936180786/1091372086477459557/1228056720132280461">this thread</a>.</li>
</ul>

Link mentioned: Figma (@futureatmozilla) | Figma: The latest files and plugins from Mozilla Innovation Projects (@futureatmozilla) — We're building products that focus on creating a more personal, private and open-source internet


Mozilla AI ▷ #llamafile (4 messages):

Link mentioned: Burn: no description found


AI21 Labs (Jamba) ▷ #jamba (4 messages):