Frozen AI News archive

Lilian Weng on Video Diffusion

**OpenAI** expands with a launch in **Japan**, introduces a **Batch API**, and partners with **Adobe** to bring the **Sora video model** to Premiere Pro. **Reka AI** releases the **Reka Core multimodal language model**. **WizardLM-2** is released showing impressive performance, and **Llama 3** news is anticipated soon. Geoffrey Hinton highlights AI models exhibiting **intuition, creativity, and analogy recognition** beyond humans. The **Devin AI model** notably contributes to its own codebase. **Opus** demonstrates the ability to recognize its own generated outputs. **Sam Altman** warns startups about being steamrolled by OpenAI if they don't adapt quickly. **Yann LeCun** discusses AGI timelines, emphasizing it is inevitable but not imminent or solely from LLMs. Lilian Weng's blog on **diffusion models for video generation** highlights **training-free adaptation** as a breakthrough technique.

Canonical issue URL

One thing we missed covering in the weekend rush is Lilian Weng's blog on Diffusion Models for Video Generation. While her work is rarely breaking news on any particular day, it is almost always the single most worthwhile resource on a given important AI topic, and we would say this even if she did not happen to work at OpenAI.

Anyone keen on Sora, the biggest AI launch of the year so far (now rumored to be coming to Adobe Premiere Pro), should read this. Unfortunately for most of us, the average diffusion paper requires 150+ IQ to read.

image.png

image.png

We are only half joking. As per Lilian's style, she takes us on a wild tour of all the SOTA videogen techniques of the past 2 years, humbling every other AI summarizooor on earth:

image.png

The surprise find of the day comes from her highlight of Training-free adaptation, which is exactly as wild as it sounds:

"Somehow surprisingly, it is possible to adapt a pre-trained text-to-image model to output videos without any training 🤯."

image.png

She unfortunately only spends 2 sentences discussing Sora, and she definitely knows more she can't say. Anyway, this is likely the most authoritative explanation to How SOTA AI Video Actually Works you or I are ever likely to get unless Bill Peebles takes to paper writing again.


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/Singularity. Comment crawling works now but has lots to improve!

AI Companies and Releases

New Model Releases and Advancements in AI Capabilities

Industry Trends, Predictions and Ethical Concerns

Technical Discussions and Humor


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

WizardLM-2 Release and Withdrawal

Reka Core Release

Open Source Model Developments

LLM Architecture Developments

Miscellaneous Discussions


AI Discord Recap

A summary of Summaries of Summaries

1. New Language Model Releases and Benchmarks

2. Open Source AI Tools and Community Contributions

3. AI Hardware and Deployment Advancements

4. AI Safety, Ethics, and Societal Impact Debates

5. Misc


PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord


Unsloth AI (Daniel Han) Discord

Benchmark Bonanza: Engineers shared positive feedback on a first benchmark's results, praising its performance. There was also a conversation around extracting tokenizer.chat_template for model template identification in leaderboards.

Progressive Techniques in LoRA Tuning: Community members exchanged tips on LoRA fine-tuning, suggesting that the alpha parameter to tweak could be double the rank. They discussed ORPO's resource-optimization in model training and discouraged the use of native CUDA, advocating for Triton instead for learning and development benefits.

Data Hygiene Takes Center Stage: The ShareGPT90k dataset was presented in a cleaned and ChatML format to facilitate training with Unsloth AI, and users highlighted the key role of data quality in model training, alluding to a community preference for hands-on experimentation in learning model training approaches.

Collaboration and Contributions on the Rise: Open calls for contributions to Unsloth documentation and projects such as Open Empathic were made, indicating a receptive attitude toward community involvement. A member announced the development of an "emotional" LLM and collaboration with a Chroma contributor on libSQL and WASM integration.

Navigating Unsloth's Notebook Nuggets: Assistance with formatting personal messages for AI training was given, complete with a Python script link and a guide to use the ShareGPT format. Advice on packing and configurations for Gemma models were discussed to mitigate unexpected training issues.


Modular (Mojo šŸ”„) Discord

Bold Python Package Sets to Conquer Mojo Code: The creation of mojo2py, a Python package to convert Mojo language code into Python, indicates a trend toward developing tools for Python and Mojo interoperability.

Grammar Police Tackle Code Aesthetics: Engaging discussions highlighted the importance of indenting code, considered laughable yet significant for readability, and there was a sense of light-hearted camaraderie over code formatting conventions.

Accolades for Achieving Level 9 in Modular: A community member was congratulated for reaching level 9, indicating a point system or achievement metric within the Modular community.

Modular Tweets Tease the Tech-Savvy: A series of mysterious tweets from Modular sparked speculation and interest among the community, serving as an intriguing marketing puzzle.

Nightly Updates Kindle Community Interest: A fresh Mojo nightly update was announced, directing engineers to update their version to nightly/mojo and review the latest changes and enhancements detailed on GitHub's diff and the changelog.


Perplexity AI Discord

Billing Confusion and API Misalignments: Users express dissatisfaction with unexpected charges and discrepancies between Perplexity AI and API usage, pointing to instances where promo codes don't appear and seeking an understanding of parameters such as temperature for consistent results between different platforms.

Pro Feature Puzzlement: Changes to the Pro message counter in Perplexity AI led to mixed reactions, with some users enjoying "reduced stress" but others questioning the rationale behind such feature tweaks.

Model Performance Scrutiny: A divergence in opinion emerges on AI coding competencies, with GPT-4 seen as inadequate by some users, while others ponder the delicate trade-offs between various Perplexity models' abilities and performance.

Cultural Curiosity and Tech Talk: The community engages in a range of searches, from probing Microsoft's ad-testing endeavors to celebrating global cultural days, reflecting an eclectic mix of technical and creative interests.

API Result Inconsistencies Provoking Discussions: Queries in the community focus on aligning outcomes from Perplexity Pro and the API, with an undercurrent of worries about hallucinations and source credibility in the API's content.


LM Studio Discord

Windows Cleared for Model Takeoff: Responding to queries, members confirmed that the Windows executables for LM Studio are signed with an authenticode certificate and discussed the cost differences between Windows certificates and Apple developer licenses, with the former requiring a hardware security module (HSM).

The Trouble with VRAM Detection: Users reported errors related to AMD hardware on Intel-based systems in Linux, despite attempts to solve the issue with ocl-icd-opencl-dev. It led to a broader discussion about hardware misidentification and the challenges it poses in configurations.

WizardLM-2 Sharpens Its Conversational Sword: The WizardLM 2 7B model was praised for its ability in multi-turn conversations and its training methods, with its availability announced on Hugging Face. The WaveCoder ultra 6.7b was also recognized for its coding prowess following fine-tuning on Microsoft's CodeOcean.

Model Showdown: Users shared performance experiences with models like WizardLM-2-8x22B and Command R Plus, voicing mixed reactions. They exchanged views on what defines a "Base" AI model and the nuances of model fine-tuning and continuous learning, sparking debates over AI memory and bias.

Diverse Coding Prowess Under the Microscope: Within the guild, members delved into Python coding model capabilities, like Deepseek Coder and Aixcoder, urging others to check 'human eval' scores. Skepticism was expressed over claims about WaveCoder Ultra's superiority, with some implying exaggerated results, while discussions on model fine-tuning and quantization illuminated varying preferences for coding models and AI agent creation tools.


Nous Research AI Discord


OpenRouter (Alex Atallah) Discord


CUDA MODE Discord

PyTorch Book Still Flares Interest: Despite being 4 years old, "Deep Learning with PyTorch" is seen as a useful foundation for PyTorch fundamentals, while chapters on transformers, LLMs, and deployment are dated. Anticipation grows for a new edition to cover recent advancements.

Torch and CUDA Grapple with Optimization: Understanding and implementing custom backward operations in Llama exhibit challenges for AI engineers, while the use of torch.nn.functional.linear and the stable-fast library are leading discussions for optimizing inference in the CUDA environment.

Novel Approaches in Transcript Processing: An automated transcript for a CUDA talk utilizing cutting-edge tools is provided by Augmend Replay, offering the AI community OCR and segmentation features for video content analysis.

Quantum Leaps with HQQ and GPT-Fast: Significant strides in token generation speeds are observed after implementing torchao int4 kernel in the generation pipeline for transformers, rising to 152 tokens/sec. The HQQ+ method also marked an accuracy increase, spurring discussions around quantization axis and integration with other frameworks.

llm.C at the Forefront of CUDA Exploration: The llm.C project ignites discussions on CUDA optimizations, underscoring the balance between education and creating efficient kernels. Optimizations, profiling, potential strategies, and applicable datasets all jostle for attention in this growing space.


Eleuther Discord


OpenAI Discord


LlamaIndex Discord

Tutorial Treasure Trove: LlamaIndex announced an introductory tutorial series for agent interfaces and applications, aiming to clarify usage of core agent interfaces. In collaboration, LlamaIndex and Qdrant Engine introduced a hybrid cloud service offering, and a new tutorial was shared highlighting the integration of LlamaIndex with Azure AI to leverage hybrid search in RAG applications, crafted by Khye Wei from Microsoft found here.

AI Chat Chops: Within the LlamaIndex community, discussion ranged from implementing async compatibility with Claude in Bedrock (where async has not yet been implemented) to complex query construction help available in the documentation. Integration issues with gpt-3.5-turbo and LlamaIndex were likely related to outdated versions or account balances, and configuring fallbacks for decision-making with incomplete data remains an open challenge.

Reasoning Chains Revolution: Revealing advancements in reasoning chain integration with LlamaIndex, a key article titled "Unlocking Efficient Reasoning" can be found here. Solutions for token counting in RAGStringQueryEngine and hierarchical document organization in LlamaIndex were discussed in detail, with the community providing a concrete token counter integration guide involving a TokenCountingHandler and CallbackManager as per LlamaIndex's reference documentation.


LAION Discord

Hugging Face Rings in New TTS Library: A high-quality TTS model library, parler-tts, for both inference and training was showcased, bolstered by its hosting on Hugging Face's community-driven platform.

Scaling Down CLIP – Less Data, Equal Power: A study on CLIP demonstrates that strategic data use and augmentation can allow smaller datasets to match the performance of the full model, introducing new considerations for data-efficient model training.

Deepfakes – Legislation Incoming, Controversies Continue: The community debated newly proposed laws against deepfakes as well as unethical practices in AI, raising awareness about a potential scam promoted through a suspicious site advertised on Facebook, found here.

Safety Benchmarking Becomes ALERT: Discussion on the importance of safety in AI highlighted the release of the ALERT benchmark, designed to evaluate large language models for handling potentially harmful content and reinforcing conversations around safety versus creative freedom.

Audio Generation Advancements on the Horizon: Research involving the Tango model to enhance text-to-audio generation shed light on improvements in relevance and order of audio events, marking progress for audio generation from text in data-scarce setups.


HuggingFace Discord


Cohere Discord

Command-R Struggles with Macedonian: Discussions flagged that Command-R doesn't perform well in Macedonian, with concerns raised on the community-support channel. Issues raised highlight the need for multilingual model improvements.

Asynchronous Streaming with Command-R: Engineers queried the best practices for converting synchronous code to asynchronous in Python, aiming to enhance the efficiency of chat streaming with the Command-R model.

Trial API Limits Clarified: For Cohere's API, engineers discovered that the ā€˜generate, summarize’ endpoint has a limit of 5 calls per minute, while other endpoints permit 100 calls per minute, with a shared pool of 5000 calls per month for all trial keys.

Commander R+ Gains Traction: A discussion took root around accessing Commander R+ using Cohere’s paid Production API, highlighting existing documentation for potential subscribers.

Rubiks.ai Introduces AI Powerhouse: Engineers took note of the launch of Rubiks.ai, which offers a suite of models including Claude 3 Opus, GPT-4 Turbo, Mistral Large, and Mixtral-8x22B, with an introductory offer of 2 months of premium access on Groq servers.


OpenAccess AI Collective (axolotl) Discord

Deepspeed's Multi-node Milestone: A guide for multi-node distributed fine-tuning using Axolotl with Deepspeed 01 and 02 configurations was shared. The pull request outlines steps to address configuration issues.

Idefics2 Raises the Bar: The newly released Idefics2 8B on Hugging Face surpasses Idefics1 in OCR, document understanding, and visual reasoning with fewer parameters. Access the model on Hugging Face.

Pacing for RTX 5090's Big Reveal: Anticipation builds for Nvidia's upcoming RTX 5090 graphics card, speculated to debut at the Computex trade show. This early release may be fueled by competitive pressure as discussed on PCGamesN.

Gradient Accumulation Spotlighted: Queries on gradient accumulation's memory conservation in the context of sample packing and dataset length led to explorations of its impact on training time.

Streamline Model Saving with Axolotl: Configuring Axolotl to save models only upon training completion rather than after each epoch involves setting save_strategy to "no". Additionally, "TinyLlama-1.1B-Chat-v1.0" was recommended for tight computational spaces, with its setup in the examples/tiny-llama directory of Axolotl's repository.


Latent Space Discord

Rewound Now Unbound as Limitless: The wearable tech previously referred to as Rewind has been rebranded to Limitless, sparking a discussion about its real-time application potential and the implications for future AI advancements. Concerns regarding data privacy and HIPAA compliance for cloud-stored information were vocalized by members.

The Birth of Reka Core: Reka Core enters the chat as a multimodal language model that comprehends video. The community appears intrigued by the small team achievement in AI democratization and the technical report released at publications.reka.ai.

Cohere Compass Beta Steers In: Cohere's Compass Beta was unveiled as a next-level data search system, meriting discussion around its embedding model and the beta testing opportunities for applicants eager to explore its functional boundaries.

Payman AI Explores AI-Human Marketplaces: Payman AI piqued interest with its innovative concept of a marketplace where AI can hire humans, driving conversations around implications for data generation and advancing AI training methodologies.

Strong Compute Serves Resources on Silver Platter: Strong Compute revealed a grant program for AI researchers, dangling the carrot of $10k-$100k and substantial GPU resources for initiatives in Explainable AI, post-transformer models, and other groundbreaking areas, with a swift application deadline signaled by the end of April. Details on the offer and the application process were outlined at Strong Compute research grants page.


OpenInterpreter Discord

AI Innovation Storm Brewing: The OpenInterpreter community launched a brainstorming space to ideate on uses of the platform, focusing on features, bugs, and innovative applications.

Voice Communication Soars with Airchat: There’s a buzz around Airchat within the community as engineers exchange usernames and scrutinize its features and usability, signaling a growing interest in diverse communication platforms.

Open Source AI Generates Excitement: Opensource AI models, notably WizardLm2, are receiving attention for providing transparent access to powerful AI capabilities akin to GPT-4, highlighting community interest in open-source alternatives.

Navigating the 01 Pre-order Process: For those reconsidering their 01 pre-orders, they can easily cancel by reaching out to [email protected], and there’s growing discussion on Windows 11 installation woes and hardware compatibility improvisations using parts from AliExpress.

Linux Love for OpenInterpreter: Linux users are directed to rbrisita's GitHub branch, agglomerating all the latest PRs for the 01 device, and the community is also optimizing their 01 setups with custom designs and battery life improvements.


LangChain AI Discord


tinygrad (George Hotz) Discord


Interconnects (Nathan Lambert) Discord

AI Models Flood the Market: EleutherAI has introduced the Pile-T5 with details shared in a blog post, while WizardLM 2 is drawing interest with its foundation transformer tech and guide on WizardLM's page. Additionally, Reka Core breaks onto the scene as explained in its technical report, and Idefics2's debut is narrated on the Hugging Face blog, amid Dolma going open-source under an ODC-BY license.

Graph Love and Hefty Models Emit Buzz: The community is showing keen interest in turning sophisticated graphs into a Python library for model exploration, while expressing mixed reactions to LLAMA 3's massive training scale of 30 trillion tokens.

WizardLM Vanishes with Abrupt Apology: Tension rose with the unexplained removal of WizardLM, with its model weights and posts erased, prompting speculation and an apology from WizardLM AI over a missed toxicity test, and a potential re-release in the pipeline.

Exploration vs. Intervention: A member considers whether to leave a bot to its own learning process or to step in, illustrating the fine line between letting algorithms explore and manual intervention.


Datasette - LLM (@SimonW) Discord


Alignment Lab AI Discord


DiscoResearch Discord

LLama-Tokenizer Training Troubles: Engineering members shared challenges in training a Llama-tokenizer with the goal of achieving hardware compatibility via reduced embedding and output perceptron sizes. They explored scripts like convert_slow_tokenizer.py from Hugging Face and convert.py from llama.cpp to aid in the process.

Hunt for EU Copyright-Compliant Resources: There's an active quest to find text and multimodal datasets compatible with EU copyright laws for training a multimodal model. Suggestions for starting points included Wikipedia, Wikicommons, and CC Search to gather permissive or free data.

Sampling Strategies Examined: Discourse in the engineering circles revolved around decoding strategies for language models, emphasizing the need for academic papers to include modern methods like MinP/DynaTemp/Quadratic Sampling. A shared Reddit post offers a layman's comparison, while the conversation called for more rigorous research into these strategies.

Decoding Methodology Deserves a Closer Look: An examination of decoding methods in LLMs has exposed a gap in current literature, specifically related to open-ended tasks seen in operational models. Members expressed the need for in-depth research on advanced sampling methods and their impacts on model performance.

Creative Writing Boost with MinP Sampling: A notable performance boost in creative writing tasks was highlighted, with the alpaca-eval style elo score increasing by +8 and the eq-bench creative writing test seeing a +10 increment due to min_p sampling parameters. Such improvements signify the potential impacts of fine-tuning sampling strategies on LLM outputs.


Skunkworks AI Discord


Mozilla AI Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Stability.ai (Stable Diffusion) ā–· #general-chat (1015 messagesšŸ”„šŸ”„šŸ”„):

Links mentioned:


Unsloth AI (Daniel Han) ā–· #general (430 messagesšŸ”„šŸ”„šŸ”„):

Links mentioned:


Unsloth AI (Daniel Han) ā–· #random (6 messages):

Link mentioned: GitHub - l4b4r4b4b4/go-chroma: Go port of Chroma vector storage: Go port of Chroma vector storage. Contribute to l4b4r4b4b4/go-chroma development by creating an account on GitHub.


Unsloth AI (Daniel Han) ā–· #help (322 messagesšŸ”„šŸ”„):

Links mentioned:


Unsloth AI (Daniel Han) ā–· #showcase (47 messagesšŸ”„):

Links mentioned:


Modular (Mojo šŸ”„) ā–· #general (60 messagesšŸ”„šŸ”„):

Links mentioned:


Modular (Mojo šŸ”„) ā–· #šŸ’¬ļø±twitter (5 messages):


Modular (Mojo šŸ”„) ā–· #ai (2 messages):

Link mentioned: Unlock AI Agent real power?! Long term memory & Self improving: How to build Long term memory & Self improving ability into your AI Agent?Use AI Slide deck builder Gamma for free: https://gamma.app/?utm_source=youtube&utm...


Modular (Mojo šŸ”„) ā–· #šŸ”„mojo (541 messagesšŸ”„šŸ”„šŸ”„):

Links mentioned:


Modular (Mojo šŸ”„) ā–· #community-projects (4 messages):

Links mentioned:


Modular (Mojo šŸ”„) ā–· #nightly (12 messagesšŸ”„):

Links mentioned:


Perplexity AI ā–· #general (549 messagesšŸ”„šŸ”„šŸ”„):

Links mentioned:


Perplexity AI ā–· #sharing (12 messagesšŸ”„):


Perplexity AI ā–· #pplx-api (3 messages):


LM Studio ā–· #šŸ’¬-general (210 messagesšŸ”„šŸ”„):

Links mentioned:


LM Studio ā–· #šŸ¤–-models-discussion-chat (108 messagesšŸ”„šŸ”„):

Links mentioned:


LM Studio ā–· #šŸ“-prompts-discussion-chat (15 messagesšŸ”„):


LM Studio ā–· #šŸŽ›-hardware-discussion (21 messagesšŸ”„):

Links mentioned:


LM Studio ā–· #🧪-beta-releases-chat (26 messagesšŸ”„):


LM Studio ā–· #amd-rocm-tech-preview (8 messagesšŸ”„):


LM Studio ā–· #open-interpreter (1 messages):

rouw3n: Anyone here got the 01light software running on windows without problems ?


LM Studio ā–· #model-announcements (2 messages):

Links mentioned:


Nous Research AI ā–· #off-topic (18 messagesšŸ”„):

Links mentioned:


Nous Research AI ā–· #interesting-links (4 messages):

Links mentioned:


Nous Research AI ā–· #general (208 messagesšŸ”„šŸ”„):

Links mentioned:


Nous Research AI ā–· #ask-about-llms (39 messagesšŸ”„):

Links mentioned:


Nous Research AI ā–· #rag-dataset (9 messagesšŸ”„):

Links mentioned:


Nous Research AI ā–· #world-sim (87 messagesšŸ”„šŸ”„):

Links mentioned:


OpenRouter (Alex Atallah) ā–· #announcements (15 messagesšŸ”„):

Links mentioned:


OpenRouter (Alex Atallah) ā–· #app-showcase (6 messages):

Link mentioned: Rubik's AI - AI research assistant & Search Engine: no description found


OpenRouter (Alex Atallah) ā–· #general (258 messagesšŸ”„šŸ”„):

Links mentioned:


CUDA MODE ā–· #general (8 messagesšŸ”„):


CUDA MODE ā–· #cuda (30 messagesšŸ”„):

Links mentioned:


CUDA MODE ā–· #torch (2 messages):


CUDA MODE ā–· #cool-links (2 messages):

Link mentioned: Advancing GPU Computing: Shared Memory and Synchronization in CUDA: no description found


CUDA MODE ā–· #beginner (2 messages):


CUDA MODE ā–· #pmpp-book (4 messages):


CUDA MODE ā–· #youtube-recordings (8 messagesšŸ”„):

Link mentioned: Lecture 14: Practitioners Guide to Triton: https://github.com/cuda-mode/lectures/tree/main/lecture%2014


CUDA MODE ā–· #torchao (2 messages):

Links mentioned:


CUDA MODE ā–· #triton-puzzles (4 messages):


CUDA MODE ā–· #hqq (35 messagesšŸ”„):

Links mentioned:


CUDA MODE ā–· #llmdotc (129 messagesšŸ”„šŸ”„):

Links mentioned:


CUDA MODE ā–· #recording-crew (9 messagesšŸ”„):


Eleuther ā–· #announcements (1 messages):

<ul>
  <li><strong>Introducing Pile-T5</strong>: EleutherAI has released <strong>Pile-T5</strong>, an enhanced T5 model family trained on the Pile with up to 2 trillion tokens, showing improved performance on SuperGLUE, code tasks, MMLU, and BigBench Hard. The models leverage the new LLAMA tokenizer and can be further finetuned for better results.</li>
  <li><strong>Intermediate Checkpoints Available</strong>: Intermediate checkpoints of Pile-T5 have been made available in both HF and original T5x versions, inviting the community to explore and build upon this advance in NLP models.</li>
  <li><strong>Comprehensive Resources for Pile-T5</strong>: Check out the <a href="https://blog.eleuther.ai/pile-t5/">detailed blog post</a> introducing Pile-T5 and the rationale behind its development, and access the code on <a href="https://github.com/EleutherAI/improved-t5">GitHub</a> to implement these improvements in your own projects.</li>
  <li><strong>Spreading the Word on Twitter</strong>: The release of Pile-T5 has also been announced on <a href="https://x.com/arankomatsuzaki/status/1779891910871490856">Twitter</a>, providing insights into the model's training process and highlighting its open-source availability.</li>
</ul>

Links mentioned:


Eleuther ā–· #general (61 messagesšŸ”„šŸ”„):

Links mentioned:


Eleuther ā–· #research (137 messagesšŸ”„šŸ”„):

Links mentioned:


Eleuther ā–· #lm-thunderdome (27 messagesšŸ”„):

Links mentioned:


Eleuther ā–· #gpt-neox-dev (7 messages):


OpenAI ā–· #ai-discussions (167 messagesšŸ”„šŸ”„):

Links mentioned:


OpenAI ā–· #gpt-4-discussions (7 messages):


OpenAI ā–· #prompt-engineering (3 messages):


OpenAI ā–· #api-discussions (3 messages):


LlamaIndex ā–· #blog (3 messages):


LlamaIndex ā–· #general (117 messagesšŸ”„šŸ”„):

Links mentioned:


LlamaIndex ā–· #ai-discussion (15 messagesšŸ”„):

Links mentioned:


LAION ā–· #general (108 messagesšŸ”„šŸ”„):

Links mentioned:


LAION ā–· #research (17 messagesšŸ”„):

Links mentioned:


HuggingFace ā–· #announcements (10 messagesšŸ”„):

Links mentioned:


HuggingFace ā–· #general (72 messagesšŸ”„šŸ”„):

Links mentioned:


HuggingFace ā–· #cool-finds (4 messages):

Links mentioned:


HuggingFace ā–· #i-made-this (15 messagesšŸ”„):

Links mentioned:


HuggingFace ā–· #reading-group (3 messages):

Link mentioned: LLM Reading Group (March 5, 19; April 2, 16, 30; May 14; 28): Come and meet some of the authors of some seminal papers in LLM/NLP research and hear them them talk about their work


HuggingFace ā–· #computer-vision (2 messages):


HuggingFace ā–· #NLP (8 messagesšŸ”„):

Link mentioned: Home: Leveraging BERT and a class-based TF-IDF to create easily interpretable topics.


HuggingFace ā–· #diffusion-discussions (8 messagesšŸ”„):

Link mentioned: error in using stable cascade with long prompt Ā· Issue #7672 Ā· huggingface/diffusers: Hi, When I use stable cascade model with long prompt, I get below error. Token indices sequence length is longer than the specified maximum sequence length for this model (165 > 77). Running this s...


Cohere ā–· #general (100 messagesšŸ”„šŸ”„):

Links mentioned:


Cohere ā–· #project-sharing (1 messages):

Link mentioned: Rubik's AI - AI research assistant & Search Engine: no description found


OpenAccess AI Collective (axolotl) ā–· #general (56 messagesšŸ”„šŸ”„):

Links mentioned:


OpenAccess AI Collective (axolotl) ā–· #axolotl-dev (5 messages):


OpenAccess AI Collective (axolotl) ā–· #general-help (11 messagesšŸ”„):


OpenAccess AI Collective (axolotl) ā–· #axolotl-help-bot (15 messagesšŸ”„):

Links mentioned:


OpenAccess AI Collective (axolotl) ā–· #axolotl-phorm-bot (4 messages):

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


Latent Space ā–· #ai-general-chat (89 messagesšŸ”„šŸ”„):

Links mentioned:


Latent Space ā–· #ai-announcements (1 messages):

Link mentioned: Research Grants: no description found


OpenInterpreter ā–· #general (51 messagesšŸ”„):

Links mentioned:


OpenInterpreter ā–· #O1 (25 messagesšŸ”„):

Links mentioned:


LangChain AI ā–· #announcements (1 messages):

Link mentioned: Introduction | šŸ¦œļøšŸ”— LangChain: LangChain is a framework for developing applications powered by large language models (LLMs).


LangChain AI ā–· #general (43 messagesšŸ”„):

Link mentioned: Safeguarding AI: Strategies and Solutions for LLM Protection | LLM Security: Explore the security challenges and solutions of LLMs in this comprehensive guide. We cover potential risks, control mechanisms, and the latest tools for safer LLM application


LangChain AI ā–· #langserve (1 messages):


LangChain AI ā–· #share-your-work (6 messages):

Links mentioned:


LangChain AI ā–· #tutorials (3 messages):

Link mentioned: Unlock AI Agent real power?! Long term memory & Self improving: How to build Long term memory & Self improving ability into your AI Agent?Use AI Slide deck builder Gamma for free: https://gamma.app/?utm_source=youtube&utm...


tinygrad (George Hotz) ā–· #general (32 messagesšŸ”„):

Link mentioned: hotfix: bump line count to 7500 for NV backend Ā· tinygrad/tinygrad@e14a9bc: You like pytorch? You like micrograd? You love tinygrad! ā¤ļø - hotfix: bump line count to 7500 for NV backend Ā· tinygrad/tinygrad@e14a9bc


tinygrad (George Hotz) ā–· #learn-tinygrad (7 messages):


Interconnects (Nathan Lambert) ā–· #news (4 messages):


Interconnects (Nathan Lambert) ā–· #random (29 messagesšŸ”„):

Links mentioned:


Interconnects (Nathan Lambert) ā–· #sp2024-history-of-open-alignment (1 messages):

natolambert: should I wizardLM 2 as a troll lol


Interconnects (Nathan Lambert) ā–· #posts (2 messages):


Datasette - LLM (@SimonW) ā–· #ai (8 messagesšŸ”„):


Datasette - LLM (@SimonW) ā–· #llm (22 messagesšŸ”„):

Link mentioned: Initial LLM WebUI: Initial LLM WebUI. GitHub Gist: instantly share code, notes, and snippets.


Alignment Lab AI ā–· #oo (7 messages):

Link mentioned: WizardLM - a microsoft Collection: no description found


DiscoResearch ā–· #general (6 messages):

Links mentioned:


DiscoResearch ā–· #benchmark_dev (1 messages):

Link mentioned: Reddit - Dive into anything: no description found


Skunkworks AI ā–· #off-topic (4 messages):

Links mentioned:


Mozilla AI ā–· #llamafile (2 messages):

Link mentioned: Publish container to Docker Hub by dzlab Ā· Pull Request #59 Ā· Mozilla-Ocho/llamafile: Build and Publish container to Docker Hub on release using Github Actions #29 For this to work, need to setup the repository secrets: DOCKER_HUB_USERNAME DOCKER_HUB_ACCESS_TOKEN