Frozen AI News archive

Not much happened today

**Anthropic** released a team plan and iOS app about 4 months after **OpenAI**. The **Command-R 35B** model excels at creative writing, outperforming larger models like **Goliath-120** and **Miqu-120**. The **Llama-3 8B** model now supports a 1 million token context window, improving long-context understanding with minimal training on a single 8xA800 GPU machine. **TensorRT-LLM** benchmarks show it is 30-70% faster than **llama.cpp** on consumer hardware. A benchmark suggests **GPT2-Chat** may have better reasoning than **GPT-4-Turbo**, though results are debated. Demos include a self-learning **Llama-3** voice agent running locally on Jetson Orin and a Self-Learning Large Action Model (LAM). **Amazon CodeWhisperer** was renamed to **Q Developer**, expanding its generative AI assistant capabilities. **Apple** plans an AI-enabled Safari browser with an on-device LLM in iOS 18 and macOS 15. Big Tech dominates AI lobbying in Washington, while major U.S. newspapers sued **OpenAI** and **Microsoft** for copyright infringement. **DeepMind's AlphaZero** became the greatest chess player in 9 hours, and their Naturalized Execution Tuning (NExT) method improves LLM code reasoning by 14-26%. **Stable Diffusion** is used for diverse image generation applications.

Canonical issue URL

Anthropic continues its pattern being (merely) 4 months behind OpenAI in releasing a team plan and iOS app in an otherwise relatively quiet day in AI. Perplexity is teasing a private Pages feature with a signup form you can access via Discord:

image.png


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

LLM Models and Frameworks

AI Agents and Robotics

AI Assistants

AI Ethics and Governance

AI Research

Stable Diffusion and Image Generation


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

Claude iOS App Launch and New Features by Anthropic

AI Experts Share Insights

Personal Experiences and Reflections

AI Research and Updates

Other Topics


AI Discord Recap

A summary of Summaries of Summaries

1. Large Language Model (LLM) Advancements and Benchmarks

2. Optimizations and Techniques for Efficient LLM Inference

3. Open-Source AI Tools, Libraries, and Frameworks

4. Multimodal and Retrieval-Augmented AI Capabilities


PART 1: High level Discord summaries

CUDA MODE Discord


Unsloth AI (Daniel Han) Discord

GPU Efficiency Sparks Interest: The A4000 16GB GPU is lauded for its efficiency in training, with its cost-effectiveness earning praise when compared to the A100. The B200 is touted for its potential, being forecasted to be 25x more efficient than the current H100 at a comparable cost.

Techniques in Question: Debate over employing LoRA versus QLoRA revealed that QLoRA may offer a 75% VRAM usage reduction at the possible expense of 1-2% in model accuracy. The 80-10-10 split for training data was suggested to ensure model robustness, while language model fine-tuning is progressing, evidencing its application in Turkish translation.

Innovations in Model Training: Users reported quantization issues with llama.cpp, leading to GitHub issues such as #3759 and #4180. Workflows for fine-tuning and training were a point of clarification, with strategies for checkpointing and inference providers like Jan and GPT4All being put forward, available at repositories like janhq/jan.

AI Development Roadmapping Proposed: Advocates for a straightforward AI project roadmap emphasized its importance, while the potential of smaller models for enhanced conversational skills is under exploration. Additionally, the concept of retrieval augmentation is gaining traction, with references to implementations such as FlagEmbedding's GitHub repository.

Size and Performance: A noteworthy mention was that the Phi3 Mini 4k outperforms the larger 128k version in open LLM rankings, prompting a reevaluation of the efficacy of model sizes. There's an inclination toward models like Phi3 Mini 4k for their efficiency over larger counterparts.


LM Studio Discord

For updates on technical advancements and fixes, the community is advised to check the respective GitHub repositories and release pages for the latest commits and build updates.


Nous Research AI Discord


Stability.ai (Stable Diffusion) Discord


OpenAI Discord


Perplexity AI Discord

Pages Feature Prepares for Beta Liftoff: Perplexity AI announces an upcoming feature named Pages designed for crafting shareable, in-depth explorations of topics; early access to the beta version is available for interested users.

API Citations The Missing Piece: Engineers express concerns about accessing citations through API requests when using Perplexity-online models, alongside discussions of discrepancy between Pro UI and API model results; the API documentation is clarified to be the go-to resource for model details.

Limitations and Glitches in Spotlight: Members discussed the 50 daily usage limit for Opus, the presence of glitches in Pro Search and referencing tools, and slow responses from AI models, with technical advice offered around possible email filtering from service providers for login issues.

Discovery Through Shared Content: Users actively shared insights and links on diverse topics, including Microsoft Research Asia, the Vimeo API, and Tesla's self-driving tech; plus, a shared newsletter provided a window into product development insights.

Claude 3 Policy and Model Utilization Clarified: Queries about the usage policy of Claude 3 led to discussions on whether Perplexity's or Anthropic's policies are applicable, while the usage of online models in the Pro UI was explained to be either finetuned or employing a search engine-style vector database for responses.


Eleuther Discord


HuggingFace Discord


LlamaIndex Discord


Modular (Mojo 🔥) Discord

Mojo Marches On: The Mojo developer community celebrated the first anniversary of Mojo’s launch, praising the addition of traits, references, and lifetimes which significantly enhanced the standard library. Concerning enhancements, it was suggested to improve Mojo by allowing negative numbers, and implementing a fallback for scalar processing, inspired by linkage to articles within the issues.

Performance Power-ups: Innovative optimization of string allocations and conversions in Mojo cut processing time from 18.5 to 12.5 seconds for 100M records, with the latest effort reducing it further to 3.8 seconds utilizing multi-core processing techniques. A call was made to form Team-Mojo for the One Billion Row Challenge, seeing it as an opportunity for showcase and community collaboration.

Syntax and Semantics Synergy: Discussions on syntax and semantics highlighted the importance of Mojo’s syntax alignment for users and how inout in Mojo bears similarity to pass-by-reference in C++, but with its nuances. Questions about the __source_location() function led to a conversation pondering the inclusion of function_name in its output and the replacement of these features in the nightly branch.

Exploring Concurrency Considerations: The conversation speculated on Mojo's concurrency model potential, theorizing it might mirror an actor model more than the golang-style, with a spotlight on avoiding heavy runtime inclusion. The Mojo compiler, with an LLVM backbone, has a dedicated YouTube video explaining its underpinnings.

Tweet Teasers Lead to Speculation: Modular spurred curiosity with a series of unspecified tweets, teasing intriguing developments without revealing the specifics, piquing interest for details beyond the announcements.


OpenAccess AI Collective (axolotl) Discord

Exploring Axolotl's Model Support: In a discussion within the #axolotl-phorm-bot channel, it was clarified that Axolotl supports GaLore but not phi-3 format. Community advice recommended checking out the Hugging Face documentation for details on enabling GaLore. Meanwhile, an untested PR was highlighted as a resource for those looking to add command-r model to Axolotl.

Strategies for Effective Chat-Tokenization: Members in #general channel debated the impact of the Beginning of Sentence (BOS) token in tokenizer behavior, and the importance of specifying it correctly in different scenarios. Also, a study on generalist foundation models prompted discussions on the effectiveness of complex prompting strategies and the challenges in rendering academic theory practical.

Best Practices for Fine-Tuning New Models: The #general-help channel was abuzz with communities engaged in fine-tuning processes, where certain criteria such as using smaller models like an 8b model for beginners were recommended. Practical tips for dataset conversion for ShareGPT loader, and inquiries regarding fsdp compatibility with lora were discussed.

Tutorial Collaboration Strikes a Chord: In the #community-showcase, a tutorial illustrating the combination of axolotl and dstack, an open-source container orchestrator, was shared and well-received, emphasizing ease-of-use and flexibility. Contributors are directed to GitHub for detailed usage.

Compute Resources for Collaboration: An offer in the #axolotl-dev channel extended compute resources to other members for the purpose of helping with triage and troubleshooting, which could be particularly useful for those involved in bug fixes and enhancements.


LAION Discord

AI Enters TOS Grey Zone: A discussion emerged around users employing AI products without consenting to terms of service, highlighting a gray area in user agreement enforcement and prompting debate on legal implications for both users and providers.

Leaderboard Integrity Challenged: There's a push for a more transparent AI model leaderboard, emphasizing the need for openness and verifiability, while members expressed skepticism over LMSYS's Chatbot Arena, raising concerns of lack of objectivity and opaque data practices. The notion of incorporating only open source models and filtering by open weights was put forth as a criterion for improved leaderboards.

Eager for Efficiency: Engineering conversations revolved around a multitude of optimization strategies, from considering GANs for superior model reconstruction to discussions about Natten's cuda implementation, and the development of projects like magvit2.

Breaking New Ground in AI and Medicine: The community took note of a published study on cardiac ultrasound utilizing OpenCLIP that was recently featured in Nature Medicine, despite some existing issues with the study.

Revolutionizing Networks and Fact-Checking: Enthusiasm was evident for the innovative Kolmogorov-Arnold Networks (KANs), poised to outdo MLPs in accuracy and interpretability (the paper on KAN), and the introduction of VisualFactChecker, a training-free pipeline designed to bolster visual content captioning fidelity (the paper on VFC).


Latent Space Discord

Decentralizing AI's Compute Power: Prime Intellect has plunged into the exploration of decentralized AI training methodologies, aiming to rival the expansive GPU clusters employed by larger corporations. Their platform is geared towards leveraging globally distributed compute resources, as detailed in their extensive blog post.

Starcoder Rises: Hugging Face has launched a new Large Language Model called StarCoder2-15B-Instruct-v0.1, focusing primarily on code generation. They've made the model and pipeline open-source, inviting the community to engage, as outlined on their announcement page.

Simulating AI Societies on Consumer Tech: An experimental setup involving 300 AI agents called AI Town is reported to operate seamlessly on a MacBook M1 Max. The intriguing tweet reveals the capabilities and potential of AI simulations on consumer-level hardware.

LLM Paper Club: Ring in the Discussion: The LLM Paper Club's upcoming event features a collaborative discussion with the StrongCompute team on the Ring Attention paper. Engineers interested in the latest research findings can join via this Zoom link.

Video Meet for the Tech-Elite: A Zoom meeting video call has been set up for a more visual interactive discussion, likely concerning ongoing work or a paper club event. The community members can join using the provided Zoom Meeting link.


OpenInterpreter Discord

Respect Is Tech's Best Friend: A community reminder underscored the imperative of respect and constructive interaction; as the group expands, it is vital that everyone feel welcomed and valued for a collaborative future.

Open Interpreter Becomes Browser-Savvy: The Open Interpreter tool was confirmed to possess capabilities for web browsing and data scraping tasks without the need for traditional browser control, fostering direct web interactions through the AI.

Hitting the Right Note with DIY Speaker Amp: To boost the audio output from speakers, one solution recommended was an external amplifier, highlighting one potential amplifier on Amazon, though real-world application awaits confirmation upon testing.

R1's AI Unboxing Sparks Integration Talks: An MKBHD YouTube review on the AI product, Rabbit R1, watch here, ignited discussions on its potential integration with OpenInterpreter, with engineers eager to push the envelope of interconnected AI systems.

Tunnel Vision for Successful OI Connection: Engineers traded know-how on establishing a stable connection with an OpenInterpreter server, including the method for setting up new domains with ngrok and modifying the tunnel.py file, aiming to iron out connection wrinkles—more details at ngrok domains page.


OpenRouter (Alex Atallah) Discord


AI Stack Devs (Yoko Li) Discord

Crisp Visuals Spark Interest: Hexagen World surprised members with high-quality diffusion model outputs, suggesting promising avenues for interactive AI game development.

Retro Games Reimagined with AI: The Guild discussed reviving retro games like Farmville using Generative AI, with WebSim as a potential platform for these nostalgic reboots.

Spy Games Meet Generative Towns: An intriguing concept for a 1950s-themed AI town with a communist spy character was proposed, generating interest in creating an immersive cat-and-mouse game within WebSim.

Join the AI-Animated Conversation: Those curious about AI-driven animation received an invitation to a specialized Discord group via a community link, offering room for collaborative discussions and projects in interactive AI.

Dev Discussions Highlight Compatibility Issues: AI devs tackled local setup processes, noting particular issues with Windows systems and the importance of using the correct Node version (nvm use 19). Some even considered switching to Linux, especially since games like Stellaris are supported, as evidenced by information found on WineHQ.


Cohere Discord

Command R Impresses: The Cohere community has expressed appreciation for the CommandR/R+ models, highlighting their polished performance which seemingly surpasses other large language models for an enterprise-level experience.

LLM Grammar Secrets Exposed: A discussion on LLMs (Large Language Models) and their ability to generate grammatically correct sentences revealed insights into word and sentence embeddings, and the significance of the self-attention mechanism, with a resource provided for in-depth understanding.

AI Legal Eagle Takes Flight: A webinar on constructing an AI legal assistant using Cohere's RAG saw the community engaged, with a link to the recording made available on YouTube.

Azure Meets OAuth: Instructions for setting up OAuth with connectors on Azure using the Cohere toolkit were clarified, highlighting the ability for azure integration while keeping data internal as detailed on their GitHub page.

Multilingual Mastery in the Making: The implementation and potential of multilingual support in Command-R is under active evaluation by the community, with particular attention to languages like Norwegian and the desire for enhanced benchmarks.


LangChain AI Discord

PDF Table Extraction Proves Tricky: Engineers shared challenges with table extraction from PDFs using unstructure library, noting subpar outcomes particularly with multi-page tables. No solution was provided, indicating an area ripe for development or an opportunity for tool recommendation.

LangChain and Llama 3 Join Forces: There was a conversation about integrating Llama 3 with LangChain, directing users to utilize Fireworks and corresponding API keys. Additionally, a mention about the re-inclusion of Google Drive libraries in a project was noted, highlighting the cyclical nature of tech dependencies.

Launch, Updates, and Spec Introductions: Noteworthy developments include the launch of QuickVid for summarizing YouTube content, the update of LangChain chatbot to 0.1.17, and the introduction of Word Loom as a potential standard for AI language management, feedback solicited at their GitHub Gist. Queries about the usefulness of a detailed performance report comparing various LLMs for content creation were also raised.

Knowledge Graph Aspirations and AI Sales Agents: Members shared insights into tools for converting documents into knowledge graphs and the development of AI-powered Sales Agents. For the former, layout parsers and Azure Doc AI were proposed, alongside exploring LangChain's documented graph construction methods. The latter involved SalesGPT logic and a call for partnerships.

RAG Innovations and Language-Focused Tutorials: Engineers discussed a variety of RAG applications, including an Advanced RAG assistant for the French-speaking community, local training of Llama3, and an Adaptive RAG technique that responds based on query complexity. Related instructional videos were shared: French RAG Assistant, Local Agentic RAG w/ llama3, and LangGraph + Adaptive Rag + LLama3 Python Project.


Mozilla AI Discord

Mozilla AI is Hiring, Wave at Lm-buddy: Mozilla AI is currently expanding its team, with opportunities posted on their official Discord channel, and has also released Lm-buddy, a new open-source tool aimed at improving model evaluation efficiency.

LLaMA3:8b on M1 MacBook Air Confirmed for Testing: After users encountered issues with LLaMA3:8b running on M1 MacBook Air, the response indicated that testing on M1 will become a priority once other support issues are resolved.

Whispering to Llamafile: Proposals were made to integrate whisper.cpp models into llamafile for enhanced inference, despite the challenges in adding microphone and speaker functionalities.

Performance Debate Clarified: An article by Justine Tunney suggesting np.matmul performs at 29 gflops was contested, leading to a clarification that this was specific to an Intel computer on Ubuntu and that actual performance may vary.

Simultaneous Llamafiles and Path Customization Explained: Discussions in the guild confirmed that running multiple llamafiles with different models is possible, with operating systems managing the resources. Users also learned that customization using the --server --path PUBLIC_PATH option is limited to replacing .html and .js files in the zip file.


tinygrad (George Hotz) Discord

Tinygrad Undergoes Tensor Transformations: The tinygrad project implemented major updates with a commit renaming Scalar to ConstType, contributing to standardization efforts in the codebase. Discussions spotlighted the potential to optimize constant handling in operations by introducing const support variables and the significance of const Variables for operations linked to symbolic dimensions.

Graph Visualization Interest Piques for Backward Passes: The conversation included curiosity about visualizing graph diagrams for backward operations with a focus on issue #3572. There are hints at using dot files and setting GRAPH=1 for visual aid in understanding these operations.

Symbolic Dimensions Step into the Spotlight: Georgehotz shared insights on working with symbolic shapes and introduced a pull request with a skipped test for symbolic arange. This indicates an ongoing effort to enhance tinygrad's capabilities with symbolic dimensions.

JIT Crafting and Mean Calculations: A dialogue on improving tinygrad's Just-In-Time (JIT) compilation with symbolic variables led to the suggestion that a robust test would involve calculating the mean of variable-length 2D tensors. Such enhancements could refine the efficiency and performance of the JIT compiler.

CUDA Challenges on Nvidia Xavier: Technical discussions touched upon challenges faced while running EfficientNet examples on Nvidia Xavier, emphasizing the need to ensure CUDA=1 for proper script execution. Members also deliberated on whether Rednode's representation in tinygrad could be complicating symbolic compiler logic.


Interconnects (Nathan Lambert) Discord


Alignment Lab AI Discord


Skunkworks AI Discord


DiscoResearch Discord

LLaMA beats GPT-4 in Language Showdown: Results from scandeval.com indicate that LLaMA 3 outperforms GPT-4 in the ScandEval benchmark for German natural language tasks, sparking discussions about new AI model capabilities.

Accelerated Local Loads Trump Sluggish Cloud: An engineer reported that a program loads in 3 seconds on a local machine, pointing towards issues other than storage affecting slower load times when running jobs elsewhere.

Qdora Expands LLaMA's Middleway: Exciting progress in Large Language Model (LLM) expansion has emerged with the mention of qdora, a solution fostering the growth of models like LLaMA; the process is outlined in an Answer.ai blog post.

Avoiding Forgetfulness in AI Training: The guild discussed methods to prevent catastrophic forgetting during post-pretraining, referencing an Arxiv paper on enhancing Transformer blocks that helps LLMs retain old skills while learning new ones.

Fusing AI Past and Present: Guild engagement highlighted the prospect of "Non-forgetful Learning" in LLMs, where expansion techniques are crucial for merging traditional AI skills with newer, more advanced capabilities.


Datasette - LLM (@SimonW) Discord


PART 2: Detailed by-Channel summaries and links

CUDA MODE ▷ #general (4 messages):

Link mentioned: CCCL - Google Drive: no description found


CUDA MODE ▷ #triton (13 messages🔥):

Link mentioned: Lecture 14: Practitioners Guide to Triton: https://github.com/cuda-mode/lectures/tree/main/lecture%2014


CUDA MODE ▷ #cuda (14 messages🔥):

Link mentioned: Strangely, Matrix Multiplications on GPUs Run Faster When Given "Predictable" Data! [short]: Great minds discuss flops per watt.


CUDA MODE ▷ #algorithms (5 messages):


CUDA MODE ▷ #triton-puzzles (2 messages):


CUDA MODE ▷ #llmdotc (809 messages🔥🔥🔥):

Links mentioned:


CUDA MODE ▷ #rocm (8 messages🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (572 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (6 messages):


Unsloth AI (Daniel Han) ▷ #help (254 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (18 messages🔥):

Links mentioned:


LM Studio ▷ #💬-general (204 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (123 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (35 messages🔥):

Links mentioned:


LM Studio ▷ #⚙-configs-discussion (9 messages🔥):


LM Studio ▷ #🎛-hardware-discussion (272 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (141 messages🔥🔥):

Links mentioned:


LM Studio ▷ #autogen (4 messages):


LM Studio ▷ #amd-rocm-tech-preview (40 messages🔥):

Links mentioned:


LM Studio ▷ #crew-ai (2 messages):


LM Studio ▷ #🛠-dev-chat (1 messages):

yagilb: https://x.com/lmstudioai/status/1785796240656957514


Nous Research AI ▷ #ctx-length-research (25 messages🔥):

Links mentioned:


Nous Research AI ▷ #off-topic (25 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (9 messages🔥):

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

Links mentioned:


Nous Research AI ▷ #general (468 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (16 messages🔥):

Links mentioned:


Nous Research AI ▷ #rag-dataset (16 messages🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (24 messages🔥):

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (497 messages🔥🔥🔥):

Links mentioned:


OpenAI ▷ #annnouncements (1 messages):


OpenAI ▷ #ai-discussions (375 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (10 messages🔥):


OpenAI ▷ #prompt-engineering (30 messages🔥):


OpenAI ▷ #api-discussions (30 messages🔥):


Perplexity AI ▷ #announcements (1 messages):


Perplexity AI ▷ #general (241 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (19 messages🔥):

Link mentioned: How Perplexity builds product: Johnny Ho, co-founder and head of product, explains how he organizes his teams like slime mold, uses AI to build their AI company, and much more


Perplexity AI ▷ #pplx-api (14 messages🔥):

Link mentioned: Supported Models: no description found


Eleuther ▷ #general (28 messages🔥):

Links mentioned:


Eleuther ▷ #research (192 messages🔥🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (34 messages🔥):

Links mentioned:


HuggingFace ▷ #announcements (2 messages):

Links mentioned:


HuggingFace ▷ #general (151 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (3 messages):

Link mentioned: Med-Gemini: A High-Level Overview: A high-level overview on Med-Gemini, Google's "Family" (said in the voice of Vin Diesel) of Multimodal GenAI models for medicine. Med-Gemini has folks in the...


HuggingFace ▷ #cool-finds (8 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (11 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (18 messages🔥):

Links mentioned:


HuggingFace ▷ #computer-vision (17 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (5 messages):


HuggingFace ▷ #diffusion-discussions (1 messages):

sayakpaul: Might be a better question for A1111 forums.


HuggingFace ▷ #gradio-announcements (1 messages):

Link mentioned: Gradio Status: no description found


LlamaIndex ▷ #blog (4 messages):

Link mentioned: no title found: no description found


LlamaIndex ▷ #general (130 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (6 messages):

Link mentioned: Word Loom proposed update: Word Loom proposed update. GitHub Gist: instantly share code, notes, and snippets.


Modular (Mojo 🔥) ▷ #general (22 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (4 messages):


Modular (Mojo 🔥) ▷ #🔥mojo (58 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (1 messages):

Link mentioned: PMADDUBSW — Multiply and Add Packed Signed and Unsigned Bytes: no description found


Modular (Mojo 🔥) ▷ #community-blogs-vids (3 messages):

Link mentioned: Mojo Lang - Tomorrow's High Performance Python? (with Chris Lattner): Mojo is the latest language from the creator of Swift and LLVM. It’s an attempt to take some of the best techniques from CPU/GPU-level programming and packag...


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (7 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #nightly (20 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (23 messages🔥):

Link mentioned: Can Generalist Foundation Models Outcompete Special-Purpose Tuning? Case Study in Medicine: Generalist foundation models such as GPT-4 have displayed surprising capabilities in a wide variety of domains and tasks. Yet, there is a prevalent assumption that they cannot match specialist capabil...


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (2 messages):


OpenAccess AI Collective (axolotl) ▷ #general-help (14 messages🔥):

Link mentioned: Axolotl - Conversation: no description found


OpenAccess AI Collective (axolotl) ▷ #rlhf (1 messages):

gbourdin: add to my bookmarks. Thanks for this !


OpenAccess AI Collective (axolotl) ▷ #community-showcase (2 messages):

Link mentioned: dstack/examples/fine-tuning/axolotl/README.md at master · dstackai/dstack: An open-source container orchestration engine for running AI workloads in any cloud or data center. https://discord.gg/u8SmfwPpMd - dstackai/dstack


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (51 messages🔥):

Links mentioned:


LAION ▷ #general (60 messages🔥🔥):

Links mentioned:


LAION ▷ #research (25 messages🔥):

Links mentioned:


Latent Space ▷ #ai-general-chat (70 messages🔥🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

Link mentioned: LLM Paper Club (Ring Attention!) · Zoom · Luma: The StrongCompute gang (@adam_peaston, @fennecs) is covering Ring Attention today! https://arxiv.org/abs/2310.01889 Also submit and vote for our next paper:…


Latent Space ▷ #llm-paper-club-west (2 messages):

Link mentioned: Join our Cloud HD Video Meeting: Zoom is the leader in modern enterprise video communications, with an easy, reliable cloud platform for video and audio conferencing, chat, and webinars across mobile, desktop, and room systems. Zoom ...


OpenInterpreter ▷ #general (36 messages🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (31 messages🔥):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #announcements (2 messages):

<ul>
    <li><strong>Snowflake Arctic 480B and FireLLaVA 13B Models Launched</strong>: Announcing new models <strong>Snowflake Arctic 480B</strong>, excellent at coding with a hybrid transformer architecture, available at <a href="https://openrouter.ai/models/snowflake/snowflake-arctic-instruct">Snowflake Arctic 480B</a> and <strong>FireLLaVA 13B</strong>, an open source, multimodal model by Fireworks at <a href="https://openrouter.ai/models/fireworks/firellava-13b">FireLLaVA 13B</a>. Both come with new pricing and detailed specifications for developers.</li>
    <li><strong>Improved Load Balancing and Detailed Provider Stats</strong>: OpenRouter introduced <strong>load balancing</strong> to manage providers' load surges and now allows monitoring of latency and providers' finish reasons, enhancing performance for users, accessible on the <a href="https://openrouter.ai/activity">Activity page</a>.</li>
    <li><strong>Streamlined Docs for Developers</strong>: Documentation updates for Image and multimodal requests, plus tool calls and function calling, are now available to guide usage on <a href="https://openrouter.ai/docs#images-_-multimodal-requests">Image Requests</a> and <a href="https://openrouter.ai/docs#tool-calls">Tool Calls</a>.</li>
    <li><strong>Feature Expansion and Price Adjustments</strong>: Announced support for <strong>logit_bias</strong> and <strong>min_p</strong> on Lepton models, a significant 40% price cut on Mythomax Extended, and a slight 4% reduction for Mixtral 8x7b Instruct. These changes reflect OpenRouter's commitment to cost-effective and advanced AI capabilities.</li>
    <li><strong>Impending API Changes and Developer Notifications</strong>: Developers are alerted about the upcoming removal of the <code>total_cost</code> field from non-streaming completions and a potential requirement of the <code>User-Agent</code> header in requests to improve service security and efficiency.</li>
</ul>

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):

Link mentioned: Skribler | Skriv med AI: no description found


OpenRouter (Alex Atallah) ▷ #general (64 messages🔥🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #app-showcase (28 messages🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #ai-town-discuss (2 messages):


AI Stack Devs (Yoko Li) ▷ #ai-town-dev (33 messages🔥):

Link mentioned: WineHQ - Stellaris: no description found


Cohere ▷ #general (35 messages🔥):

Links mentioned:


Cohere ▷ #collab-opps (1 messages):

There are no sufficient details or discussion points provided in the single message history given to create a summary. If more chat contents were provided, a summary could be created following the guidelines.


LangChain AI ▷ #general (24 messages🔥):

Links mentioned:


LangChain AI ▷ #langserve (1 messages):


LangChain AI ▷ #share-your-work (7 messages):

Links mentioned:


LangChain AI ▷ #tutorials (3 messages):

Links mentioned:


Mozilla AI ▷ #announcements (1 messages):

<ul>
  <li><strong>Join the Mozilla AI Team</strong>: Mozilla AI is expanding its team and is currently hiring. Interested parties can check out the employment opportunities on their official Discord channel [here](https://discord.com/channels/1089876418936180786/1230938514955436242/1234870020916510823).</li>
  <li><strong>Introducing Lm-buddy</strong>: Mozilla AI has released a new open-source tool named **Lm-buddy** designed to help evaluate models more efficiently. For more details and access, visit the announcement in their channel [here](https://discord.com/channels/1089876418936180786/1230938514955436242/1234589599733518378).</li>
  <li><strong>Local LLM as Digital Jurist</strong>: There's a discussion about using a **Local LLM** as a judge via the Prometheus framework. Details are available on the Discord channel, accessible [here](https://discord.com/channels/1089876418936180786/1234890301143912599/1234890301143912599).</li>
</ul>

Mozilla AI ▷ #llamafile (34 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #general (8 messages🔥):

Link mentioned: tensor variable by geohot · Pull Request #4362 · tinygrad/tinygrad: no description found


tinygrad (George Hotz) ▷ #learn-tinygrad (13 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (11 messages🔥):


Interconnects (Nathan Lambert) ▷ #reads (1 messages):

Link mentioned: AI leaderboards are no longer useful. It's time to switch to Pareto curves.: What spending $2,000 can tell us about evaluating AI agents


Interconnects (Nathan Lambert) ▷ #posts (2 messages):


Alignment Lab AI ▷ #ai-and-ml-discussion (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #programming-help (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #looking-for-collabs (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #general-chat (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #landmark-dev (1 messages):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


Alignment Lab AI ▷ #landmark-evaluation (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #open-orca-community-chat (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #leaderboard (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #looking-for-workers (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #looking-for-work (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #join-in (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #fasteval-dev (1 messages):

No summary can be provided as the content does not contain relevant topics or discussion points related to AI or the Alignment Lab AI Discord chatbot messages. Further, the content appears to be inappropriate and not aligned with the expected academic or professional discussions typically summarized.

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Alignment Lab AI ▷ #qa (1 messages):

Link mentioned: Join the e-girl paradise 🍑🍒 // +18 Discord Server!: Check out the e-girl paradise 🍑🍒 // +18 community on Discord - hang out with 16457 other members and enjoy free voice and text chat.


Skunkworks AI ▷ #general (11 messages🔥):

Links mentioned:


Skunkworks AI ▷ #off-topic (2 messages):

Link mentioned: NEVER GIVE UP YOUR WAAAAAAAAAAAAY: NEVA GIVE UP - https://bit.ly/2VrgAcKSong is Before my Body is Dry instrumental version from the anime Kill La KillConsider donating to our Patreon!https://w...


DiscoResearch ▷ #general (1 messages):


DiscoResearch ▷ #benchmark_dev (1 messages):

le_mess: llama 3 seems to beat gpt4 on scandeval https://scandeval.com/german-nlg/


DiscoResearch ▷ #discolm_german (1 messages):

Link mentioned: LLaMA Pro: Progressive LLaMA with Block Expansion: Humans generally acquire new skills without compromising the old; however, the opposite holds for Large Language Models (LLMs), e.g., from LLaMA to CodeLLaMA. To this end, we propose a new post-pretra...


Datasette - LLM (@SimonW) ▷ #llm (2 messages):