Frozen AI News archive

Perplexity, the newest AI unicorn

**Perplexity** doubles its valuation shortly after its Series B with a Series B-1 funding round. Significant developments around **Llama 3** include context length extension to **16K tokens**, new multimodal **LLaVA models** outperforming Llama 2, and fine-tuning improvements like QDoRA surpassing QLoRA. The **Llama-3-70B** model is praised for instruction following and performance across quantization formats. **Phi-3 models** by **Meta AI** released in multiple sizes show competitive benchmark results, with the 14B model achieving **78% on MMLU** and the 3.8B model nearing **GPT-3.5** performance.

Canonical issue URL

Just 3 months after the Series B, Perplexity doubles its valuation again with a Series B-1, with mostly the same list of stellar investors as last time, but a rare split of Daniel Gross not co-leading with Nat Friedman. Dan seems to have a special relationship with the company - Aravind shared a Dec 2022 email on Dan's product feedback.

image.png


Table of Contents

[TOC]


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

Llama 3 Variants and Optimizations

Llama 3 Performance and Capabilities

Phi-3 and Llama 3 Models Push Boundaries of Open-Source Language AI

Miscellaneous


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

Perplexity AI Raises $62.7M at $1.04B Valuation

Meta's Llama-3 Model Achieves Top Performance

Microsoft Releases Phi-3 Language Models

Google's Gemini 1.5 Pro Achieves Strong Performance

Other Notable Releases and Benchmarks


AI Discord Recap

A summary of Summaries of Summaries

1. Evaluating and Comparing Large Language Models

2. Advancements in Retrieval-Augmented Generation (RAG)

3. Fine-tuning and Optimizing Large Language Models

4. Multimodal and Vision Model Developments

5. Misc

Let me know if you would like me to elaborate on any part of the summary or if you have additional questions!


PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord


Perplexity AI Discord

Perplexity AI Hits $1 Billion Valuation: After a successful funding round, Perplexity AI has been valued at a whopping $1 billion, even appearing in Bloomberg articles, with potential collaborations hinted involving AI expert Yann LeCun. The enterprise version, dubbed Perplexity Enterprise Pro, boasts enhanced data privacy and management features, drawing attention from major companies.

New Product Launch Brings Expectations and App Woes: The launch of Perplexity AI's Enterprise Pro for $40/month has stirred excitement and anticipation for possible upcoming features, although some frustration was voiced over technical difficulties with the iOS app on iPads. Despite the issues, the enthusiasm suggests high expectations from the current user base.

Data Privacy Takes Center Stage: In light of the Enterprise Pro introduction, users discussed data privacy concerns, prompting moderator references to official statements about user consent for data use in models. Separately, the sharing channel instructed users on compliances necessary to share Perplexity AI's search threads.

Anticipation Grows for Perplexity's High Valuation Fundraise: Community conversations buzzed about Perplexity AI seeking to raise $250 million at a $2.5 to $3 billion valuation, as members shared a TechCrunch article and a CNBC interview with CEO Aravind Srinivas, signifying rapid company growth and market interest.

API User Looks for Cutting-Edge Features: A request on the pplx-api channel highlighted a thirst for an API providing up-to-date web information, like GPT but with browsing capabilities; Perplexity's sonar online models were recommended, found in their documentation, with additional advice on prompt enhancement for improved model performance.


Stability.ai (Stable Diffusion) Discord


Nous Research AI Discord


LM Studio Discord


CUDA MODE Discord


Eleuther Discord

Local LLMs on Smartphone Horizon: Discussions explored the feasibility of running large language models (LLMs) on smartphones, considering memory bandwidth (up to 51.2 GB/s) and GPU capabilities (Exynos 2400 chipset specs), suggesting even 7-8B models might be workable. Community members examined existing apps like MLC-LLM and discussed how Hugging Face's downtime raises questions about free AI model hosting sustainability.

SpaceByte Makes Tokenization Obsolete: A new byte-level LLM architecture, SpaceByte, promises to eliminate the need for tokenization, addressing potential information leakage from tokenizers. Other discussions critiqued Fineweb's relation to LLaMA and the novel application of ProGen2 for AI-designed CRISPR-Cas proteins, showcasing LLMs' role in accelerating scientific discovery.

Scale Wisely with Tactful Debates: A clash over data rounding in a publication sparked wider conversation about constructive criticism and tone in technical debates. The skirmish illuminated misunderstandings around attributions of rounding data to the Chinchilla paper versus the replication team, unraveling deeper issues in replication methodologies.

RWKV Integration Ramps Up: GPT-NeoX developers are busy implementing RWKV (Rethinking Weighted Key-Value Memory Networks) with support for fp16 and JIT kernel compilation. Progress and tasks are detailed in GitHub Issue #1167, and developers are pushing for a version numbering system to streamline the iteration process.

AI Designs High-Performance Proteins: Profluent Bio successfully employed LLM ProGen2 to design new CRISPR-Cas protein sequences, yielding variants with increased specificity. The accomplishment demonstrates LLMs' expanding utility in biotechnology sectors.


HuggingFace Discord

Chatting with PDFs, Now with Math!: ai_pdf is an open-source project enabling conversations with PDF documents, excelling with math PDFs by converting them to LaTeX.

Voice Directed AI Artistry: A 2.5-minute video generated in real-time from voice commands has been shared on Reddit, pointing towards a future of AI-driven dynamic video creation.

AI Gets Reasonable: Transformers.js allows running HuggingFace Transformers directly in the browser, expanding the playfield for AI applications in web environments.

Rust Helps Minify BPE: minbpe-rs is a Rust port of minbpe with functions for tokenization and training, improving performance for NLP tasks. The project is available on GitHub.

Diffusion Dilemmas and AI Video Debates: Users discuss the feasibility of creating a 1-minute video on "AI Horse" using Diffusion, and others tackle various implementation challenges, demonstrating the teething issues of burgeoning AI applications.


Modular (Mojo 🔥) Discord

Code Instructions Boost Hermes: After integrating code instruction examples, Hermes 2.5 has been observed to outperform Hermes 2 in various benchmarks, with notable improvements in metrics such as the MMLU benchmark score.

Mistral's Capacity Challenge: Discussions concluded that Mistral cannot be scaled beyond 8k without ongoing pretraining. Focus shifts to enhancements in model merging strategies, such as applying differences between UltraChat and base Mistral to Mistral-Yarn.

Empathy in AI: The Open Empathic project seeks assistance in expanding categories; contributors are guided by a YouTube tutorial and encouraged to leverage movie scenes from YouTube for diversity in empathic response training.

Mojo Delights in Differences: Clarifications were made on Mojo around parameters and arguments with the latter being runtime values, while parameters in the language remain compile-time constants. Complex patterns like 'Type State' are being explored, and performance comparison to Python reveals ongoing efficiency issues, notably in IO operations.

In the Trenches with Mojo SIMD and Multithreading: Implementing SIMD patterns in Mojo yielded close performance to Rust in a CPU-limited context. However, optimization challenges exist, such as the best practices for parallelize. In other discussions, the use of UnsafePointer and the phasing out of LegacyPointer indicate a maturation of memory handling within the language.


OpenAccess AI Collective (axolotl) Discord


OpenRouter (Alex Atallah) Discord


OpenAI Discord


LAION Discord


LlamaIndex Discord

DREAM Big with Distributed RAG: LlamaIndex introduces DREAM, a Distributed RAG experimentation framework, while also launching various RAG enhancements like ColBERT with a Twist and LoRA Fine-Tuning. Dig into the discussions about CRAG, an innovative layer improving RAG retrieval, and open-source rerankers in LlamaIndex tweets.

Using AI Models Beyond OpenAI: Within #general, users tackle different retrieval methods for LLMs, while addressing integration bugs and API key annoyances. There's a spotlight on techniques for improved context management and interest in using alternatives to OpenAI's options, as detailed in numerous LlamaIndex docs.

From LinkedIn to Google Sheets, AI Funding Data Draws Interest: A member shares an Infini Attention explainer on LinkedIn, while AI funding distribution by city is accessible on Google Sheets. New LLM-Ready Markdown integrations excite the community, and WhyHow.AI's boosted Knowledge Graph SDK invites beta testers on Medium.

Database Debates and Fine-tuning: Members in #ai-discussion actively debate database types optimal for LLM training. They underscore the importance of understanding database schema and vector store possibilities when training large language models.


OpenInterpreter Discord

Caught a Case of the Compatibility Blues: Members noted that Open Interpreter, despite successful implementations, encountered challenges with Windows and mix-ups regarding model support, specifically clarifying that OI currently only supports OpenAI for the cloud option, not Groq or the Llama 3 70b model. They also discussed stability issues with the Llama 3 70b compared to its 8b counterpart.

Say What, Interpreter?: Various functionalities and integration challenges with Open Interpreter were highlighted, such as installation issues on Windows systems and pytesseract errors, the latter mitigated by using pip install --upgrade litellm. Detailed troubleshooting videos, e.g., on YouTube for integrating OI with GROQ API, show community eagerness for cost-effective solutions.

Screen Vision, but No Prophecy: In the AI vision domain, it was clarified that Open Interpreter leverages the GPT-4-vision-preview for screenshot recognition tasks, indicating a mix of text and vision capabilities within the tool.

Helping Hands and Config Stands: The community celebrated reaching 100 GitHub contributors for Open Interpreter and displayed a strong collaboration spirit. There’s a push for sharing default configuration files, as seen in a pull request, to improve interactions with various models.

M1 Mac Spacebar Conspiracy: Specifically, for M1 Mac users, troubleshooting a recording issue where pressing the spacebar didn't work as intended, diverse solutions were proposed, including installing ffmpeg, checking microphone permissions or switching Python versions using conda.

Cloudy with a Chance of Compatibility: There's a desire among members to see OI aligned with cloud services, with calls to enable compatibility for broader cloud platform support, including but not limited to platforms like brev.dev and Scaleway.


Interconnects (Nathan Lambert) Discord

Clickbait vs. Substance: The debate over AGI article titles in the community reflects a push for engaging yet truthful headlines. The discord in opinions, varying from AGI's ontological status to being a faith, indicates a search for thought-provoking yet honest discourse, as illustrated by titles like "AGI Isn't Real" and Mistral CEO Arthur Mensch's interview in Business Insider.

Phi-3 Under the Microscope: There is skepticism around the integrity of the Phi-3 benchmarks due to perceived overfitting on benchmarks like the MMLU, calling into question their relevance for OOD performance. Criticism also extends to the model's evaluation presentation and undisclosed data pipelines, amidst excitement for Phi-3's anticipated MIT license release and multilingual capabilities.

Benchmarking Evals: The utility of AI model evaluations is scrutinized, noting the trade-offs between automated benchmarking tools like MMLU, BIGBench, and human-intensive evaluations like ChatBotArena. Perplexity-based evaluations, like AI2's Paloma, were confirmed to be more for internal training checkpoints rather than public competitions.

Discord Community Dynamics: Anecdotes about the community include a researcher's ephemeral tweeting habits, the surprising low membership despite free subscription, and candid aspirations of engaging with industry figures like Ross Taylor post NDA-laden periods.

A Tangle of Instruction and CRINGE: The ecosystem of instruction tuning is expounded with references to an introductory blog and appreciation for the classification in the MT Bench paper. Additionally, the CRINGE paper's novel training approach using negative examples gains attention and is further discussed in relation to instruction tuning.


Cohere Discord


LangChain AI Discord

Webpage Wizardry with LLM Scraper: The newly unveiled LLM Scraper on GitHub presents a method to transform any webpage into structured data, leveraging LLM's parsing capabilities, and cacheing previous replies to subsequent requests.

Stock Analysis at Your Fingertips: AllMind AI, an AI tool that promises speedy and economical financial insights, is gunning for the top spot on Product Hunt.

Automated Graphs Get Smarter: WhyHow.AI has rolled out a major upgrade with schema-controlled automated knowledge graphs, aiming to structure user-uploaded content more efficiently. The new feature and its beta program were introduced on a Medium post.

Conversational Query Crafting: A blog post breaks down how the Self-querying retriever creates structured queries from natural language inputs, enhancing semantic similarity searches with filtering based on metadata.

Watermark Warnings for LLMs: The community delved into the concept of watermarking in AI-generated texts, a technique for planting identifiable patterns, as detailed on this resource page: Watermarking LLMs.


tinygrad (George Hotz) Discord

TinyGrad Tackles Segfaults and Training Woes: Discussions highlighted challenges with setting up tinygrad post-ROCm 6.1 release due to segfaults, while George Hotz assured that the master branch is stable thanks to robust CI.

AI Hardware Hyped to Outperform Cloud: The community debated the merits of decentralized AI services like TinyBox against traditional cloud services, focusing on points such as censor resistance, local training feasibility, and the importance of real-time user data training.

Inside TinyGrad's Mechanics: In the realm of tinygrad, members dove into deep discussions about stacking tensors, shape tracking, and memory management, exchanging tutorials and documentation that reveal the innards of the minimalist deep learning library.

Windows Walks a Tightrope with CUDA: Windows users shared their experiences and workarounds for running tinygrad with CUDA, using tools like WSL and Docker, while acknowledging the platform's official unsupported status for this setup.

George Hotz Chronicles Upcoming Tinygrad Evolutions: In a weekly roundup, Hotz mentioned focus areas for upcoming discussions, highlighting mlperf progress, potential NVIDIA CI strategies, and the goal of keeping the tinygrad codebase succinct.

ShapeTracker Tutorial, Uops Documentation, and CUDA Tensor Core Guide were shared as educational resources, while Meta AI was cited in the discussion.


DiscoResearch Discord

Mixtral Edges Out Llama3: Mixtral-8x7B-Instruct-v0.1 demonstrated superior performance to Llama3 70b instruct in a German RAG evaluation, according to shared dataset results. However, members noted potential issues with the evaluation metrics, especially the "question to context" metric, and suggested a possible formatting bug in the query template which might impact results.

Enhancing Chatbots with Execution Models and Haystack: Armifer91 is prototyping an "execute_model" function for chatbots, grouping certain functionalities and paralleling the MoE approach, while a GitHub notebook illustrates using the Haystack LLM framework for dynamically invoking services. Developers are exploring improvement techniques for Llama related to tokenization for fine-tuning, despite facing platform instability complaints with Hugging Face.

Whispers of German Speech Recognition: Members are trialing various Whisper models for German speech recognition such as whisper-tiny-german and whisper-base-quant-ct2, with a consensus on potential finetuning or quantization for enhanced functionality on smartphones.

Template Troubles and Tokenization Tangles: Complexities related to templates and tokenizer configurations in Llama-3 models were prevalent in discussions, with talk on zero weights for special tokens and alternative eos_tokens in conversational contexts. The ChatML template is standard, yet there are tokenizer-related challenges.

DiscoLM's German Precision Problem: Fine-tuning DiscoLM for German language applications prompted debates over the model's tokenization issues and potential strategies for improvement, with Instruct model serving as a possible foundation. Suggestions were made to follow the LeoLM training approach and connect with the occiglot team to bolster Llama3's performance in German.


Latent Space Discord

Expanding the LLM Horizon: Engineers debated the prospect of using rope to expand large language models' context window, showing enthusiasm and referencing a Perplexity AI article for in-depth understanding.

FineWeb Stirs Excitement: The announcement of FineWeb, a massive web data trove of 15 trillion tokens drew attention, with expectations high due to its superior performance markers over predecessors like RefinedWeb and C4, as disclosed on Twitter.

Frameworks in Focus: Discordants shared mixed feelings about the Hydra framework, with some appreciating its sophisticated application configuration capabilities, while others pondered over its distinctions; interest peaked with references to Hydra's GitHub repository.

Microsoft's Mighty Phi-3 Emerges: Phi-3 sparked interest with its release—operating at a grander scale than its predecessor, Phi-2, and speculated to compete with notable models like llama 3 8B; speculations fueled by insights shared through a Tweet on Phi-3's capabilities.

Perplexity.ai Makes a Financial Leap: The technical crowd took note of Perplexity.ai's successful fundraising round, touted to enhance its search engine prowess—announcement revealed in a Tweet detailing the $62.7M fundraise.


Mozilla AI Discord


Skunkworks AI Discord


LLM Perf Enthusiasts AI Discord


Datasette - LLM (@SimonW) Discord


AI21 Labs (Jamba) Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Unsloth AI (Daniel Han) ▷ #general (1118 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (167 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (716 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (76 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #suggestions (73 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):


Perplexity AI ▷ #general (1005 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (29 messages🔥):

Links mentioned:


Perplexity AI ▷ #pplx-api (3 messages):


Stability.ai (Stable Diffusion) ▷ #general-chat (1044 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ctx-length-research (5 messages):


Nous Research AI ▷ #off-topic (22 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (20 messages🔥):

Links mentioned:


Nous Research AI ▷ #general (650 messages🔥🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (78 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #project-obsidian (7 messages):

Links mentioned:


Nous Research AI ▷ #rag-dataset (89 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #world-sim (343 messages🔥🔥):

Links mentioned:


LM Studio ▷ #💬-general (635 messages🔥🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (314 messages🔥🔥):

Links mentioned:


LM Studio ▷ #announcements (1 messages):

Link mentioned: Tweet from LM Studio (@LMStudioAI): Model search / download within LM Studio may be impacted by this Hugging Face downtime. Stay tuned for updates ↘️ Quoting Hugging Face Status (@hf_status) We're experiencing some downtime on h...


LM Studio ▷ #🧠-feedback (27 messages🔥):


LM Studio ▷ #📝-prompts-discussion-chat (12 messages🔥):

Link mentioned: bartowski/Llama-3-Smaug-8B-GGUF · Hugging Face: no description found


LM Studio ▷ #🎛-hardware-discussion (59 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (10 messages🔥):


LM Studio ▷ #autogen (20 messages🔥):


LM Studio ▷ #memgpt (1 messages):


LM Studio ▷ #amd-rocm-tech-preview (42 messages🔥):

Links mentioned:


CUDA MODE ▷ #general (34 messages🔥):

Links mentioned:


CUDA MODE ▷ #triton (46 messages🔥):

Links mentioned:


CUDA MODE ▷ #cuda (8 messages🔥):


CUDA MODE ▷ #torch (2 messages):


CUDA MODE ▷ #announcements (1 messages):


CUDA MODE ▷ #algorithms (1 messages):

andreaskoepf: https://x.com/AliHassaniJr/status/1766108184630943832


CUDA MODE ▷ #beginner (27 messages🔥):

Link mentioned: Discord - A New Way to Chat with Friends & Communities: Discord is the easiest way to communicate over voice, video, and text. Chat, hang out, and stay close with your friends and communities.


CUDA MODE ▷ #pmpp-book (2 messages):


CUDA MODE ▷ #youtube-recordings (1 messages):

.bexboy: I suppose that this one session will be uploaded too?


CUDA MODE ▷ #jax (1 messages):

Link mentioned: equinox/equinox/internal/_loop/common.py at main · patrick-kidger/equinox: Elegant easy-to-use neural networks + scientific computing in JAX. https://docs.kidger.site/equinox/ - patrick-kidger/equinox


CUDA MODE ▷ #ring-attention (3 messages):

Link mentioned: GitHub - cuda-mode/axolotl at ring_attention_patching: Go ahead and axolotl questions. Contribute to cuda-mode/axolotl development by creating an account on GitHub.


CUDA MODE ▷ #off-topic (4 messages):


CUDA MODE ▷ #hqq (15 messages🔥):

Links mentioned:


CUDA MODE ▷ #llmdotc (600 messages🔥🔥🔥):

Links mentioned:


CUDA MODE ▷ #massively-parallel-crew (29 messages🔥):

Link mentioned: Flash Attention 2.0 with Tri Dao (author)! | Discord server talks: ❤️ Become The AI Epiphany Patreon ❤️https://www.patreon.com/theaiepiphany👨‍👩‍👧‍👦 Join our Discord community 👨‍👩‍👧‍👦https://discord.gg/peBrCpheKEHey g...


Eleuther ▷ #general (262 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (443 messages🔥🔥🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (35 messages🔥):

Link mentioned: Tweet from Kyo (@kyo_takano): You ARE rounding the original estimate lol Try inspecting the TeX source like you did PDF figures. To be more specific, you rounded: - E from exp(0.5267228) to 1.69 - A from exp(6.0073404) to 406.4 ...


Eleuther ▷ #interpretability-general (2 messages):

Link mentioned: Residual stream norms grow exponentially over the forward pass — LessWrong: Summary: For a range of language models and a range of input prompts, the norm of each residual stream grows exponentially over the forward pass, wit…


Eleuther ▷ #lm-thunderdome (8 messages🔥):


Eleuther ▷ #gpt-neox-dev (14 messages🔥):

Links mentioned:


HuggingFace ▷ #general (473 messages🔥🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (13 messages🔥):

Links mentioned:


HuggingFace ▷ #cool-finds (21 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (25 messages🔥):

Links mentioned:


HuggingFace ▷ #computer-vision (4 messages):

Link mentioned: GitHub - qaz812345/TrackNetV3: Implementation of paper - TrackNetV3: Enhancing ShuttleCock Tracking with Augmentations and Trajectory Rectification: Implementation of paper - TrackNetV3: Enhancing ShuttleCock Tracking with Augmentations and Trajectory Rectification - qaz812345/TrackNetV3


HuggingFace ▷ #NLP (10 messages🔥):

Link mentioned: GitHub - gnp/minbpe-rs: Port of Andrej Karpathy's minbpe to Rust: Port of Andrej Karpathy's minbpe to Rust. Contribute to gnp/minbpe-rs development by creating an account on GitHub.


HuggingFace ▷ #diffusion-discussions (10 messages🔥):


Modular (Mojo 🔥) ▷ #general (77 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (6 messages):


Modular (Mojo 🔥) ▷ #ai (3 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #🔥mojo (338 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #community-projects (35 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (19 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #🏎engine (24 messages🔥):


Modular (Mojo 🔥) ▷ #nightly (36 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (462 messages🔥🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (19 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #general-help (19 messages🔥):

Link mentioned: Axolotl - Dataset Formats: no description found


OpenAccess AI Collective (axolotl) ▷ #community-showcase (1 messages):

Link mentioned: mattshumer/Llama-3-8B-16K · Hugging Face: no description found


OpenAccess AI Collective (axolotl) ▷ #runpod-help (1 messages):

duh_kola: not axolotl related but yeah i canlt uplaod shit to hub using runpod


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (22 messages🔥):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #announcements (7 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #app-showcase (3 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (474 messages🔥🔥🔥):

Links mentioned:

📙Release Blog:…": no description founddreamgen/opus-v1.2-llama-3-8b · Hugging Face: no description foundOpenRouter: A router for LLMs and other AI modelsOpenRouter: A router for LLMs and other AI modelsFireAttention — Serving Open Source Models 4x faster than vLLM by quantizing with ~no tradeoffs: Serving Open Source Models 4x faster than vLLM by quantizing with ~no tradeoffsmicrosoft/Phi-3-mini-4k-instruct · Hugging Face: no description foundMeta: Llama 3 70B Instruct (nitro) by meta-llama | OpenRouter: Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated stron...Lynn: Llama 3 Soliloquy 8B by lynn | OpenRouter: Soliloquy-L3 is a fast, highly capable roleplaying model designed for immersive, dynamic experiences. Trained on over 250 million tokens of roleplaying data, Soliloquy-L3 has a vast knowledge base, ri...


OpenAI ▷ #ai-discussions (303 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (33 messages🔥):


OpenAI ▷ #prompt-engineering (24 messages🔥):


OpenAI ▷ #api-discussions (24 messages🔥):


LAION ▷ #general (298 messages🔥🔥):

Links mentioned:


LAION ▷ #research (38 messages🔥):

Links mentioned:


LAION ▷ #learning-ml (6 messages):


LlamaIndex ▷ #blog (6 messages):


LlamaIndex ▷ #general (188 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (5 messages):

Links mentioned:


OpenInterpreter ▷ #general (110 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (22 messages🔥):


Interconnects (Nathan Lambert) ▷ #ideas-and-feedback (39 messages🔥):

Link mentioned: AI CEO says people's obsession with reaching artificial general intelligence is 'about creating God': Arthur Mensch doesn't feel concerned about AI surpassing human intelligence, but he does worry about American tech giants dominating the field.


Interconnects (Nathan Lambert) ▷ #news (44 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-questions (9 messages🔥):


Interconnects (Nathan Lambert) ▷ #random (25 messages🔥):

Link mentioned: no title found: no description found


Interconnects (Nathan Lambert) ▷ #memes (9 messages🔥):

Link mentioned: Tweet from near (@nearcyan): current state of llm benchmarks


Interconnects (Nathan Lambert) ▷ #reads (5 messages):

Links mentioned:


Cohere ▷ #general (71 messages🔥🔥):

Links mentioned:


Cohere ▷ #project-sharing (8 messages🔥):

Links mentioned:


Cohere ▷ #collab-opps (1 messages):


LangChain AI ▷ #general (63 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (9 messages🔥):

Links mentioned:


LangChain AI ▷ #tutorials (1 messages):

Link mentioned: Building a Rental Apartment Search with Langchain's Self-Querying Retriever: In this blog post, we delve into the capabilities of Langchain's self-querying retriever, a powerful tool for bridging the gap between natural language and structured data retrieval. This retriev...


tinygrad (George Hotz) ▷ #general (26 messages🔥):

Link mentioned: React App: no description found


tinygrad (George Hotz) ▷ #learn-tinygrad (45 messages🔥):

Links mentioned:


DiscoResearch ▷ #mixtral_implementation (5 messages):

Links mentioned:


DiscoResearch ▷ #general (6 messages):

Link mentioned: notebooks/haystack2x-demos/haystack_rag_services_demo.ipynb at main · vblagoje/notebooks: Contribute to vblagoje/notebooks development by creating an account on GitHub.


DiscoResearch ▷ #discolm_german (45 messages🔥):

Links mentioned:


Latent Space ▷ #ai-general-chat (53 messages🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

Link mentioned: LLM Paper Club (TimeGPT paper WITH AUTHORS) · Zoom · Luma: This week @Vibhu hasa invited Nixtla to cover TimeGPT: https://arxiv.org/abs/2310.03589 Also submit and vote for our next paper:…


Latent Space ▷ #ai-in-action-club (1 messages):

alan_95125: Selfcheck, both the Evauator & Evaluatee models are the same by definition.


Mozilla AI ▷ #llamafile (24 messages🔥):

Links mentioned:


Skunkworks AI ▷ #general (3 messages):

Link mentioned: config.json · mattshumer/Llama-3-8B-16K at main: no description found


Skunkworks AI ▷ #datasets (1 messages):

noob_master169: OCR dataset for less popular languages? mainly looking for doc type data


Skunkworks AI ▷ #finetuning (10 messages🔥):


Skunkworks AI ▷ #moe-main (1 messages):

getovahit: Enjoyed this! Thanks for sharing your work


LLM Perf Enthusiasts AI ▷ #general (3 messages):


LLM Perf Enthusiasts AI ▷ #speed (5 messages):

Link mentioned: OpenAI API and other LLM APIs response time tracker: no description found


Datasette - LLM (@SimonW) ▷ #ai (2 messages):


Datasette - LLM (@SimonW) ▷ #llm (2 messages):

Link mentioned: Options for accessing Llama 3 from the terminal using LLM: Llama 3 was released on Thursday. Early indications are that it’s now the best available openly licensed model—Llama 3 70b Instruct has taken joint 5th place on the LMSYS arena …


AI21 Labs (Jamba) ▷ #general-chat (4 messages):

Link mentioned: Join the NSFW // 18 🍑🍒 Discord Server!: Check out the NSFW // 18 🍑🍒 community on Discord - hang out with 31716 other members and enjoy free voice and text chat.