Frozen AI News archive

Cursor reaches >1000 tok/s finetuning Llama3-70b for fast file editing

**Cursor**, an AI-native IDE, announced a **speculative edits** algorithm for code editing that surpasses **GPT-4** and **GPT-4o** in accuracy and latency, achieving speeds of over **1000 tokens/s** on a **70b** model. **OpenAI** released **GPT-4o** with multimodal capabilities including audio, vision, and text, noted to be **2x faster and 50% cheaper** than GPT-4 turbo, though with mixed coding performance. **Anthropic** introduced streaming, forced tool use, and vision features for developers. **Google DeepMind** unveiled **Imagen Video** and **Gemini 1.5 Flash**, a small model with a **1M-context** window. **HuggingFace** is distributing **$10M** in free GPUs for open-source AI models like **Llama**, **BLOOM**, and **Stable Diffusion**. Evaluation insights highlight challenges with LLMs on novel problems and benchmark saturation, with new benchmarks like **MMLU-Pro** showing significant drops in top model performance.

Canonical issue URL

AI News for 5/15/2024-5/16/2024. We checked 7 subreddits, 384 Twitters and 30 Discords (428 channels, and 6173 messages) for you. Estimated reading time saved (at 200wpm): 696 minutes.

As an AI-native IDE, Cursor edits a lot of code, and needs to do it fast, particularly Full-File Edits. They have just announced a result that

"surpasses GPT-4 and GPT-4o performance and pushes the pareto frontier on the accuracy / latency curve. We achieve speeds of >1000 tokens/s (just under 4000 char/s) on our 70b model using a speculative-decoding variant tailored for code-edits, called speculative edits."

image.png

Because the focus is solely on the "fast apply" task, the team used a synthetic data pipeline tuned to do just that:

image.png

They are a little cagey about the speculative edit alogirthm - this is all they say:

"With code edits, we have a strong prior on the draft tokens at any point in time, so we can speculate on future tokens using a deterministic algorithm rather than a draft model."

If you can figure out how to do it on gpt-4-turbo, there is a free month of Cursor Pro for you.


Table of Contents

[TOC]


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

OpenAI GPT-4o Release

Anthropic, Google, and AI Developments

AI Evaluation and Safety Considerations

AI Startups, Products and Courses

Memes and Humor


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Model Releases and Capabilities

AI Safety and Ethics

AI Applications and Use Cases

Technical Discussions and Tutorials

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries


PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord


Stability.ai (Stable Diffusion) Discord

SD3 Release: Will We Ever See It?: Despite skepticism surrounding the release and quality of Stability AI's SD3, members keep hope alive with rumors that SD3 might be held back to boost sales, but there is no firm release date or pricing information provided.

GPU Wars: 4060 TI vs 4070 TI Smackdown: The 4060 TI 16GB was pitted against the 4070 TI 12GB, with the former being recommended for ComfyUI usage while the latter was touted as better for gaming performance, though specifics were not detailed.

API Alternatives High in Demand: Members are actively seeking and debating APIs, with Forge being equated to A1111's UI for model training and asset design, and Invoke also being part of the discussion.

Workhorse GPUs Get a Benchmark: An informative benchmark site was circulated for evaluating GPU performance; it offers data on models such as 1.5 and XL Models and filters for specific hardware including Intel Arc a770.

The Divide of Dollars and Sense: Intense dialogues opened up about economic inequality, with some members emphasizing the moral and well-being costs of chasing wealth, though these were general philosophical conversations rather than specific AI-focused discourse.


OpenAI Discord


Perplexity AI Discord


HuggingFace Discord

In essence, community dialogues peaked around novel AI tools and model fine-tuning within a vibrant tapestry of technological advancements and practical implementations.


Nous Research AI Discord

Streaming the Brain for Better AI: Engineers suggest that AI could adopt a streaming-like method akin to human thought processes, referencing the Infini-attention paper as a potential framework to improve LLM's handling of long context without overwhelming their finite working memory.

Beneath the Needles, a Tougher Benchmark: The Needle in a Needlestack (NIAN) benchmark has been introduced as a more challenging test for evaluating LLMs, posing a hurdle even for robust models like GPT-4-turbo; further info available on NIAN's website and GitHub.

Unveiling Nordic NLP Treasure, Viking 7B: Viking 7B emerges as the first open-source multilingual LLM for Nordic languages, while SUPRA is presented as a cost-effective approach to retrofitting large transformers by enhancing them into Recurrent Neural Networks to improve scaling.

Hermes 2 Ω: Merging LLMs for Superior Results: Nous Research heralds the release of Hermes 2 Ω, a model merger of Hermes 2 Pro and Llama-3 Instruct, refined further, showing promising results on benchmarks and accessible on HuggingFace.

Multimodal Meld and Finetuning: Meta's release of ImageBind raises the bar with a new AI model capable of joint embedding across various modalities, while discussions enter on the potential of finetuning existing models like PaliGemma for enhanced interactivity.


LM Studio Discord


Modular (Mojo 🔥) Discord


CUDA MODE Discord

Tensor Tug of War: Engineers discussed the use of torch.tensor Accessors versus directly passing tensor.data_ptr to kernels in CUDA, with some concerned about the potential unsigned char pointers and lack of clear documentation. The conversation pointed to PyTorch's CppDocs for using Accessors and the implications on tensor efficiency.

Solving Vexing CUDA Puzzles: Members tackled the dot product problem from the CUDA puzzle repo, noting the pitfalls of floating-point overflow with naive approaches, while reduction-based kernels maintain fp32 precision. A user's experiences and code snippets, including a floating-point overflow error, were shared on GitHub Gist.

Battle Against Non-Contiguous Tensors: Discussions on torch.compile issues and custom ops in PyTorch highlighted challenges with non-contiguous tensor strides and memory cache constraints. Engineers exchanged ideas on using tags in custom op definitions, as suggested by [torch library](https://pytorch.org/docs/main/library.html) and advocated for plans to reduce compilation times for torch.compile, pointing to conversations on the PyTorch forum.

Exploring Bitnet's Quantization Quest: Enthusiasm bubbled up for Bitnet 1.58, with calls for organizing it on platforms like GitHub and digging into training-aware quantization for linear layers and 2-bit kernels. The discussions recommended centralizing efforts in the Torch AO repository, and highlighted HQQ and BitBLAS as existing solutions for bitpacking and 2-bit GPU kernels.

Footnotes on Kernel Kinks and Gadgets: A user posted a link to an article about instant apply techniques without further context, while another shared wisdom with the GPU Programming Guide, and yet another user ran into a CUDA-related ONNXRuntimeError.

Torching Into Precision and Performance: The discussions have converged on a collective effort to recalibrate CUDA streams, with suggestions on wiping the slate clean and starting over, resulting in significant discourse and corresponding GitHub Pull Requests. The elusive dream of direct NVMe to GPU DMA transfer was also mentioned, with a nod to the ssd-gpu-dma repository.


LlamaIndex Discord

Vertex AI Welcomes LlamaIndex: LlamaIndex teamed up with Vertex AI for a new RAG API, aiming to enhance users' ability to implement retrieval-augmented generation models on Vertex’s cloud platform. The community can explore the announcement via LlamaIndex's Twitter post.

GPT-4o Quartz gets Friendly with LlamaIndex: The update to LlamaIndex's create-llama now incorporates GPT-4o, providing an intuitive way to create chatbots using a simple Q&A format over user data. For additional information, there's a comprehensive breakdown on LlamaIndex’s Twitter.

LlamaParse Merges with Quivr: LlamaIndex has forged a collaboration with Quivr, resulting in LlamaParse—a tool designed to parse multifaceted document formats (.pdf, .pptx, .md) by leveraging advanced AI. A link to Twitter provides more insights on this development.

UI Tweaks Spark Joy in LlamaParse: The LlamaIndex team unveiled major enhancements to the LlamaParse UI, promising a broadened suite of functionalities for users. The GUI improvements can be seen in the latest Twitter update.

Empower Your SQL with the Right Model: The #general channel saw concerns on choosing the right embedding models for SQL tables, with users suggesting a glance at models on the MTEB Leaderboard. However, a snag was noticed since these models are generally text-centric and may not cater specifically to SQL data.

Chat through Docs with RAG: In the #ai-discussion channel, a user needed assistance for integrating Cohere AI's retrieval-augmented generation (RAG) capabilities within Llama, aspiring to create a "Chat with your docs" application. They sought community advice on methods and resources for an effective implementation.


LAION Discord

AI Powers Up the Grid: There's a buzz about the energy demands of AI, with a highlight on how a fleet of 5000 H100 GPUs can idle at a massive 375kW. This speaks volumes about the increasing energy footprint of AI technologies.

Stable Diffusion Goes Native on Mac: A project named DiffusionKit, in partnership with Stability AI, has successfully brought Stable Diffusion 3 on-device for Mac users, signaling advances in accessibility to powerful AI tools. The news arrived via a tweet, raising expectations for the open-source release.

The Open Source Compromise: A heated debate simmered around the choice between the innovative spirit of open-source ventures and the financial lure of proprietary companies, intensified by concerns over restrictive non-compete clauses, becoming more focused in light of the FTC's recent rule banning such agreements (FTC announcement).

GPT-4o Leads the Multimodal Revolution: Discussion pointed towards GPT-4o's prowess in multimodal functions, including image generation and editing, suggesting a growing consensus that multimodal models stand at the forefront of AI development.

Breakthroughs in Video Dataset and Sampling Approaches: From unveiling VidProM, a substantial dataset to accelerate text-to-video research, found in an arXiv paper, to a novel approach in overcoming the limitations of bilinear sampling for neural networks, these discussions underscored the relentless pursuit of innovation. Meanwhile, Google's Imagen 3 is making waves as a leading image generation model, with a role in creating synthetic data sets discussed eagerly by community members (Imagen 3 information).


Eleuther Discord


Interconnects (Nathan Lambert) Discord

Neural Networks Agree on Reality: Members engaged in discussions suggesting that neural networks, despite varying objectives and data, are displaying convergence towards a universal statistical model of reality within their representation spaces. Phillip Isola's recent insights support this, as shared through his project site, academic paper, and Twitter thread, showing how large language models and vision models begin sharing representations as they scale.

OpenAI Tokenization Enigma: The community pondered if OpenAI's tokenizer could be "fake," speculating that different modalities would likely necessitate distinct tokenizers. Despite skepticism, some members advocated for giving the benefit of the doubt, suggesting detailed methodologies may exist even within seemingly chaotic projects.

Anthropic Swings to Product Focus: Transitioning to a product-based approach, Anthropic embraces the necessity for marketable deliverables to enhance data refinement, amidst discussions of broader challenges facing AI organizations such as OpenAI and Anthropic, including the sustainability of their valuations and their dependence on external infrastructures.

AGI Timing Tug-of-War: Dialogues on the plausibility of approaching AGI, prompted by a Dwarkesh interview, revealed a stark divide in the community, ranging from optimism to criticism on the practicality and impact of AGI timeline predictions.

Transparency in AI Model Metrics Called Into Question: The community flagged the unexplained drop in GPT-4o’s ELO ratings and reduction in LMsys evaluation detail, sparking discussions on the need for clear communication and consistent update protocols. Resources and perspectives on this issue were exchanged through various tweets and video content.


LangChain AI Discord


OpenRouter (Alex Atallah) Discord


OpenInterpreter Discord


Datasette - LLM (@SimonW) Discord

Google I/O Overlooks LLM Reliability: Engineers in the guild highlighted the absence of discussion on LLM reliability issues at Google I/O, expressing concern for the lack of acknowledgment on the matter by key presenters.

"Sober" Take on AI: A concept for a "Sober AI" showcase was proposed to display practical, reliable AI without the hype, aiming to set realistic expectations for large language model applications.

Transforming AI: The group discussed the potential of rebranding AI as "transformative" instead of "generative" to better reflect its capabilities in altering and processing information, suggesting that this could lead to a more accurate and productive discourse.

Prompt Caching For Efficiency: Technical discussion touched on using Gemini's prompt caching to lower the cost of token usage by maintaining prompts in GPU memory, albeit with an operational cost of $4.50 per million tokens per hour.

Model Switching And Desktop Client Concerns: The technical community raised concerns about switching between LLMs mid-conversation and the potential data integrity issues it might cause. Additionally, worries were voiced that SimonW's Mac desktop solution had been abandoned, prompting discussions on alternatives for a seamless experience.


Latent Space Discord


OpenAccess AI Collective (axolotl) Discord

Falcon Versus LLaMA in Licensing Showdown: Falcon 11B and LLaMA 3's licenses sparked debate, with concerns raised about Falcon's Acceptable Use Policy updates potentially being unenforceable. Original prompt fidelity is key when applying LORA to models like LLaMA 3.

Docker Dilemmas and Data Discussions: A Docker setup for 8xH100 PCIe was successful but the SXM variant status was unclear. Meanwhile, the STEM MMLU dataset has been expanded, creating a more detailed benchmark for STEM-related AI evaluation.

Tiny But Mighty: TinyLlama Issues and Fixes: TinyLlama presented training troubles, necessitating manual launches with accelerate. Members are seeking fixes for this discrepancy, which seems to be a current challenge.

Cross-Format Conversations: The Alpaca format for training chatbots faced criticism for its inconsistent follow-up questions, driving the preference for maintaining consistent chat formats during AI training.

Hunyuan-DiT Throws Its Hat in the Ring: Attention was drawn to the Hunyuan-DiT model, a new multi-resolution diffusion transformer tailored for Chinese language processing and detailed in their arXiv paper.

Using the Right Tokens: Queries related to LLaMA 3 and ChatML tokenization were resolved with confirmation that ChatML's ShareGPT format is compatible without requiring additional special tokens.


AI Stack Devs (Yoko Li) Discord


tinygrad (George Hotz) Discord

CORDIC Conquers Complexity: Engineers discussed the advantages of the CORDIC algorithm over Taylor series for calculating trigonometric functions, addressing simplicity and speed benefits. A Python implementation and approaches for handling large argument values were deliberated, expressing concerns over precision and efficacy in machine learning applications.

Taming Trigonometry: The conversation shifted towards efficient ways to reduce arguments in trigonometric functions, ensuring precise results in an acceptable range (-π to π or -π/2 to π/2). Potential optimization paths for GPUs and fallbacks using Taylor approximations were considered for tackling large trigonometric values.

Efficient Visualization Utilized for Shape Indexing: A visualization tool to aid in understanding shape expressions in tensor reshaping operations was introduced, addressing the challenge of complex mappings. This tool is public and can be found here.

Exploring TACO for Code Generation: The community evaluated TACO, a code generator for tensor algebra, as an efficient resource for tensor computations. An exploration into using custom CUDA kernels for large tensor reductions in Tinygrad was also suggested for direct result accumulation.

Seeking Clarity on Tinygrad Operations: Clarification was sought regarding uops in a compute graph, particularly the DEFINE_GLOBAL operation and the output buffer tag, emphasizing a need for clearer documentation in low-level operations. Additionally, UseAdrenaline was recommended as a learning aid for understanding various repositories, including Tinygrad.


MLOps @Chipro Discord

Catch-Up with Members at Data AI Summit: Engineer colleagues are coordinating an informal meet-up during the Data AI Summit, scheduled for June 6-16 in the Bay Area. The suggestion has sparked mutual interest among members for an in-person connection.

Put a Pin in Monthly Casuals: The regularly scheduled casual event organized by Chip is on hold for the next few months, leaving participants to wonder about when the next social mixer might occur.

Interactive Learning Opportunity at Snowflake Dev Day: Members of the Discord have received an invitation to visit a booth at Snowflake Dev Day on June 6, promising potential insights into Snowflake's integration with data science workflows.

NVIDIA Ups the Ante with Developer Contest: There's excitement about NVIDIA & LangChain's Generative AI Agents Developer Contest, which includes the NVIDIA® GeForce RTX™ 4090 GPU among its rewards, even if geo-restrictions have dampened the spirits for some.

Exploring the Evolution of AI Hardware: An in-depth article was shared, dissecting the historical development of machine learning microprocessors and projecting future trends, noting the transformative impact of transformer-based models with a nod to Nvidia's soaring valuation. It forecasts exciting advances for NVMe drives and Tenstorrent technology, but posits a cooling period for GPUs in the mid-term future.


Cohere Discord


DiscoResearch Discord

Ilya Sutskever Bids Farewell to OpenAI: The announcement of Ilya Sutskever's departure from OpenAI ignited debate over the organization's appeal to alignment researchers, stirring concerns about its future research direction.

GPT-4-turbo Meets its Match with NIAN: The Needle in a Needlestack (NIAN) benchmark presents a new level of challenge for context-sensitive responses in large language models, with reports that "even GPT-4-turbo struggles with this benchmark." Explore the code and the website for details.


LLM Perf Enthusiasts AI Discord


Mozilla AI Discord


The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Skunkworks AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Unsloth AI (Daniel Han) ▷ #general (1022 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (27 messages🔥):

Link mentioned: Replete-AI/code_bagel · Datasets at Hugging Face: no description found


Unsloth AI (Daniel Han) ▷ #help (448 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (5 messages):


Unsloth AI (Daniel Han) ▷ #community-collaboration (1 messages):

starsupernova: Oh fantastic - if u need help - ask away!


Stability.ai (Stable Diffusion) ▷ #general-chat (966 messages🔥🔥🔥):

Links mentioned:


OpenAI ▷ #ai-discussions (280 messages🔥🔥):

Link mentioned: Reddit - Dive into anything: no description found


OpenAI ▷ #gpt-4-discussions (103 messages🔥🔥):


OpenAI ▷ #prompt-engineering (192 messages🔥🔥):

Link mentioned: ChatGPT can now access the live Internet. Can the API?: Given the news announcement I am wondering if the API now has that same access to the Internet. Thanks in advance!


OpenAI ▷ #api-discussions (192 messages🔥🔥):

Link mentioned: ChatGPT can now access the live Internet. Can the API?: Given the news announcement I am wondering if the API now has that same access to the Internet. Thanks in advance!


OpenAI ▷ #api-projects (4 messages):


Perplexity AI ▷ #general (477 messages🔥🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (12 messages🔥):


Perplexity AI ▷ #pplx-api (10 messages🔥):


HuggingFace ▷ #announcements (3 messages):

Links mentioned:


HuggingFace ▷ #general (261 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (13 messages🔥):

Link mentioned: Curiosity-driven Exploration by Self-supervised Prediction: Pathak, Agrawal, Efros, Darrell. Curiosity-driven Exploration by Self-supervised Prediction. In ICML, 2017.


HuggingFace ▷ #cool-finds (6 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (7 messages):

Link mentioned: Terminus XL - a ptx0 Collection: no description found


HuggingFace ▷ #reading-group (12 messages🔥):

Link mentioned: Join the Hugging Face Discord Server!: We're working to democratize good machine learning 🤗Verify to link your Hub and Discord accounts! | 79111 members


HuggingFace ▷ #computer-vision (10 messages🔥):

Link mentioned: tonyassi/sales1 · Datasets at Hugging Face: no description found


HuggingFace ▷ #NLP (9 messages🔥):


HuggingFace ▷ #diffusion-discussions (7 messages):

Link mentioned: Explaining the SDXL latent space: no description found


Nous Research AI ▷ #ctx-length-research (2 messages):

Link mentioned: Reddit - Dive into anything: no description found


Nous Research AI ▷ #off-topic (16 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (10 messages🔥):

Links mentioned:


Nous Research AI ▷ #announcements (1 messages):

Links mentioned:


Nous Research AI ▷ #general (199 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (55 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #project-obsidian (3 messages):

Links mentioned:


Nous Research AI ▷ #world-sim (3 messages):

Link mentioned: “Wait, this Agent can Scrape ANYTHING?!” - Build universal web scraping agent: Build an universal Web Scraper for ecommerce sites in 5 min; Try CleanMyMac X with a 7 day-free trial https://bit.ly/AIJasonCleanMyMacX. Use my code AIJASON ...


LM Studio ▷ #💬-general (145 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (66 messages🔥🔥):

Link mentioned: HuggingFaceM4/idefics-9b-instruct at main: no description found


LM Studio ▷ #🧠-feedback (8 messages🔥):

Link mentioned: VirusTotal: no description found


LM Studio ▷ #⚙-configs-discussion (2 messages):


LM Studio ▷ #🎛-hardware-discussion (46 messages🔥):


LM Studio ▷ #🧪-beta-releases-chat (9 messages🔥):


LM Studio ▷ #amd-rocm-tech-preview (7 messages):


Modular (Mojo 🔥) ▷ #general (59 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (2 messages):


Modular (Mojo 🔥) ▷ #✍︱blog (1 messages):


Modular (Mojo 🔥) ▷ #🔥mojo (166 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #nightly (35 messages🔥):

Links mentioned:


CUDA MODE ▷ #cuda (15 messages🔥):

Links mentioned:


CUDA MODE ▷ #torch (25 messages🔥):

Links mentioned:


CUDA MODE ▷ #algorithms (1 messages):

andreaskoepf: https://www.cursor.sh/blog/instant-apply


CUDA MODE ▷ #cool-links (1 messages):


CUDA MODE ▷ #beginner (8 messages🔥):


CUDA MODE ▷ #pmpp-book (5 messages):

Link mentioned: Programming-Massively-Parallel-Processors-A-Handson-Approach/Chapter 2 Heterogeneous data parallel computing/device_vector_addition_gpu.cu at main · longlnOff/Programming-Massively-Parallel-Processors-A-Handson-Approach: Contribute to longlnOff/Programming-Massively-Parallel-Processors-A-Handson-Approach development by creating an account on GitHub.


CUDA MODE ▷ #jax (1 messages):

prometheusred: https://x.com/srush_nlp/status/1791089113002639726


CUDA MODE ▷ #off-topic (3 messages):


CUDA MODE ▷ #triton-puzzles (1 messages):

Links mentioned:


CUDA MODE ▷ #llmdotc (141 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #bitnet (12 messages🔥):

Links mentioned:


LlamaIndex ▷ #blog (6 messages):

Link mentioned: RSVP to GenAI Summit Pre-Game: Why RAG Is Not Enough? | Partiful: Note: This is an in-person meetup @LlamaIndex HQ in SF! Stop by our meetup to learn about latest innovations in building production-grade retrieval augmented generation engines for your company from ...


LlamaIndex ▷ #general (155 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (3 messages):


LAION ▷ #general (144 messages🔥🔥):

Links mentioned:


LAION ▷ #research (20 messages🔥):

Links mentioned:


Eleuther ▷ #general (31 messages🔥):


Eleuther ▷ #research (51 messages🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (22 messages🔥):


Eleuther ▷ #interpretability-general (1 messages):

alofty: https://x.com/davidbau/status/1790218790699180182?s=46


Eleuther ▷ #lm-thunderdome (3 messages):


Eleuther ▷ #gpt-neox-dev (31 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #other-papers (10 messages🔥):

Link mentioned: Tweet from Phillip Isola (@phillip_isola): We survey evidence from the literature, then provide several new results including: As LLMs get bigger and better, they learn representations that are more and more similar to those learned by visi...


Interconnects (Nathan Lambert) ▷ #ml-questions (16 messages🔥):


Interconnects (Nathan Lambert) ▷ #random (64 messages🔥🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #memes (1 messages):


Interconnects (Nathan Lambert) ▷ #posts (4 messages):


LangChain AI ▷ #general (86 messages🔥🔥):

Links mentioned:


LangChain AI ▷ #langserve (2 messages):


LangChain AI ▷ #share-your-work (2 messages):


LangChain AI ▷ #tutorials (1 messages):


OpenRouter (Alex Atallah) ▷ #app-showcase (3 messages):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (82 messages🔥🔥):

Links mentioned:


OpenInterpreter ▷ #general (19 messages🔥):


OpenInterpreter ▷ #O1 (55 messages🔥🔥):

Links mentioned:


Datasette - LLM (@SimonW) ▷ #ai (64 messages🔥🔥):

Links mentioned:


Datasette - LLM (@SimonW) ▷ #llm (2 messages):


Latent Space ▷ #ai-general-chat (54 messages🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

swyxio: new pod drop! https://twitter.com/latentspacepod/status/1791167129280233696


OpenAccess AI Collective (axolotl) ▷ #general (33 messages🔥):

Link mentioned: Tencent-Hunyuan/HunyuanDiT · Hugging Face: no description found


OpenAccess AI Collective (axolotl) ▷ #general-help (5 messages):


OpenAccess AI Collective (axolotl) ▷ #datasets (3 messages):


OpenAccess AI Collective (axolotl) ▷ #runpod-help (8 messages🔥):


AI Stack Devs (Yoko Li) ▷ #ai-town-discuss (13 messages🔥):

Link mentioned: Tweet from Hugo Duprez (@HugoDuprez): Folks building @discord activities, I made a ready-to-use starter template 🕹️ Batteries included with physics and multiplayer ⚡️ @JoshLu @RamonDarioIT


AI Stack Devs (Yoko Li) ▷ #ai-town-dev (10 messages🔥):


AI Stack Devs (Yoko Li) ▷ #late-night-lounge (2 messages):

Link mentioned: The Arc PMF framework: This framework outlines three distinct archetypes of PMF which help you understand your product’s place in the market and determine how your company operates.


tinygrad (George Hotz) ▷ #general (10 messages🔥):


tinygrad (George Hotz) ▷ #learn-tinygrad (14 messages🔥):

Links mentioned:


MLOps @Chipro ▷ #events (9 messages🔥):

Link mentioned: Generative AI Agents Developer Contest by NVIDIA & LangChain: Register Now! #NVIDIADevContest #LangChain


MLOps @Chipro ▷ #general-ml (1 messages):

Link mentioned: The Past, Present, and Future of AI Hardware - SingleLunch: no description found


Cohere ▷ #general (10 messages🔥):

Link mentioned: GitHub - hkulekci/cohere-php: Contribute to hkulekci/cohere-php development by creating an account on GitHub.


DiscoResearch ▷ #general (2 messages):


DiscoResearch ▷ #benchmark_dev (1 messages):

Link mentioned: Reddit - Dive into anything: no description found


LLM Perf Enthusiasts AI ▷ #jobs (2 messages):

Link mentioned: Remote Senior Web Developer (Full-Stack) at Ambush: Ambush is looking to hire a Remote Senior Web Developer (Full-Stack) to join their team. This is a full-time position that can be done remotely anywhere in Americas, Asia, Europe or the United Kingdom...


Mozilla AI ▷ #llamafile (2 messages):

Links mentioned: