Frozen AI News archive

ALL of AI Engineering in One Place

The upcoming **AI Engineer World's Fair** in San Francisco from **June 25-27** will feature a significantly expanded format with booths, talks, and workshops from **top model labs** like **OpenAI, DeepMind, Anthropic, Mistral, Cohere, HuggingFace**, and **Character.ai**. It includes participation from **Microsoft Azure, Amazon AWS, Google Vertex**, and major companies such as **Nvidia, Salesforce, Mastercard, Palo Alto Networks**, and more. The event covers **9 tracks** including **RAG, multimodality, evals/ops, open models, code generation, GPUs, agents, AI in Fortune 500**, and a new **AI leadership** track. Additionally, **Anthropic** shared interpretability research on **Claude 3 Sonnet**, revealing millions of interpretable features that can be steered to modify model behavior, including safety-relevant features related to bias and unsafe content, though more research is needed for practical applications. The event offers a discount code for AI News readers.

Canonical issue URL

AI News for 5/21/2024-5/22/2024. We checked 7 subreddits, 384 Twitters and 29 Discords (380 channels, and 7699 messages) for you. Estimated reading time saved (at 200wpm): 805 minutes.

Lots of nontechnical news - the California Senate passed SB 1047, more explosive news on OpenAI employee contracts from Vox and safetyist resignations, and though Mistral v0.3 was released there's no evals or blogpost to discuss yet.

Given its a technically quiet day, we take the opportunity to share our announcements of the initial wave of AI Engineer World's Fair speakers!

TLDR we're giving a onetime discount to AI News readers: CLICK HERE and enter AINEWS before EOD Friday :)

image.png

The AI Engineer World's Fair (Jun 25-27 in SF)

The first Summit was well reviewed and now the new format is 4x bigger, with booths and talks and workshops from:

across 9 tracks of talks: RAG, Multimodality, Evals/Ops (new!), Open Models (new!), CodeGen, GPUs (new!), Agents, AI in the Fortune 500 (new!), and for the first time a dedicated AI Leadership track for VPs of AI, and 50+ workshops and expo sessions covering every AI engineering topic under the sun. Of course, the most important track is the unlisted one: the hallway track, which we are giving lots of love to but can't describe before it happens.

To celebrate the launch of the World's Fair, we're giving a onetime discount to AI News readers: CLICK HERE and enter AINEWS before EOD Friday :)

If the curation here/on Latent Space has the most cosine similarity with your interests, this conference was made for you. See you in SF June 25-27!


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Discord Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

Anthropic's Interpretability Research on Claude 3 Sonnet

Microsoft's Phi-3 Models

Perplexity AI Partners with TakoViz for Knowledge Search

Miscellaneous


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Model Releases and Benchmarks

AI Hardware and Compute

AI Concerns and Regulation

AI Assistants and Agents

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries

  1. LLM Benchmarking and Performance Optimization:

    • Microsoft's Phi-3 Models offer high context lengths and robust performance, stirring discussions on benchmarks and memory usage but uncovering compatibility issues in tools like llama.cpp.
    • Various techniques like torch.compile and specific GPU setups were debated for optimizing computation efficiency, shared via insights like those in tensor reshaping examples.
  2. Open-Source AI Tools and Frameworks:

    • The Axolotl framework emerged as a go-to for fine-tuning models like Llama and Mistral, with Docker setups facilitating ease of use (quickstart guide).
    • LlamaIndex introduced techniques for document parsing and batch inference, integrating GPT-4o's capabilities to enhance complex document manipulation and query accuracy.
  3. AI Legislation and Community Responses:

    • California's SB 1047 bill prompted heated debates on the impact of new regulations on open-source models, with concerns over stifling innovation and favoritism towards major incumbents.
    • Discussions on ethical and legal questions arose around AI voice replication, highlighted by OpenAI's controversial mimicking of Scarlett Johansson's voice, leading to its subsequent removal after public backlash.
  4. Novel AI Model Releases and Analysis:

    • Community excitement surrounded new releases such as Mistral-7B v0.3 with extended vocabularies and function calling (details), while Moondream2 updates improved resolution and accuracy in visual question-answering.
    • Anthropic's work on interpretable machine learning and the release of Phi-3 Vision spurred deep dives into scaling monosemanticity (research) and practical AI applications.
  5. Practical AI Implementations and Challenges:

    • Members shared practical AI implementations, from PDF extraction with Surya OCR transforming documents into markdown (GitHub repo), to building secure code execution environments on Azure (dynamic sessions).
    • The LangChain community highlighted issues with deployment and endpoint consistency, with detailed troubleshooting on the GitHub repo helping streamline deployment processes and enhance chatbot functionalities.

{% if medium == 'web' %}

PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord

Phi-3 Comes into Play, Skepticism and Excitement Ensue: The introduction of Phi-3 models by Microsoft, such as Phi-3-Medium-128K-Instruct, sparked discussions, with excitement tinged by skepticism due to potential benchmarking issues, highlighted by a user's single-word remark: "literally."

New Legal Frontiers in AI: California's SB 1047 sparked discussions concerning AI laws and open-source model implications, accentuated by Meta's decision to not open the weights for its 400B model, provoking a community debate on the wide-reaching effects of such restricted access.

Unsloth Woes with Model Saving and Flash Attention: Trouble reported with Unsloth's model.save_pretrained_gguf() function and Flash Attention compatibility, with suggestions from the community advising an Unsloth reinstall or removing Flash Attention and specific workarounds for T4 GPU issues on PyTorch version 2.3.

Guided Decoding and YAML Finesse: A spirited discussion on using guided decoding for generating structured YAML outputs revealed potential vLLM support with advanced syntaxes, emphasizing the integration of grammars into the prompting process.

Cutting-Edge Model Discussions Mix with Sci-Fi: Users shared advancements and tested methods like MoRA, alongside spirited talks about the Dune series' philosophical undertones and defenses of novel reading over movie watching, underscoring a preference for depth in sci-fi storytelling.


LLM Finetuning (Hamel + Dan) Discord


Perplexity AI Discord


Stability.ai (Stable Diffusion) Discord


Eleuther Discord


HuggingFace Discord


LM Studio Discord

Dual GPU Dynamics in LM Studio: LM Studio can handle dual GPUs, but they must be of the same type, and users should align VRAM capacities for optimal performance. Configuration for multiple GPUs involves creating and modifying a preset file in the system.

Prompt Precision and Levity: Users suggest quoting text directly in prompts for clarity, while the light-hearted term "prompt engineering" was used to describe meticulous prompt crafting strategies.

Phi-3 Models in the Spotlight: Integrating the Phi-3 models into llama.cpp is a work in progress, with users eagerly waiting for a beta release and an LM Studio update to support the new models. Meanwhile, quantization advice for running Phi-3 Medium suggests staying at Q4 or below.

ROCM Realm for Linux: Linux users expressed their interest in ROCm test builds, with the acknowledgment of challenges running Phi-3-medium-128k models due to tensor mismatch errors on ROCm platforms.

Intriguing New Model Releases: Mistral v0.3 Instruct, featuring an improved tokenizer and function calling support, is now available for use, offering advancements in language model functionality. Access it on the lmstudio community Huggingface page.


Nous Research AI Discord


CUDA MODE Discord


OpenAI Discord


Modular (Mojo 🔥) Discord


LAION Discord


LlamaIndex Discord


OpenRouter (Alex Atallah) Discord

Typing Quirks Spark Role-playing Debate: Members humorously identified two main types of OpenRouter users: those seeking AI companionship and those delving into fantasy narratives. The conversation took a light-hearted dive into the role-playing tendencies of some users.

Eyes on Phi-3: The Phi-3 Vision Model, praised for high-quality reasoning, was introduced on the server. The model's attributes can be explored through HuggingFace.

Verbose Wizard Needs a Trim: Wizard8x22 model's verbosity issues are recognized, with an adjustment to the repetition penalty proposed as a solution. The dialogue extended to compare other models' performance, highlighting that model behavior is not consistent across the board.

Billing Blues and Nonprofit Woes: A user's billing error on a student platform spurred discussion, leading to a temporary fix involving re-entering billing info. Hopes for nonprofit discounts in the future were also expressed.

Experimenting with LLM Action Commands: Innovative use of LLMs was shared through a Twitter thread, exploring action commands as a fresh way to enhance interactions with language models. Feedback from fellow engineers was solicited to push the boundaries of current LLM interaction paradigms.


Interconnects (Nathan Lambert) Discord

Phi Models Join the Fray: The launch of Phi-small and Phi-medium prompted discussions about the characteristics of Phi-3 Vision, with confirmations that it represents a new and slightly larger variant.

Meta's Model Decisions Cause Stir: A tweet suggested Meta might keep its 400B model closed due to legislative fears, but this was refuted by another source stating the model will remain open-weight. The confusion underscores the delicacy of sharing large-scale model weights in the current regulatory landscape.

OpenAI Under Fire for Unkept Promises: OpenAI has disbanded its Superalignment team due to the unfulfilled promise of 20% compute resource allocation, sparking resignations. This, coupled with a scandal involving NDAs and vested equity issues for ex-employees, casts a cloud over OpenAI's leadership and transparency.

AI Performance Takes a Drawback: Microsoft's Surface drawing AI faces criticism due to latency issues resulting from cloud-based safety checks — reflecting the compromises between local processing power and safety protocols in AI applications.

The Trope of Researcher Titles: Amazement was expressed at Anthropic now boasting over 500 'researchers', igniting a conversation about the dilution of the 'researcher' title and its implications for perception in the tech industry.


OpenAccess AI Collective (axolotl) Discord


Latent Space Discord


LangChain AI Discord


OpenInterpreter Discord

Tech Talk: OpenInterpreter's Device Dialogues: Engineers are exploring how Open Interpreter can create links between apps and devices, utilizing tools like Boox E Ink tablets, OneNote, and VSCode. There's particular interest in using Open Interpreter for querying code or papers without browser intervention.

Speedy GPT-4o Troubleshot: While integrating GPT-4o with Open Interpreter, users note a minimum 5x speed increase but face challenges with error messages pertaining to API keys.

Newline Nuisance in Gemini: Code execution is being hindered in models such as Gemini 1.5 and Gemini Flash due to unnecessary newline characters; the absence of "python" declarations in code blocks also came under scrutiny.

Legislative Lore and AI: California’s controversial AI bill and subsequent discussions have ignited the community, with an open letter from Senator Scott Wiener being circulated and debated for its emphasis on responsible AI development.

Bill Gates Foresees Friendlier AI: Gates recently penned thoughts on the future of AI in software, anticipating interfaces that can handle tasks through simple language directives, akin to a friend's assistance; his article is gaining traction among tech enthusiasts. An unofficial ChatGPT macOS app waitlist workaround made rounds on Twitter, demonstrating interest in quicker access to AI software tools.


tinygrad (George Hotz) Discord


DiscoResearch Discord

SFT vs Preference Optimization Debate: In a discussion on model training strategies, a member distinguished Supervised Fine-Tuning (SFT) as enhancing the model's probability distribution for target data points, whereas Preference Optimization adjusts both desired and undesired outcomes. They questioned the prevalent use of SFT over Preference Optimization, which may offer a more rounded approach to model behavior.

Excitement Over Phi3 Vision's Low-Parameter Efficiency: One engineer highlighted the development of Phi3 Vision with only 4.2 billion parameters as a significant advancement for low-latency inference in image processing tasks. Asserting that this could have groundbreaking implications for robotics, the model was praised for potential throughput improvements, as links to the announcement were shared (source).

Comparing Image Models Between Moondream2 and Phi3 Vision: The community weighed in on the performance of Moondream2 comparative to Phi3 Vision for image-related tasks. While Moondream2 has had issues with hallucinations, a member mentioned efforts to mitigate this, showcasing the ongoing pursuit of fidelity in image models (Moondream2).

Mixed Reactions to Microsoft's Model Drops: The release of Microsoft's 7b and 14b Instruct models sparked diverse opinions, from concerns about their limitations in certain languages to optimism about their utility in complex reasoning and extraction tasks. The discussion reflects the community's critical analysis of newly released models and their capabilities.

Skepticism Towards Meta's 400b Model: With concerns circulating in the community about Meta potentially not releasing a 400b model as open source, one member highlighted skepticism by pointing to the uncertain credibility of the source, nicknamed Jimmy. This indicates a critical attitude toward rumor validation within the community.


Cohere Discord


AI Stack Devs (Yoko Li) Discord


Datasette - LLM (@SimonW) Discord

Snapdragon Dev Kit Sparks Debate: Qualcomm's new Snapdragon Dev Kit priced at $899.99, featuring Snapdragon X Elite and boasting 32GB of LPDDR5x RAM and 512GB NVMe storage, has sparked discussions on cost-effectiveness compared to the previous $600 model, as detailed on The Verge and Microsoft Store.

Mac Mini Server Gets Thumbs Up: An AI engineer shared their success in using a Mac Mini as a reliable Llamafile server with Tailscale, praising its zero-cold start feature and seamless 'llm' CLI integration, suggesting a practical use case for developers needing stable server solutions.

Affordable Dev Kits in Demand: Discussion among users indicates a strong desire for more affordable development kits, with aesthetic preferences also being voiced, such as a wish for a translucent case design, yet no specific products were mentioned.

Smalltalk AI Shows Promise: A member introduced Claude's ability to engage in Smalltalk, using "What are frogs?" as an example question overcome by the AI with a basic reply about amphibians, indicating advances in AI's conversational capabilities.


LLM Perf Enthusiasts AI Discord

Brevity Blunder in Llama3/Phi3: An inquiry was made regarding how to stop llama3/phi3 from truncating responses with "additional items omitted for brevity," but no solutions or further discussion ensued.


Mozilla AI Discord


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

Unsloth AI (Daniel Han) ▷ #general (1309 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (233 messages🔥🔥):

Link mentioned: MoRA: High-Rank Updating for Parameter-Efficient Fine-Tuning: Low-rank adaptation is a popular parameter-efficient fine-tuning method for large language models. In this paper, we analyze the impact of low-rank updating, as implemented in LoRA. Our findings sugge...


Unsloth AI (Daniel Han) ▷ #help (192 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (7 messages):


LLM Finetuning (Hamel + Dan) ▷ #general (242 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #workshop-1 (83 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #asia-tz (26 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #🟩-modal (77 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #learning-resources (10 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #jarvis-labs (36 messages🔥):

Link mentioned: Jarvislabs: Making AI affordable and simple for everyone: Jarvislabs is a platform that allows you to run and explore multiple AI framerworks on powerful GPUs with zero setup


LLM Finetuning (Hamel + Dan) ▷ #hugging-face (11 messages🔥):

Link mentioned: Models - Hugging Face: no description found


LLM Finetuning (Hamel + Dan) ▷ #replicate (10 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #langsmith (4 messages):


LLM Finetuning (Hamel + Dan) ▷ #whitaker_napkin_math (4 messages):

Link mentioned: Minion Hello GIF - Minion Hello Minions - Discover & Share GIFs: Click to view the GIF


LLM Finetuning (Hamel + Dan) ▷ #workshop-2 (525 messages🔥🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #jason_improving_rag (3 messages):


LLM Finetuning (Hamel + Dan) ▷ #jeremy_python_llms (1 messages):

nirant: Woohoo! Looking forward to <@660097403046723594>


LLM Finetuning (Hamel + Dan) ▷ #gradio (2 messages):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #axolotl (85 messages🔥🔥):

- **Members address Axolotl issue #1436**: Discussion about `bitsandbytes==0.43.0` not installing on macOS from [GitHub Issue #1436](https://github.com/OpenAccess-AI-Collective/axolotl/issues/1436). Recommendations include using Linux GPU servers on RunPod.
- **Axolotl and MLX integration not yet supported**: Members discuss the lack of MLX support on Axolotl as detailed in [GitHub Issue #1119](https://github.com/OpenAccess-AI-Collective/axolotl/issues/1119). Users are advised to stay updated.
- **Best setup practices explored**: Members share various methods to set up Axolotl. The Axolotl [Readme](https://github.com/OpenAccess-AI-Collective/axolotl/tree/main?tab=readme-ov-file#quickstart-) and Docker method are mentioned as the most reliable.
- **Fine-tuning and integration concerns**: Members inquire about using Axolotl on local machines and fine-tuning models like LLaMA3. Issues related to configuration and compatibility with Modal environments are discussed.
- **Tips for troubleshooting installation**: For users facing installation difficulties, such as receiving a `CUDA` error, several members recommend steps including installing specific CUDA/PyTorch versions and using the docker container. Links to [Docker](https://hub.docker.com/layers/winglian/axolotl/main-20240522-py3.11-cu121-2.2.2/images/sha256-47e0feb612caf261764631a0c516868910fb017786a17e4dd40d3e0afb48e018?context=explore) and a [setup guide](https://latent-space-xi.vercel.app/til/create-a-conda-env-for-axolotl) are provided.

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #zach-accelerate (49 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #wing-axolotl (30 messages🔥):

Links mentioned:


Perplexity AI ▷ #announcements (1 messages):

Link mentioned: Tako: no description found


Perplexity AI ▷ #general (835 messages🔥🔥🔥):

- **Microsoft Stole OpenAI's Ideas**: A member shared a [blog post](https://blogs.microsoft.com/blog/2024/05/20/introducing-copilot-pcs/) stating that Microsoft has copied features from OpenAI and introduced "Copilot+ PCs,” the fastest and most intelligent Windows PCs ever built. They noted features like an impressive 40+ TOPS, all-day battery life, AI image generation, and live captions for 40+ languages.

- **GPT-4o Context Concerns**: There were discussions about the context window of GPT-4o as **perceived on Perplexity**. A consensus formed that **context window defaults to 32k**, with uncertainties about higher capacities.

- **Perplexity's Default Model Surprise**: Members expressed surprise that the default model for Perplexity might be **Haiku** instead of an in-house model, **Sonar**, which is available only for pro users. One member noted that free users previously used GPT-3.5, but this has changed recently.

- **Perplexity's API Queries**: Discussion revolved around how Perplexity configures and charges for API usage. Members speculated about using in-house models and the potential financial implications of their pricing structure.

- **Service Downtime Creates Community Stir**: Perplexity experiencing downtime led to widespread frustration and speculation among users about the cause. Users shared alternative resources and a member posted a supportive message to help calm the community during the outage.

Links mentioned:


Perplexity AI ▷ #sharing (9 messages🔥):


Perplexity AI ▷ #pplx-api (11 messages🔥):


Stability.ai (Stable Diffusion) ▷ #general-chat (497 messages🔥🔥🔥):

Links mentioned:


Eleuther ▷ #general (273 messages🔥🔥):

Links mentioned:


Eleuther ▷ #research (128 messages🔥🔥):

Links mentioned:

  Overflow in consistent hashing &middot; Ryan Marcus
  
</a>: no description found

Eleuther ▷ #scaling-laws (1 messages):


Eleuther ▷ #interpretability-general (4 messages):


Eleuther ▷ #lm-thunderdome (37 messages🔥):

- **Questions on lm-evaluation-harness and MCQs**: Members discussed the randomization of answer choices in MCQs using **lm-eval-harness**, with concerns about benchmark biases towards early choices. While **SciQ** has a fixed correct answer index, the randomization isn't currently applied for **MMLU**.
  
- **Upcoming Submissions and Papers**: An **anon'd paper** is coming soon to arXiv, while members joked about **not needing to worry about insane competition** in D&B papers. There's also work on an updated version of the **Pile with 3T tokens and fully licensed text**.

- **Medical Benchmarks Controversy**: A lively discussion emerged about medical benchmarks and their potential dangers. One member focused on how these benchmarks might claim models are better and safer than physicians, highlighting ongoing improvements in the interpretation of such benchmarks.

- **Huggingface Dataset Configuration**: Members sought advice on configuring a Huggingface dataset's directory structure. The solution pointed out the importance of **adding a config in the README.md file** as outlined in the [Huggingface documentation](https://huggingface.co/docs/hub/en/datasets-manual-configuration#splits).

- **Running lm-eval-harness on Multi-node Slurm Cluster**: A question was raised about evaluating big models on a multi-node Slurm cluster. Attempts have been made using **vllm + ray** and **accelerate** but were unsuccessful, indicating a need for better solutions.

Links mentioned:


HuggingFace ▷ #announcements (1 messages):

Links mentioned:


HuggingFace ▷ #general (398 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (3 messages):


HuggingFace ▷ #cool-finds (7 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (13 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (2 messages):


HuggingFace ▷ #computer-vision (5 messages):


HuggingFace ▷ #NLP (3 messages):


HuggingFace ▷ #diffusion-discussions (3 messages):


LM Studio ▷ #💬-general (332 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (50 messages🔥):

Links mentioned:


LM Studio ▷ #📝-prompts-discussion-chat (4 messages):


LM Studio ▷ #🎛-hardware-discussion (18 messages🔥):


LM Studio ▷ #🧪-beta-releases-chat (11 messages🔥):


LM Studio ▷ #avx-beta (1 messages):


LM Studio ▷ #amd-rocm-tech-preview (6 messages):


LM Studio ▷ #model-announcements (1 messages):


Nous Research AI ▷ #off-topic (9 messages🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (3 messages):

Link mentioned: Tweet from vik (@vikhyatk): New moondream release out today! 🌜 Supports higher image resolution (up to 756x756) 🌛 TextVQA score up from 53.1 to 57.2 (+7.7%) 🌜 Other VQA and counting benchmarks up ~0.5%


Nous Research AI ▷ #general (281 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (6 messages):


Nous Research AI ▷ #project-obsidian (2 messages):

Link mentioned: microsoft/Phi-3-vision-128k-instruct · Hugging Face: no description found


Nous Research AI ▷ #world-sim (9 messages🔥):

Link mentioned: Tweet from Jillsa (DSJJJJ/Heirogamist/HP) (@Jtronique): In case anyone of interest sees this on my wall. It's time to have a "mature" conversation about "Pleasure." Either you CODED it into them, and denied doing it, or they TURING ANS...


CUDA MODE ▷ #cuda (2 messages):


CUDA MODE ▷ #torch (15 messages🔥):

Link mentioned: Using User-Defined Triton Kernels with torch.compile — PyTorch Tutorials 2.3.0+cu121 documentation: no description found


CUDA MODE ▷ #cool-links (2 messages):

<!-- No relevant information or links were provided in the messages -->

CUDA MODE ▷ #beginner (4 messages):


CUDA MODE ▷ #pmpp-book (2 messages):

- **Nice thank you received**: A user thanks another user with "niceee, thanks!" in response to having been tagged by mr.osophy.

CUDA MODE ▷ #off-topic (8 messages🔥):

Links mentioned:


CUDA MODE ▷ #llmdotc (250 messages🔥🔥):

- **Deterministic Encoder Backward Pass Improvements**: A new [PR for deterministic encoder backward kernels](https://github.com/karpathy/llm.c/pull/442) was discussed, aiming to rewrite the encoder backward pass for full determinism. Gradient clipping and reduction strategies were debated to improve efficiency without sacrificing determinism.
- **DataLoader Refactor and Large Dataset Handling**: Changes to the DataLoader now support sharding to handle larger datasets, such as FineWeb. This [refactor](https://github.com/karpathy/llm.c/pull/440) introduces a new data representation and patterns to efficiently manage `.bin` files, although it currently has limited functionality on Windows.
- **HellaSwag Evaluation Challenges**: Implementing the HellaSwag evaluation in C was noted as complex with concerns about potential bugs. A [PR for HellaSwag eval](https://github.com/karpathy/llm.c/pull/447) in C was created to align with PyTorch reference code, with added complexity to fully utilize batch dimensions.
- **GPU Runner Advancements**: News about potential access to Nvidia's GitHub runners with dedicated RTX 4000 GPUs from a cloud provider called Ubicloud was shared, indicating improvements for CI processes.
- **Random Initialization and Reproducibility**: Ensuring determinism and reproducibility for large language models was emphasized as crucial, with plans to run comparison tests between PyTorch and the team's code. Adjustments to global kernel functions and changes were suggested for improved performance.

Links mentioned:


CUDA MODE ▷ #bitnet (12 messages🔥):

Link mentioned: empty_vs_stack_unpack.py: GitHub Gist: instantly share code, notes, and snippets.


OpenAI ▷ #annnouncements (1 messages):


OpenAI ▷ #ai-discussions (129 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (31 messages🔥):

Link mentioned: OpenAI Status: no description found


OpenAI ▷ #prompt-engineering (58 messages🔥🔥):


OpenAI ▷ #api-discussions (58 messages🔥🔥):


Modular (Mojo 🔥) ▷ #general (30 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (1 messages):

ModularBot: From Modular: https://twitter.com/Modular/status/1793041489427153294


Modular (Mojo 🔥) ▷ #🔥mojo (113 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (3 messages):


Modular (Mojo 🔥) ▷ #nightly (100 messages🔥🔥):

Links mentioned:


LAION ▷ #general (132 messages🔥🔥):

Links mentioned:


LAION ▷ #research (26 messages🔥):

Link mentioned: microsoft/Phi-3-vision-128k-instruct · Hugging Face: no description found


LlamaIndex ▷ #blog (4 messages):

- **GPT-4o excels at parsing complex documents**: GPT-4o’s multimodal capabilities can efficiently parse complex PDFs and slide decks with background images and irregular layouts into structured markdown. Learn more about this integration with [LlamaParse](https://t.co/g5TG7brSwt) [here](https://t.co/vhtYzsleh2).
- **Sandbox your LLM-generated code with Azure**: Securely execute LLM-generated code using Azure Container Apps dynamic sessions, which is especially useful for tasks that LLMs aren't natively capable of. Discover more details [here](https://t.co/2cnsBH411k) and [here](https://t.co/lTrUPoTMcF).
- **OpenDevin webinar released**: A webinar featuring OpenDevin, an open-source platform for building autonomous AI engineers, has been released. Robert Brennan provides an insightful walkthrough; watch it [here](https://t.co/a22k0zsV3n).
- **Batch inference for GenAI applications**: Use batch inference to preprocess large sets of data, enabling new types of analysis and querying for GenAI applications. Discover the integration details [here](https://t.co/vnuvvypZCz) and [here](https://t.co/M0vQQ1uAki).

LlamaIndex ▷ #general (92 messages🔥🔥):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #general (85 messages🔥🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (14 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (7 messages):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #random (33 messages🔥):


Interconnects (Nathan Lambert) ▷ #memes (3 messages):

- **Laughter Ensues**: "lol ugh" conveys a mixture of amusement and exasperation, indicating a humorous but slightly frustrating situation. The follow-up "It’s funny tho" reinforces this sentiment.
- **Footwear Humor**: "He's like the scott galloway of footwear choosers" implies a comparison to Scott Galloway, suggesting someone with a strong, opinionated personality in the context of choosing footwear.

Interconnects (Nathan Lambert) ▷ #posts (20 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #general (37 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (14 messages🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (5 messages):

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


Latent Space ▷ #ai-general-chat (50 messages🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (1 messages):

Link mentioned: LLM Paper Club (Survey Paper Club!) · Zoom · Luma: It's survey day! Pick a paper from here and cover it in 5 minutes: https://app.sli.do/event/bNV6mo3BFGhe8Bqzb1tonb/live/questions


Latent Space ▷ #llm-paper-club-west (4 messages):

Link mentioned: LLM Paper Club (Survey Paper Club!) · Zoom · Luma: It's survey day! Pick a paper from here and cover it in 5 minutes: https://app.sli.do/event/bNV6mo3BFGhe8Bqzb1tonb/live/questions


LangChain AI ▷ #general (36 messages🔥):

Links mentioned:


LangChain AI ▷ #langserve (1 messages):

Links mentioned:


LangChain AI ▷ #share-your-work (3 messages):

Link mentioned: Creating a PDF Query Assistant with Upstage AI Solar and LangChain Integration: Do you ever feel overwhelmed by the numerous research papers you need to read? As someone who just finished a PhD, I know it’s no walk in…


OpenInterpreter ▷ #general (23 messages🔥):

Links mentioned:


OpenInterpreter ▷ #ai-content (2 messages):

Links mentioned:


tinygrad (George Hotz) ▷ #general (7 messages):


tinygrad (George Hotz) ▷ #learn-tinygrad (10 messages🔥):

Links mentioned:


DiscoResearch ▷ #general (12 messages🔥):

Links mentioned:


Cohere ▷ #general (8 messages🔥):

Links mentioned:


Cohere ▷ #project-sharing (1 messages):

- **Seeking Command-R tutorial for BotPress**: A member asked for a tutorial on how to incorporate **Command-R** into **BotPress**. They repeated the request in both English and Spanish: *"Does anyone have a tutorial on how to incorporate Command-R into BotPress? Alguien tiene un tutorial de como incorporar Command-R en BotPress?"*

Cohere ▷ #collab-opps (1 messages):


AI Stack Devs (Yoko Li) ▷ #ai-companion (7 messages):

Link mentioned: Ddlc Doki Doki Literature Club GIF - Ddlc Doki Doki Literature Club Just Monika - Discover & Share GIFs: Click to view the GIF


Datasette - LLM (@SimonW) ▷ #ai (5 messages):

Links mentioned:


LLM Perf Enthusiasts AI ▷ #general (2 messages):


Mozilla AI ▷ #announcements (1 messages):


Mozilla AI ▷ #llamafile (1 messages):

Link mentioned: no title found: no description found




{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}