Frozen AI News archive

Hybrid SSM/Transformers > Pure SSMs/Pure Transformers

**NVIDIA**'s Bryan Catanzaro highlights a new paper on **Mamba models**, showing that mixing Mamba and Transformer blocks outperforms either alone, with optimal attention below **20%**. **Mixture-of-Agents (MoA)** architecture improves LLM generation quality, scoring **65.1% on AlpacaEval 2.0** versus **GPT-4 Omni's 57.5%**. The **LiveBench AI benchmark** evaluates reasoning, coding, writing, and data analysis. A hybrid **Mamba-2-Hybrid** model with **7% attention** surpasses a Transformer on MMLU accuracy, jumping from **50% to 53.6%**. **GPT-4** performs better at temperature=1. **Qwen 72B** leads open-source models on LiveBench AI. **LaminiAI Memory Tuning** achieves **95% accuracy** on a SQL agent task, improving over instruction fine-tuning. **Sakana AI Lab** uses evolutionary strategies for preference optimization. **Luma Labs Dream Machine** demonstrates advanced text-to-video generation. The **MMWorld benchmark** evaluates multimodal video understanding, and **Table-LLaVa 7B** competes with GPT-4V on multimodal table tasks.

Canonical issue URL

AI News for 6/12/2024-6/13/2024. We checked 7 subreddits, 384 Twitters and 30 Discords (414 channels, and 3646 messages) for you. Estimated reading time saved (at 200wpm): 404 minutes. You can now tag @smol_ai for AINews discussions!

Lots of fun image-to-video and canvas-to-math demos flying around today, but not much technical detail, so we turn elsewhere, to Bryan Catanzaro of NVIDIA calling attention to their new paper studying Mamba models:

image.png

As Eugene Cheah remarked in the Latent Space Discord, this is the third team (after Jamba and Zamba that has independently found the result that mixing Mamba and Transformer blocks does better than either can alone. And the paper does conclude empirically that the optimal amount of Attention is <20%, being FAR from all you need.

image.png

image.png


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

LLM Capabilities and Evaluation

LLM Training and Fine-Tuning

Multimodal and Video Models

Open-Source Models and Datasets


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

Stable Diffusion 3 Medium Release

Issues and Limitations of SD3 Medium

Comparisons with Other Models

Community Reactions and Speculation

Memes and Humor


AI Discord Recap

A summary of Summaries of Summaries

  1. Stable Diffusion 3 Faces Scrutiny but Offers Alternatives:

    • SD3 Faces Criticism for Model Quality: Users conveyed dissatisfaction with SD3—highlighting anatomical inaccuracies and prompt issues—while medium models can be downloaded on Huggingface.
    • Preferred Interfaces & Tools Discussed: ComfyUI emerged as the favored interface, with suggested samplers like uni_pc and ddim_uniform for optimal performance. Alternatives like Juggernaut Reborn and Playground are highlighted for their specific capabilities.
  2. Boosting AI Performance and Infrastructure Insights:

    • LLM Performance Boosted by Higher Model Rank: Shifting from rank 16 to 128 resolved Qwen2-1.5b's gibberish output, aligning it with llama-3 caliber outputs.
    • Perplexity AI's Efficient LLM Use: Quick results are achieved by leveraging NVIDIA A100 GPUs, AWS p4d instances, and TensorRT-LLM optimizations.
  3. Innovations in Fine-Tuning and Quantization:

    • Fine-Tuning LLMs with New Models: The discussion covered the legal aspects of using GPT-generated data, referencing OpenAI's business terms. Experimentations with ToolkenGPT show creative approaches to synthetic data for fine-tuning.
    • CUDA Quantization Project discussions: Projects like the BiLLM showcase rapid quantization of large models, essential for efficient AI deployments.
  4. Model Management and Deployment Techniques:

    • Strategies for Handling Large Embeddings: Queries about 170,000 embedding indexes led to recommendations on using Qdrant or FAISS for faster retrieval. Specific fixes for erroneous queries were shared here.
    • Docker and GPU Configuration Troubleshooting: Users dealing with Docker GPU detection on WSL found solutions by consulting the official NVIDIA toolkit guide.
  5. AI Community Trends and Updates:

    • OpenAI's Revenue Milestone and Focus Shift: OpenAI's revenue doubled, reflecting sales direct from ChatGPT and other services, not primarily facilitated by Microsoft (source).
    • Partnerships and Conferences Engage Community: Aleph Alpha and Silo AI joined forces to advance European AI (read more), and Qwak's free virtual conference promises deep dives into AI mechanisms and networking opportunities.

PART 1: High level Discord summaries

Stability.ai (Stable Diffusion) Discord

SD3's Rocky Release: Users have expressed dissatisfaction with Stable Diffusion 3 (SD3), citing issues such as anatomical inaccuracies and non-compliance with prompts compared to SDXL and SD1.5. Despite the critiques, the medium model of SD3 is now downloadable on Huggingface, requiring form completion for access.

Preferred Interfaces and Samplers: ComfyUI is currently the go-to interface for running SD3, and users are advising against Euler samplers. The favored samplers for peak performace with SD3 are uni_pc and ddim_uniform.

Exploring Alternatives: Participants in the channel have highlighted alternative models and tools like Juggernaut Reborn and Divinie Animemix to achieve more realism or anime style, respectively. Other resources include Playground and StableSwarm for managing and deploying models.

Keep Discussions Relevant: Moderators have had to direct conversations back on topic after detours into global politics and personal anecdotes sidetracked from the technical AI discussions.

Big Models, Bigger Needs: The 10GB model of SD3 was mentioned as a very sought-after option among the community, showing the desire for larger, more powerful models despite the mixed reception of the SD3 release.


Unsloth AI (Daniel Han) Discord


HuggingFace Discord

SD3 Revolutionizes Stable Diffusion: Stable Diffusion 3 (SD3) has dropped with a plethora of enhancements - now sporting three formidable text encoders (CLIP L/14, OpenCLIP bigG/14, T5-v1.1-XXl), a Multimodal Diffusion Transformer, and a 16 channel AutoEncoder. Details of SD3's implementation can be found on the Hugging Face blog.

Navigating SD3 Challenges: Users encountered difficulties with SD3 on different platforms, with recommendations such as applying pipe.enable_model_cpu_offload() for faster inference and ensuring dependencies like sentencepiece are installed. GPU setup tips include using RTX 4090, employing fp16 precision, and making sure paths are correctly formulated.

Hugging Face Extends Family With Argilla: In an exciting turn of events, Hugging Face welcomes Argilla into its fold, a move celebrated by the community for the potential to advance open-source AI initiatives and new collaborations.

Community and Support in Action: From universities, such as the newly created University of Glasgow organization on Hugging Face, to individual contributions like Google Colab tutorials for LLM, members have been contributing resources and sourcing support for their various AI undertakings.

Enriched Learning Through Shared Resources: Members are actively exchanging knowledge, with highlighted assets including a tutorial for LLM setup on Google Colab, a proposed reading group discussion on the MaPO technique for text-to-image models, and an Academic paper on NLP elucidating PCFGs.


OpenAI Discord


LLM Finetuning (Hamel + Dan) Discord


Nous Research AI Discord


Perplexity AI Discord


CUDA MODE Discord

Compute Intensity Discussion Left Hanging: A member inquired whether the compute intensity calculation should consider only floating-point operations on data from Global Memory. The topic remained open for discussion without a conclusive answer.

Streamlined Triton 3.0 Setup: Two practical installation methods for Triton 3.0 surfaced; one guide details installing from source, while another involves using make triton with a specific version from the PyTorch repository.

Optimizing Optimizers in PyTorch: A robust conversation on creating a fast 8-bit optimizer using pure PyTorch and torch.compile, as well as making a drop-in replacement for 32-bit with comparable accuracy was had, drawing inspiration from the bitsandbytes implementation.

Breakthroughs in Quantization and Training Dynamics: The BiLLM project boasts rapid quantization of large language models, while torchao members debate the trade-offs in speed and accuracy across various numeric representations during matrix multiplication, from INT8 to FP8 and even INT6.

Hardware Showdown and Quantization Innovations: AMD's MI300X showcases higher throughput for LLM inference than NVIDIA's H100, and Bitnet sees progress with refactoring and nightly build strategies, but a lingering build issue remains due to an unrelated mx format test.


LM Studio Discord

Gemini 1.5 JSON Woes: Engineers report that Gemini 1.5 flash struggles with JSON mode, causing intermittent issues with output. Users are invited to share insights or solutions to this challenge.

Tess Takes the Stage: The Tess 2.5 72b q3 and q4 quant models are now live on Hugging Face, offering new tools for experimentation.

AVX2 Instruction Essential: Users facing direct AVX2 errors should verify their CPU's support for AVX2 instructions to ensure compatibility with application requirements.

LM Studio Limitations and Solutions: LM Studio cannot be run on headless web servers or support safetensor files, but it succesfully employs GGUF format and Flash Attention can be enabled via alternatives like llama.cpp.

Hardware Market Fluctuations: There's a spike in the price of electronically scrapped P40 GPUs with current prices over $200, as well as a humorous note on sanctions possibly affecting Russian P40 stocks. A community member shares specs for an efficient home server build: R3700X, 128GB RAM, RTX 4090, and multiple storage options.


Eleuther Discord


LlamaIndex Discord

TiDB AI Experimentation on GitHub: PingCap demonstrates a RAG application using their TiDB database with LlamaIndex's knowledge graph, all available as open-source code with a demo and the source code on GitHub.

Paris AI Infrastructure Meetup Beckons: Engineers can join an AI Infrastructure Meetup at Station F in Paris featuring speakers from LlamaIndex, Gokoyeb, and Neon; details and sign-up are available here.

Vector Database Solutions for Quick Queries: For indexes containing 170,000 embeddings, use of Qdrant or FAISS Index is recommended; discussion includes fixing an AssertionError related to FAISS queries and direct node retrieval from a VectorStoreIndex with Chroma.

Adjacent Node Retrieval from Qdrant: A user inquiring about fetching adjacent nodes for law texts in a Qdrant vector store is advised to leverage node relationships and the latest API features for directional node retrieval.

Pushing LLM-Index Capabilities with PDF Embedding: An AI Engineer discusses embedding PDFs and documents into Weaviate using LLM-Index, demonstrating interest in expanding the ingestion of complex data types into vector databases.


Cohere Discord


LAION Discord


LangChain AI Discord


Modular (Mojo 🔥) Discord


Interconnects (Nathan Lambert) Discord


Latent Space Discord


OpenInterpreter Discord


OpenAccess AI Collective (axolotl) Discord


OpenRouter (Alex Atallah) Discord

Param Clamping in OpenRouter: Alex Atallah specified that parameters exceeding support, like Temp > 1, are clamped at 1 for OpenRouter, and parameters like Min P aren't passed through the UI, despite UI presentation suggesting otherwise.

Mistral 7B's Lag Time Mystery: Users noticed increased response times for Mistral 7B variants, attributing it to context length changes and potential rerouting, supported by data from an API watcher and a model uptime tracker.

Blockchain Developer on the Market: A senior full-stack & blockchain developer is on the lookout for new opportunities, showcasing experience in the field and eagerness to engage.

Vision for Vision Models: A request surfaced for the inclusion of more advanced vision models such as cogvlm2 in OpenRouter to enhance dataset captioning capabilities.


tinygrad (George Hotz) Discord


Datasette - LLM (@SimonW) Discord


DiscoResearch Discord


Torchtune Discord


MLOps @Chipro Discord


Mozilla AI Discord


YAIG (a16z Infra) Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI Stack Devs (Yoko Li) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Stability.ai (Stable Diffusion) ▷ #general-chat (854 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (446 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (190 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #announcements (8 messages🔥):

Links mentioned:


HuggingFace ▷ #general (185 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #cool-finds (10 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (11 messages🔥):

Links mentioned:

I'm excited to share my latest project that…": no description foundRelease v0.9.7 - stable diffusion 3 · bghira/SimpleTuner: Stable Diffusion 3 To use, set STABLE_DIFFUSION_3=true in your sdxl-env.sh and set your base model to stabilityai/stable-diffusion-3-medium-diffusers. What's Changed speed-up for training sample...GitHub - casualcomputer/llm_google_colab: A tutorial on how to set up a LLM on Google Colab for both GPU-accelerated and CPU-only session.: A tutorial on how to set up a LLM on Google Colab for both GPU-accelerated and CPU-only session. - casualcomputer/llm_google_colabGitHub - SimonThomine/CoursDeepLearning: Un regroupement de notebooks de apprendre le deep learning à partir de 0: Un regroupement de notebooks de apprendre le deep learning à partir de 0 - SimonThomine/CoursDeepLearningGitHub - piEsposito/tiny-ai-client: Tiny client for LLMs with vision and tool calling. As simple as it gets.: Tiny client for LLMs with vision and tool calling. As simple as it gets. - piEsposito/tiny-ai-clientCaptionEmporium/conceptual-captions-cc12m-llavanext · Datasets at Hugging Face: no description found


HuggingFace ▷ #reading-group (1 messages):

Link mentioned: MaPO Project Page: SOCIAL MEDIA DESCRIPTION TAG TAG


HuggingFace ▷ #computer-vision (28 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (1 messages):


HuggingFace ▷ #diffusion-discussions (79 messages🔥🔥):

Links mentioned:


OpenAI ▷ #ai-discussions (223 messages🔥🔥):

Link mentioned: DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs: Reading comprehension has recently seen rapid progress, with systems matching humans on the most popular datasets for the task. However, a large body of work has highlighted the brittleness of these s...


OpenAI ▷ #gpt-4-discussions (3 messages):


OpenAI ▷ #prompt-engineering (3 messages):

Links mentioned:


OpenAI ▷ #api-discussions (3 messages):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #general (15 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #🟩-modal (4 messages):


LLM Finetuning (Hamel + Dan) ▷ #learning-resources (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #hugging-face (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #langsmith (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #berryman_prompt_workshop (1 messages):

hamelh: humanloop, promptfoo are both popular


LLM Finetuning (Hamel + Dan) ▷ #clavie_beyond_ragbasics (4 messages):

Link mentioned: RizzConn Answering Machine - a Hugging Face Space by t0mkaka: no description found


LLM Finetuning (Hamel + Dan) ▷ #jason_improving_rag (2 messages):

Link mentioned: ParadeDB - Postgres for Search and Analytics: ParadeDB is a modern Elasticsearch alternative built on Postgres.


LLM Finetuning (Hamel + Dan) ▷ #saroufimxu_slaying_ooms (7 messages):


LLM Finetuning (Hamel + Dan) ▷ #paige_when_finetune (4 messages):


LLM Finetuning (Hamel + Dan) ▷ #gradio (2 messages):

Link mentioned: RizzConn Answering Machine - a Hugging Face Space by t0mkaka: no description found


LLM Finetuning (Hamel + Dan) ▷ #axolotl (9 messages🔥):

Link mentioned: Run Bash script in background and exit terminal: Is it possible to launch a command or Bash script exit terminal and NOT interrupt command? My solution was to run cron at a specific time of day, but I'm sure there is something easier.


LLM Finetuning (Hamel + Dan) ▷ #zach-accelerate (1 messages):

Link mentioned: From DeepSpeed to FSDP and Back Again with Hugging Face Accelerate: no description found


LLM Finetuning (Hamel + Dan) ▷ #wing-axolotl (3 messages):

Link mentioned: How to run axolotl on JarvisLabs | Tutorial: Check out axolotl on JarvisLabs : jarvislabs.ai/templates/axolotlCheck out axolotl github : https://github.com/OpenAccess-AI-Collective/axolotl


LLM Finetuning (Hamel + Dan) ▷ #simon_cli_llms (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #fireworks (4 messages):


LLM Finetuning (Hamel + Dan) ▷ #emmanuel_finetuning_dead (1 messages):

gitmaxd: great question


LLM Finetuning (Hamel + Dan) ▷ #europe-tz (3 messages):


LLM Finetuning (Hamel + Dan) ▷ #predibase (5 messages):

Link mentioned: Efficient and cost-effective multi-tenant LoRA serving with Amazon SageMaker | Amazon Web Services: In this post, we explore a solution that addresses these challenges head-on using LoRA serving with Amazon SageMaker. By using the new performance optimizations of LoRA techniques in SageMaker large m...


LLM Finetuning (Hamel + Dan) ▷ #openpipe (3 messages):


LLM Finetuning (Hamel + Dan) ▷ #openai (40 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #pawel-function-calling (92 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #interesting-links (5 messages):

Links mentioned:


Nous Research AI ▷ #general (152 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (2 messages):


Nous Research AI ▷ #rag-dataset (15 messages🔥):

Link mentioned: GitHub - VikParuchuri/marker: Convert PDF to markdown quickly with high accuracy: Convert PDF to markdown quickly with high accuracy - VikParuchuri/marker


Nous Research AI ▷ #world-sim (6 messages):


Perplexity AI ▷ #general (136 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (8 messages🔥):

Links mentioned:


Perplexity AI ▷ #pplx-api (14 messages🔥):


CUDA MODE ▷ #general (3 messages):


CUDA MODE ▷ #triton (2 messages):

Links mentioned:


CUDA MODE ▷ #torch (13 messages🔥):

Link mentioned: 8-Bit Approximations for Parallelism in Deep Learning: The creation of practical deep learning data-products often requires parallelization across processors and computers to make deep learning feasible on large data sets, but bottlenecks in communication...


CUDA MODE ▷ #algorithms (2 messages):

Link mentioned: GitHub - Aaronhuang-778/BiLLM: (ICML 2024) BiLLM: Pushing the Limit of Post-Training Quantization for LLMs: (ICML 2024) BiLLM: Pushing the Limit of Post-Training Quantization for LLMs - Aaronhuang-778/BiLLM


CUDA MODE ▷ #cool-links (1 messages):

_shivasinghbagri: https://powerinfer.ai/v2/


CUDA MODE ▷ #torchao (31 messages🔥):

Links mentioned:


CUDA MODE ▷ #llmdotc (54 messages🔥):

Links mentioned:


CUDA MODE ▷ #rocm (1 messages):

Link mentioned: AMD’s MI300X Outperforms NVIDIA’s H100 for LLM Inference: Discover if AMD's MI300X accelerator can outperform NVIDIA's H100 in real-world AI workloads. Early results are in!


CUDA MODE ▷ #bitnet (13 messages🔥):


LM Studio ▷ #💬-general (71 messages🔥🔥):

Link mentioned: GitHub - andrewyng/translation-agent: Contribute to andrewyng/translation-agent development by creating an account on GitHub.


LM Studio ▷ #🤖-models-discussion-chat (8 messages🔥):


LM Studio ▷ #🧠-feedback (2 messages):


LM Studio ▷ #🎛-hardware-discussion (9 messages🔥):


Eleuther ▷ #general (16 messages🔥):

Links mentioned:


Eleuther ▷ #research (52 messages🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (10 messages🔥):

Link mentioned: log_samples truthfulqa_gen - Pastebin.ai: no description found


Eleuther ▷ #multimodal-general (1 messages):

yash05880: icydk https://laion.ai/blog/open-flamingo/


LlamaIndex ▷ #blog (3 messages):

Links mentioned:


LlamaIndex ▷ #general (61 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):


Cohere ▷ #general (56 messages🔥🔥):

Links mentioned:


Cohere ▷ #project-sharing (3 messages):


LAION ▷ #general (41 messages🔥):

Links mentioned:


LAION ▷ #research (14 messages🔥):

Links mentioned:


LAION ▷ #resources (1 messages):

sidfeels: <@&825830190600683521>


LAION ▷ #learning-ml (1 messages):

.michu7: <@&825830190600683521>


LAION ▷ #paper-discussion (1 messages):

zuwop21: 50$ from steam steamcommunity.com/glft/918524 @everyone


LangChain AI ▷ #general (49 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (2 messages):

Link mentioned: Exécuter des LLMs en local avec Docker et Ollama: no description found


LangChain AI ▷ #tutorials (1 messages):

Link mentioned: GitHub - casualcomputer/llm_google_colab: A tutorial on how to set up a LLM on Google Colab for both GPU-accelerated and CPU-only session.: A tutorial on how to set up a LLM on Google Colab for both GPU-accelerated and CPU-only session. - casualcomputer/llm_google_colab


Modular (Mojo 🔥) ▷ #general (7 messages):

Link mentioned: Modular Community Livestream - New in MAX 24.4: MAX 24.4 is now available! Join us on our upcoming livestream as we discuss what’s new in MAX Engine and Mojo🔥 - MAX on macOS, MAX Engine Quantization API, ...


Modular (Mojo 🔥) ▷ #🔥mojo (31 messages🔥):

Link mentioned: nvim-lspconfig/doc/server_configurations.md at master · neovim/nvim-lspconfig: Quickstart configs for Nvim LSP. Contribute to neovim/nvim-lspconfig development by creating an account on GitHub.


Modular (Mojo 🔥) ▷ #nightly (13 messages🔥):


Interconnects (Nathan Lambert) ▷ #news (33 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (4 messages):

Link mentioned: Tweet from Cosmin Negruseri (@cosminnegruseri): haven't seen this in a paper before


Interconnects (Nathan Lambert) ▷ #random (6 messages):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #nlp (7 messages):

Link mentioned: Tweet from Liliang Ren (@liliang_ren): Introducing Samba 3.8B, a simple Mamba+Sliding Window Attention architecture that outperforms Phi3-mini on major benchmarks (e.g., MMLU, GSM8K and HumanEval) by a large margin.😮 And it has an infinit...


Latent Space ▷ #ai-general-chat (43 messages🔥):

Links mentioned:


OpenInterpreter ▷ #general (40 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #general (5 messages):

Link mentioned: Introducing Apple’s On-Device and Server Foundation Models: At the 2024 Worldwide Developers Conference, we introduced Apple Intelligence, a personal intelligence system integrated deeply into…


OpenAccess AI Collective (axolotl) ▷ #general-help (16 messages🔥):

Link mentioned: Installing the NVIDIA Container Toolkit — NVIDIA Container Toolkit 1.15.0 documentation: no description found


OpenRouter (Alex Atallah) ▷ #general (16 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #general (3 messages):

Link mentioned: RDNA3 assembly support by geohot · Pull Request #3637 · tinygrad/tinygrad: no description found


tinygrad (George Hotz) ▷ #learn-tinygrad (10 messages🔥):


Datasette - LLM (@SimonW) ▷ #ai (6 messages):

Link mentioned: Sober AI is the Norm: Sober AI is the quiet default, despite all the hype you hear about human-replacements and AGI. Data scientists and engineers are quietly transforming business intelligence through practical applicatio...


DiscoResearch ▷ #general (3 messages):

Link mentioned: Aleph Alpha and Silo AI enter a strategic partnership to advance open source AI and enterprise-grade solutions in Europe - ALEPH ALPHA - AI for Enterprises and Governments: To foster the adoption and fully leverage the potential of generative AI across European industrial firms, Europe’s largest AI lab Silo AI and European AI champion Aleph Alpha are announcing a long-te...


Torchtune ▷ #general (1 messages):


Torchtune ▷ #dev (1 messages):

Link mentioned: Build software better, together: GitHub is where people build software. More than 100 million people use GitHub to discover, fork, and contribute to over 420 million projects.


MLOps @Chipro ▷ #events (1 messages):

Link mentioned: Infer Summer ‘24 by Qwak | The Engineering Behind AI and ML: Infer Summer ‘24 by Qwak brings AI leaders to share how the world’s leading companies use ML and AI in production. Join live on Jun 26, 2024, 11:00 AM EDT


Mozilla AI ▷ #announcements (1 messages):


YAIG (a16z Infra) ▷ #tech-discussion (1 messages):




{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}