Frozen AI News archive

Gemini launches context caching... or does it?

**Nvidia's Nemotron** ranks #1 open model on LMsys and #11 overall, surpassing **Llama-3-70b**. **Meta AI** released **Chameleon 7B/34B** models after further post-training. **Google's Gemini** introduced context caching, offering a cost-efficient middle ground between RAG and finetuning, with a minimum input token count of 33k and no upper limit on cache duration. **DeepSeek** launched **DeepSeek-Coder-V2**, a 236B parameter model outperforming **GPT-4 Turbo**, **Claude-3-Opus**, and **Gemini-1.5-Pro** in coding tasks, supporting 338 programming languages and extending context length to 128K. It was trained on 6 trillion tokens using the **Group Relative Policy Optimization (GRPO)** algorithm and is available on Hugging Face with a commercial license. These developments highlight advances in model performance, context caching, and large-scale coding models.

Canonical issue URL

AI News for 6/17/2024-6/18/2024. We checked 7 subreddits, 384 Twitters and 30 Discords (415 channels, and 3582 messages) for you. Estimated reading time saved (at 200wpm): 397 minutes. You can now tag @smol_ai for AINews discussions!

Today was a great day for AINews followups:

But for AI Engineers, today's biggest news has to be the release of Gemini's context caching, first teased at Google I/O (our report here).

image.png

Caching is exciting because it creates a practical middle point between the endless RAG vs Finetuning debate - instead of using a potentially flawed RAG system, or lossfully finetuning a LLM to maaaaybe memorize new facts... you just allow the full magic of attention to run on the long context and but pay 25% of the cost (but you do pay $1 per million tokens per hour storage which is presumably a markup over the raw storage... making the breakeven about the 400k tokens/hr mark):

image.png

Some surprises:

We first discussed context caching with Aman Sanger on the Neurips 2023 podcast and it was assumed the difficulty was the latency/cost efficiency around loading/unloading caches per request. However the bigger challenge to using this may be the need for prompt prefixes to be dynamically constructed per request (this issue only applies to prefixes, dynamic suffixes can work neatly with cached contexts).


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku. DeepSeek-Coder-V2 Model Release

Meta AI Model Releases

Runway Gen-3 Alpha Video Model

NVIDIA Nemotron-4-340B Model

Anthropic AI Research on Reward Tampering


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

Video Generation AI Models and Capabilities

Image Generation AI Models

AI in Healthcare

AI Replacing Jobs

Robotics and Embodied AI

Humor/Memes


AI Discord Recap

A summary of Summaries of Summaries

  1. DeepMind Brings Soundtracks to AI Videos:

    • Google DeepMind's V2A technology can generate unlimited audio tracks for AI-generated videos, addressing the limitation of silent AI videos.
    • ElevenLabs launched a sound effects generator with infinite customization, promising high-quality, royalty-free audio for various media applications.
  2. Stable Diffusion 3 Faces Licensing Drama:

    • [Civitai temporarily] banned all SD3-based models](https://civitai.com/articles/5732) due to unclear licensing terms, triggering community concerns about Stability AI's control over models.
    • SD3's release was met with disappointment, labeled as the "worst base model release yet" due to both performance issues and licensing uncertainties.
  3. Exceeding Expectations with Model Optimizations:

    • The CUTLASS library outperformed CuBLAS by 10% in pure C++ for matrix multiplications but lost this edge when integrated with Python, both touching 257 Teraflops.
    • Meta introduces Chameleon, a model supporting mixed-modal inputs with promising benchmarks and open-source availability, alongside other innovative models like JASCO.
  4. AI Community Questions OpenAI Leadership:

    • Concerns arose around OpenAI's appointment of a former NSA director, with Edward Snowden's tweet cautioning against potential data security risks associated with this decision.
    • Widespread ChatGPT downtimes left users frustrated, highlighting server stability issues across different regions and pushing users to seek alternatives like the ChatGPT app.
  5. Training and Compatibility Issues Across Platforms:

    • Google Colab struggles with session interruptions during model training led to discussions about workaround tips like preemptive checkpointing.
    • AMD GPU support in Axolotl remains inadequate, especially for the MI300X, prompting users to exchange modification tips for enhanced compatibility and performance.

PART 1: High level Discord summaries

Unsloth AI (Daniel Han) Discord


CUDA MODE Discord


Stability.ai (Stable Diffusion) Discord


HuggingFace Discord


OpenAI Discord


Modular (Mojo 🔥) Discord


Cohere Discord


LM Studio Discord

Heed the Setup Cautions with New Models: While setting up the Deepseek Coder V2 Lite, users should pay close attention to certain settings that are critical during the initial configuration, as one setting incorrectly left on could cause issues.

When Autoupdate Fails, DIY: LM Studio users have encountered broken autoupdates since version 0.2.22, necessitating manual download of newer versions. Links for downloading version 0.2.24 are functioning, but issues have been reported with version 0.2.25.

Quantization's Quandary: There's a notable variability in model responses based on different quantization levels. Users found Q8 to be more responsive compared to Q4, and these differences are important when considering model efficiency and output suitability.

Config Chaos Demands Precision: One user struggled with configuring the afrideva/Phi-3-Context-Obedient-RAG-GGUF model, triggering advice on specific system message formatting. This discussion emphasizes the importance of precise prompt structuring for optimal bot interaction.

Open Interpreter Troubleshooting: Issues regarding Open Interpreter defaulting to GPT-4 instead of LM Studio models led to community-shared workarounds for MacOS and references to a YouTube tutorial for detailed setup guidance.


Nous Research AI Discord


Interconnects (Nathan Lambert) Discord


Perplexity AI Discord


OpenAccess AI Collective (axolotl) Discord


LlamaIndex Discord


LLM Finetuning (Hamel + Dan) Discord

Note: Links and specific numerical details were embedded when available for reference.


OpenRouter (Alex Atallah) Discord


Eleuther Discord


LangChain AI Discord

LangChain Learners Face Tutorial Troubles: Members experienced mismatch issues between LangChain versions and published tutorials, with one user getting stuck at a timestamp in a ChatGPT Slack bot video. Changes like the deprecation of LLMChain in LangChain 0.1.17 and the upcoming removal in 0.3.0 highlight the rapid evolution of the library.

Extracting Gold from Web Scrapes & Debugging Tips: A user was guided on company summary and client list extraction from website data using LangChain, and others discussed debugging LangChain's LCEL pipelines with set_debug(True) and set_verbose(True). Frustration arose from BadRequestError in APIs, reflecting challenges in handling unexpected API behavior.

Serverless Searches & Semantic AI Launches: An article on creating a serverless semantic search with AWS Lambda and Qdrant was shared, alongside the launch of AgentForge on ProductHunt, integrating LangChain, LangGraph, and LangSmith. Another work, YouSim, showcased a backrooms-inspired simulation platform for identity experimentation.

New Mediums, New Codes: jasonzhou1993 explored AI's impact on music creation in a YouTube tutorial, while also sharing a Hostinger website builder discount code AIJASON.

Calls for Collaboration and Sharing Innovations: A plea for beta testers surfaced for an advanced research assistant at Rubik's AI, mentioning premium features like Claude 3 Opus and GPT-4 Turbo. Hugging Face's advice to sequester environment setup from code and the embrace of tools like Bitwarden for managing credentials stressed importance of secure and clean development practices.


tinygrad (George Hotz) Discord


LAION Discord


Torchtune Discord

CUDA vs MPS: Beware the NaN Invasion: Engineers discussed an issue where nan outputs appeared on CUDA but not on MPS, tied to differences in kernel execution paths for softmax operations in SDPA, leading to softmax causing nan on large values.

Cache Clash with Huggingface: There were discussions on system crashes during fine-tuning with Torchtune due to Huggingface's cache overflowing, causing concern and a call for solutions among users.

Constructing Bridge from Huggingface to Torchtune: The guild shared a detailed process for converting Huggingface models to Torchtune format, highlighting Torchtune Checkpointers for easy weight conversion and loading.

The Attention Mask Matrix Conundrum: Clarification on the proper attention mask format for padded token inputs to avoid disparity across processing units was debated, ensuring that the model's focus is correctly applied.

Documentation to Defeat Disarray: Links to Torchtune documentation, including RLHF with PPO and GitHub pull requests, were shared to assist with implementation details and facilitate knowledge sharing among engineers. RLHF with PPO | Torchtune Pull Request


Latent Space Discord


OpenInterpreter Discord


AI Stack Devs (Yoko Li) Discord


Datasette - LLM (@SimonW) Discord


Mozilla AI Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

Unsloth AI (Daniel Han) ▷ #general (526 messages🔥🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (10 messages🔥):


Unsloth AI (Daniel Han) ▷ #help (143 messages🔥🔥):

<ul>
    <li><strong>Colab Training Sessions Woes</strong>: One user experienced issues with their Google Colab training session for Unsloth cutting out at 90% after 23 hours. They expressed frustration and received advice about preemptively enabling checkpointing within TrainingArguments() to avoid future occurrences.</li>
    <li><strong>Fine-Tuning LLMs Issues</strong>: Users gabrielsandstedt and shensmobile discussed problems related to fine-tuning large language models (LLMs) on Google Colab. The importance of enabling checkpointing and limitations of session lengths were highlighted.</li>
    <li><strong>Tokenizing Troubles</strong>: A member wanted to compare vocab before and after fine-tuning an LLM but faced storage limits on free Google Colab. Discussion revolved around the necessity of saving the tokenizer along with the model and possible space-saving methods.</li>
    <li><strong>Dataset Formatting and Schema</strong>: Thefanciestpeanut guided gbourdin on how to convert JSON to Parquet for better training efficiency in Unsloth, emphasizing mapping the data correctly for fine-tuning. They shared a detailed code snippet for dataset conversion and loading in Python.</li>
    <li><strong>Mixed GPU Usage Obstacles</strong>: Several users, including karatsubabutslower and origamidream, deliberated on challenges encountered when using multiple GPUs with Unsloth, suggesting using older versions or setting environment variables properly to circumvent usage restrictions.</li>
</ul>

Links mentioned:


CUDA MODE ▷ #general (5 messages):

Link mentioned: 100k H100 Clusters: Power, Network Topology, Ethernet vs InfiniBand, Reliability, Failures, Checkpointing: Frontier Model Scaling Challenges and Requirements, Fault Recovery through Memory Reconstruction, Rack Layouts


CUDA MODE ▷ #algorithms (2 messages):


CUDA MODE ▷ #jobs (1 messages):

niceboy2989: <@848720848282189855> I can help you


CUDA MODE ▷ #jax (1 messages):

Link mentioned: GitHub - yixiaoer/tpux: A set of Python scripts that makes your experience on TPU better: A set of Python scripts that makes your experience on TPU better - yixiaoer/tpux


CUDA MODE ▷ #torchao (25 messages🔥):

Links mentioned:


CUDA MODE ▷ #llmdotc (536 messages🔥🔥🔥):

Links mentioned:


CUDA MODE ▷ #bitnet (9 messages🔥):

Link mentioned: Issues · pytorch/ao: PyTorch dtype and layout library. 30% speedups for training. 2x speedups and 65% less VRAM for inference. Composability with FSDP and torch.compile. - Issues · pytorch/ao


CUDA MODE ▷ #sparsity (3 messages):

Link mentioned: Strangely, Matrix Multiplications on GPUs Run Faster When Given "Predictable" Data! [short]: Great minds discuss flops per watt.


Stability.ai (Stable Diffusion) ▷ #general-chat (363 messages🔥🔥):

- **Civitai bans SD3 content**: Civitai has temporarily banned all SD3 related content due to concerns about the license's clarity, as shared by a user *“due to a lack of clarity in the license associated with Stable Diffusion 3, we are temporarily banning all SD3 based models.”* ([Civitai Announcement](https://civitai.com/articles/5732)).
- **Community dissatisfaction with SD3 release**: Multiple users expressed disappointment with the SD3 model, describing it as *“the worst base model release yet.”* Complaints were directed at both the performance and licensing issues.
- **SD3 Performance and Alternatives**: Users discussed the architecture and potential of SD3, noting its *“16ch VAE allows better text understanding”*, yet also acknowledging that other models like Pixart and Lumina can do *“more with less compute.”*
- **License concerns and legal implications**: There's significant worry in the community about how the SD3 model's license might allow Stability AI *“too much power over the models.”* This has caused platforms like Civitai to seek legal clarity before allowing SD3 content.
- **Comparisons with other tools**: Discussions often referenced alternate tools and software, with one user stating *“I swapped to Pixart Sigma...prompt adherence is good but has issues with limbs.”* Other users recommended different models and interfaces for various use cases including StableSwarmUI and ComfyUI.

Links mentioned:


HuggingFace ▷ #general (311 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (8 messages🔥):

Links mentioned:


HuggingFace ▷ #diffusion-discussions (4 messages):


OpenAI ▷ #ai-discussions (187 messages🔥🔥):


OpenAI ▷ #gpt-4-discussions (17 messages🔥):


OpenAI ▷ #prompt-engineering (19 messages🔥):


OpenAI ▷ #api-discussions (19 messages🔥):


Modular (Mojo 🔥) ▷ #general (91 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (3 messages):


Modular (Mojo 🔥) ▷ #✍︱blog (1 messages):

Link mentioned: Modular: What’s New in Mojo 24.4? Improved collections, new traits, os module features and core language enhancements: We are building a next-generation AI developer platform for the world. Check out our latest post: What’s New in Mojo 24.4? Improved collections, new traits, os module features and core language enhanc...


Modular (Mojo 🔥) ▷ #🔥mojo (108 messages🔥🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #nightly (1 messages):

helehex: pollinate mojo buzz buzz


Cohere ▷ #general (150 messages🔥🔥):


Cohere ▷ #project-sharing (5 messages):


LM Studio ▷ #💬-general (55 messages🔥🔥):

<ul>
  <li><strong>Deepseek Coder V2 Lite requires caution during setup</strong>: Users discussed the importance of certain settings when loading the new Deepseek Coder V2 Lite model. One noted, *"make sure this is turned off"*, referring to a specific setting in the model setup.</li>
  <li><strong>LM Studio and Open Interpreter guidelines</strong>: A step-by-step guide was shared for using LM Studio with Open Interpreter, referencing the need to run LM Studio in the background. The guide can be found on the official <a href="https://docs.openinterpreter.com/language-models/local-models/lm-studio">Open Interpreter documentation</a>.</li>
  <li><strong>Help requests for local model loading issues</strong>: Users reported issues loading models on LM Studio, with one sharing system specs and receiving advice to try different settings and models. Model loading issues, particularly with smaller VRAM capacity, were discussed.</li>
  <li><strong>Using AMD cards with LM Studio</strong>: Discussion around using AMD GPUs for AI, noting that OpenCL is required and performance may be suboptimal. A link to OpenCL instructions was shared from the <a href="https://github.com/lmstudio-ai/configs/blob/main/Extension-Pack-Instructions.md">LM Studio Configs GitHub</a>.</li>
  <li><strong>Meta's new AI models announcement</strong>: Meta announced several new AI models including Meta Chameleon and Meta JASCO. Users were directed to more details on <a href="https://go.fb.me/tzzvfg">Facebook's official announcement</a> and the <a href="https://github.com/facebookresearch/chameleon">GitHub repository for Meta Chameleon</a>.</li>
</ul>

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (49 messages🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (3 messages):


LM Studio ▷ #📝-prompts-discussion-chat (1 messages):


LM Studio ▷ #⚙-configs-discussion (9 messages🔥):

Link mentioned: no title found: no description found


LM Studio ▷ #🎛-hardware-discussion (11 messages🔥):


LM Studio ▷ #🧪-beta-releases-chat (6 messages):

Link mentioned: no title found: no description found


LM Studio ▷ #open-interpreter (13 messages🔥):

Link mentioned: ChatGPT "Code Interpreter" But 100% Open-Source (Open Interpreter Tutorial): This is my second video about Open Interpreter, with many new features and much more stability, the new Open Interpreter is amazing. Update: Mixtral 7x8b was...


LM Studio ▷ #🛠-dev-chat (5 messages):

Link mentioned: GitHub - lmstudio-ai/lmstudio.js: LM Studio TypeScript SDK: LM Studio TypeScript SDK. Contribute to lmstudio-ai/lmstudio.js development by creating an account on GitHub.


Nous Research AI ▷ #off-topic (1 messages):

<ul>
    <li><strong>Chaotic music not a favorite</strong>: One member listened to some music and commented, "I can safely say that's not quite my preferred music XD. Very chaotic."</li>
</ul>

Nous Research AI ▷ #interesting-links (10 messages🔥):

Link mentioned: YouSim: they've simulated websites, worlds, and imaginary CLIs... but what if they simulated you?


Nous Research AI ▷ #general (105 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (19 messages🔥):

Link mentioned: GitHub - FlowiseAI/Flowise: Drag & drop UI to build your customized LLM flow: Drag & drop UI to build your customized LLM flow. Contribute to FlowiseAI/Flowise development by creating an account on GitHub.


Nous Research AI ▷ #world-sim (8 messages🔥):


Interconnects (Nathan Lambert) ▷ #news (58 messages🔥🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (4 messages):

Link mentioned: Tweet from Bill Yuchen Lin 🤖 (@billyuchenlin): M-A-P/Neo-7B-Instruct is the 1st 💎fully-open💎 LLM on WildBench leaderboard and its performance is awesome. "Fully open-source" here means that all data for pre-training & post-training are ...


Interconnects (Nathan Lambert) ▷ #random (71 messages🔥🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #rlhf (6 messages):

Link mentioned: Creativity Has Left the Chat: The Price of Debiasing Language Models: Large Language Models (LLMs) have revolutionized natural language processing but can exhibit biases and may generate toxic content. While alignment techniques like Reinforcement Learning from Human Fe...


Interconnects (Nathan Lambert) ▷ #posts (2 messages):

- **SnailBot summons the crew**: SnailBot issued a call to the community with the tag <@&1216534966205284433>. 
- **Nathan Lambert celebrates SnailBot**: Nathan Lambert adds a cute and playful touch with *"🐌 🐌 🐌 🐌"* emojis, showing affection or enthusiasm for SnailBot.

Perplexity AI ▷ #general (99 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (10 messages🔥):

Link mentioned: YouTube: no description found


Perplexity AI ▷ #pplx-api (19 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #general (59 messages🔥🔥):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general-help (4 messages):


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (6 messages):

For references and more detailed steps, see the full post on Phorm.ai.

Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (5 messages):

Links mentioned:


LlamaIndex ▷ #blog (2 messages):


LlamaIndex ▷ #general (62 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #general (18 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #🟩-modal (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #replicate (1 messages):

strickvl: When do Replicate credits expire?


LLM Finetuning (Hamel + Dan) ▷ #langsmith (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #clavie_beyond_ragbasics (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #jason_improving_rag (1 messages):

hammadkhan: https://x.com/xhluca/status/1803100958408241597?s=46&t=-TRJUfVdW8KeDqen1HJU1Q


LLM Finetuning (Hamel + Dan) ▷ #credits-questions (21 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #fireworks (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #braintrust (4 messages):


LLM Finetuning (Hamel + Dan) ▷ #west-coast-usa (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #predibase (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #openpipe (1 messages):

strickvl: When do OpenPipe credits expire?


LLM Finetuning (Hamel + Dan) ▷ #openai (1 messages):

sph3r3ical: yeah, where do you see the credits?


LLM Finetuning (Hamel + Dan) ▷ #bergum_rag (7 messages):


OpenRouter (Alex Atallah) ▷ #announcements (1 messages):


OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):

mka79: Is it from OR team?


OpenRouter (Alex Atallah) ▷ #general (60 messages🔥🔥):

Links mentioned:


OpenRouter (Alex Atallah) ▷ #일반 (1 messages):

sigridjin.eth: 와 안녕하세요.


Eleuther ▷ #general (29 messages🔥):

Links mentioned:


Eleuther ▷ #research (20 messages🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (10 messages🔥):

Links mentioned:


Eleuther ▷ #lm-thunderdome (2 messages):


LangChain AI ▷ #general (17 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (14 messages🔥):

Links mentioned:


LangChain AI ▷ #tutorials (1 messages):

Link mentioned: Did AI just end music?!: Music Gen 101 & build application with Text-to-Music APIHostinger website builder: https://www.hostinger.com/aijasonGet 10% off with my code: AIJASON🔗 Links...


tinygrad (George Hotz) ▷ #general (9 messages🔥):

Links mentioned:


tinygrad (George Hotz) ▷ #learn-tinygrad (17 messages🔥):


LAION ▷ #general (23 messages🔥):

- **RunwayML Gen-3 clip amazes users**: Members were impressed by a [RunwayML Gen-3 clip](https://fxtwitter.com/Mr_AllenT/status/1802706451586023763), calling its AI-generated details "insane". One noted, "99% of people wouldn't know this is AI."
- **DeepMind shares video-to-audio research**: A blog post on DeepMind's V2A technology was shared, explaining how video pixels and text prompts can generate soundtracks for videos. This could innovate in creating sound for silent footage and working with models like [Veo](https://deepmind.google/technologies/veo/).
- **Meta FAIR releases new research artifacts**: Meta FAIR announced several new [research artifacts](https://ai.meta.com/blog/meta-fair-research-new-releases/), including Meta Llama 3 and V-JEPA, emphasizing their commitment to open AI ecosystems. Another user was interested in the recently released Chameleon vision-only weights.
- **PKU-YuanGroup's Open-Sora Plan**: A member shared a [GitHub link](https://github.com/PKU-YuanGroup/Open-Sora-Plan) about the Open-Sora Plan, a project aimed at reproducing the Open AI T2V model. They requested community contributions to this open-source endeavor.
- **Free img2img model request**: A user expressed a need for a free img2img model using RealVision or similar, aiming to add "a touch of realism." They reminisced about potentially using their old custom Stable 2 model for this purpose.

Links mentioned:


LAION ▷ #research (3 messages):

Link mentioned: weights2weights: no description found


Torchtune ▷ #general (24 messages🔥):

Links mentioned:


Latent Space ▷ #ai-general-chat (18 messages🔥):

Links mentioned:


OpenInterpreter ▷ #general (9 messages🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (1 messages):

legaltext.ai: the one from april?


OpenInterpreter ▷ #ai-content (2 messages):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #ai-town-discuss (3 messages):

Link mentioned: Agent Hospital: A Simulacrum of Hospital with Evolvable Medical Agents: In this paper, we introduce a simulacrum of hospital called Agent Hospital that simulates the entire process of treating illness. All patients, nurses, and doctors are autonomous agents powered by lar...


Datasette - LLM (@SimonW) ▷ #ai (1 messages):

shajith: oh that is good, thanks for sharing.


Datasette - LLM (@SimonW) ▷ #llm (2 messages):

Link mentioned: Language models on the command-line: I gave a talk about accessing Large Language Models from the command-line last week as part of the Mastering LLMs: A Conference For Developers & Data Scientists six week long …


Mozilla AI ▷ #llamafile (1 messages):






{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}