Frozen AI News archive

Ways to use Anthropic's Tool Use GA

**Anthropic** launched general availability of tool use/function calling with support for streaming, forced use, and vision, alongside **Amazon** and **Google**. Alex Albert shared five architectures for agentic tool use: delegation, parallelization, debate, specialization, and tool suite experts. **Anthropic** also introduced a self-guided course on tool use. **Yann LeCun** emphasized ethical open science funding, gradual emergence of superintelligence with safety guardrails, and convolutional networks for image/video processing as competitive with vision transformers. He also noted growth in AI researchers across industry, academia, and government.

Canonical issue URL

AI News for 5/30/2024-5/31/2024. We checked 7 subreddits, 384 Twitters and 29 Discords (393 channels, and 2911 messages) for you. Estimated reading time saved (at 200wpm): 337 minutes.

Together with Anthropic's GA of tool use/function calling today on Anthropic/Amazon/Google, with support for streaming, forced use, and vision...

image.png

Alex Albert shared 5 architectures for using them in an agentic context:

  1. Delegation: Use cheaper, faster models for cost and speed gains.
  1. Parallelization: Cut latency (but not cost) by running agents in parallel.
  1. Debate: Multiple agents with different roles engage in discussion to reach better decisions.
  1. Specialization: A generalist agent orchestrates, while specialists execute tasks.
  1. Tool Suite Experts: When using 100s or 1000s of tools, specialize agents in tool subsets.

Nothing particularly groundbreaking here but a very handy list to think about for patterns. Anthropic also launched a self guided course on tool use:

image.png


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

AI Research and Development

AI Tools and Applications

Memes and Humor


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

AI Image & Video Generation

AI Ethics & Societal Impact

AI Capabilities & Advancements

OpenAI News & Developments

AI Humor & Memes


AI Discord Recap

A summary of Summaries of Summaries

1. Model Performance Optimization and Benchmarking

2. Fine-Tuning and Prompt Engineering

3. Open-Source AI Developments and Collaborations

4. AI Community Innovations and Knowledge Sharing

5. Hardware Advancements and Compatibility Challenges


PART 1: High level Discord summaries

LLM Finetuning (Hamel + Dan) Discord

These summaries encapsulate the detailed, often granular discussions among AI Engineers in the Discord guild, highlighting the collective endeavor to optimize LLM fine-tuning and deployment amidst pursuit of career growth and community building.


HuggingFace Discord

K2 Triumphs Over Llama 2: LLM360's K2 model outpaces Llama 2 70B, achieving better performance with 35% less computational effort; it's touted as fully-reproducible and is accessible under the Apache 2.0 license.

Numbers Are No Match for Positional Embeddings: Researchers cracked the nut on transformers' arithmetic abilities; with tailored positional embeddings, transformers reach a 99% accuracy on 100-digit sums, a monumental feat outlined in their paper.

NeurIPS Throws Down the Merging Gauntlet: With an $8,000 purse, the NeurIPS Model Merging Competition invites contenders to blend optimal AI models. Hugging Face, among others, sponsors this competition, more info in the announcement and competition website.

Data Dive: From 150K Datasets to Clothing Sales: A treasure trove of 150k+ datasets is now at engineers' fingertips for exploration with DuckDB, explained in a blog post, while a novel clothing sales dataset propelled the development of an image regression model which was then detailed in this article.

Learning Resources and Courses Amplify Skills: In the perpetually advancing field of AI, engineers can bolster their expertise through Hugging Face courses in Reinforcement Learning and Computer Vision, with more information accessible at Hugging Face - Learn.


Unsloth AI (Daniel Han) Discord

Quantization Quandaries and High-Efficiency Hardware: Unsloth AI guild members highlight challenges with the quantized Phi3 finetune results, noting performance issues without quantization tricks. NVIDIA's new 4nm research chip is generating buzz with its 96 int4 tera operations per second per watt (TOPs/Watt) efficiency, overshadowing Blackwell's 20T/W and reflecting industry-wide advancements in power efficiency, numerical representation, Tensor Cores' efficiency, and sparsity techniques.

Model Fine-Tuning and Upscaling Discussions: AI engineers share insights on fine-tuning strategies, including dataset merging, with one member unveiling an 11.5B upscale model of Llama-3 using upscaling techniques. An emerging fine-tuning method, MoRA, suggests a promising avenue for parameter-efficient updates.

Troubleshooting Tools and Techniques: Engineers confront various hurdles, from GPU selection in Unsloth (os.environ["CUDA_VISIBLE_DEVICES"]="0") and troubleshooting fine-tuning errors to handling dual-model dependencies and addressing VRAM spikes during training. Workarounds for issues like Kaggle installation challenges underscore the need for meticulous problem-solving.

AI in Multiple Tongues: Ghost XB Beta garners attention for its capability to support 9+ languages fluently and is currently navigating through its training stages. This progress reaffirms the guild’s commitment to developing accessible, cost-efficient AI tools for the community, especially emphasizing startup support.

Communal Cooperative Efforts and Enhancements: Guild discussions reveal a collective push for self-deployment and community backing, with members sharing updates and seeking assistance across a spectrum of AI-related endeavors such as the Open Empathic project and Unsloth AI model improvements.


Perplexity AI Discord


CUDA MODE Discord


Stability.ai (Stable Diffusion) Discord


LM Studio Discord

GPU Blues with ROCm? Not Music to Our Ears: Engineers discussed GPU performance with ROCm, lamenting the lack of support for RX 6700 and old AMD GPUs like RX580, influencing token generation speeds and overall performance. Users seeking performance benchmarks on multi-GPU systems with models such as LLAMA 3 8B Q8 reported a 91% efficiency with two GPUs compared to one.

VRAM Envy: The release of LM Studio models ignited debates on VRAM adequacy, where the 4070's 12GB was compared unfavorably to the 1070's 20GB, especially concerning suitability for large models like "codestral."

CPU Constraints Cramp Styles: CPU requirements for running LM Studio became a focal point, where AVX2 instructions proved mandatory, leading users with older CPUs to use a prior version (0.2.10) for AVX instead.

Routing to the Right Template: AI engineers shared solutions and suggestions for model templates, such as using Deepseek coder prompt template for certain models, and advised checking tokenizer configurations for optimal formatting with models like TheBloke/llama2_7b_chat_uncensored-GGUF.

New Kids on the Block - InternLM Models: Several InternLM models designed for Math and Coding, ranging from 7B to a mixtral 8x22B, were announced. Models such as AlchemistCoder-DS-6.7B-GGUF and internlm2-math-plus-mixtral8x22b-GGUF were highlighted among the latest tools available for AI engineers.


OpenRouter (Alex Atallah) Discord


OpenAI Discord

Pro Privileges Propel Chat Productivity: Pro users of OpenAI now enjoy enhanced capabilities such as higher rate limits, and exclusive GPT creation, along with access to DALL-E and real-time communication features. The alluring proposition maintains its charm despite the $20 monthly cost, marking a clear divide from the limited toolkit available to non-paying users.

AI Framework Favorites Facilitate Functional Flexibility: The Chat API is recommended over the Assistant API for those developing AI personas with idiosyncratic traits, as it offers superior command execution without surplus functionalities such as file searching.

Bias Brouhaha Besieges ChatGPT: A suspension due to calling out perceived racism in ChatGPT's outputs opened up a forum of contention around inherent model biases, spotlighting the relentless pursuit of attenuating such biases amidst the ingrained nuances of training data.

Virtual Video Ventures Verified: Sora and Veo stand as subjects of a speculative spree as the guild contemplates the curated claims and practical potency of the pioneering video generation models, juxtaposed against the realities of AI-assisted video crafting.

API Agitations and Advancements Announced: Persistent problems presented by memory leaks causing lag and browser breakdowns mar the ChatGPT experience, triggering talks on tactical chat session limits and total recall of past interactions to dodge the dreariness of repetition. Meanwhile, the anticipated arrival of real-time voice and visual features in GPT-4 has been slated to debut in an Alpha state for a select circle, broadening over subsequent months as per OpenAI's update.


Nous Research AI Discord

NeurIPS Competition: Merge Models for Glory and Cash: NeurIPS will host a Model Merging competition with an $8K prize, sponsored by Hugging Face and Sakana AI Labs—seeking innovations in model selection and merging. Registration and more info can be found at llm-merging.github.io as announced on Twitter.

AI's Quest to Converse with Critters: A striking $500K Coller Prize is up for grabs for those who can demystify communication with animals using AI, sparking excitement for potential breakthroughs (info). This initiative echoes Aza Raskin's Earth Species Project, aiming to untangle interspecies dialogue (YouTube video).

Puzzling Over Preference Learning Paradox: The community is abuzz after a tweet highlighted unexpected limitations in RLHF/DPO methods—preference learning algorithms are not consistently yielding better ranking of preferred responses, challenging conventional wisdom and suggesting a potential for overfitting.

LLMs Reigning Over Real-Time Web Content: A revelation for web users: LLMs are often churning out web pages in real-time, rendering what you see as it loads. This routine faces hiccups with lengthy or substantial pages due to context constraints, an area ripe for strategic improvements.

Google Enhances AI-Driven Search: Google has upgraded its AI Overviews for US search users, improving both satisfaction and webpage click quality. Despite some glitches, they're iterating with a feedback loop, detailed in their blog post – AI Overviews: About last week.


LlamaIndex Discord


Eleuther Discord


Modular (Mojo 🔥) Discord


LangChain AI Discord


LAION Discord


OpenAccess AI Collective (axolotl) Discord


DiscoResearch Discord


Interconnects (Nathan Lambert) Discord


Cohere Discord


Latent Space Discord

Adapter Layers Bridge the Gap: Engineers are exploring embedding adapters as a means to improve retrieval performance in AI models, with evidence showcased in a Chroma research report. The effectiveness of these can be likened to Froze Embeddings, which the Vespa team employs to eliminate frequent updates in dynamic systems (Vespa's blog insights).

ChatGPT Goes Corporate with PwC: The acquisition of ChatGPT Enterprise licenses by PwC for roughly 100,000 employees sparked debates around the estimated value of $30M/year, with member guesses on the cost per user ranging from $8 to $65 per month.

Google's Twin Stars: Gemini 1.5 Flash & Pro: Release updates for Google Gemini 1.5 Flash and Pro have been pushed to general availability, introducing enhancements such as increased RPM limits and JSON Schema mode (Google developers blog post).

TLBrowse Joins the Open Source Universe: TLBrowse, melding Websim with TLDraw, was open-sourced, allowing users to conjure up infinite imagined websites on @tldraw canvas, with access to a free hosted version.


AI Stack Devs (Yoko Li) Discord


OpenInterpreter Discord


Mozilla AI Discord


MLOps @Chipro Discord


Datasette - LLM (@SimonW) Discord


tinygrad (George Hotz) Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

{% if medium == 'web' %}

LLM Finetuning (Hamel + Dan) ▷ #general (86 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #workshop-1 (10 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #asia-tz (1 messages):

blaine.wishart: Hi everyone...I'm on Hainan for the next 3 months.


LLM Finetuning (Hamel + Dan) ▷ #🟩-modal (18 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #learning-resources (9 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #langsmith (4 messages):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #kylecorbitt_prompt_to_model (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #berryman_prompt_workshop (3 messages):

Link mentioned: ExplainPrompt: no description found


LLM Finetuning (Hamel + Dan) ▷ #whitaker_napkin_math (268 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #workshop-2 (7 messages):


LLM Finetuning (Hamel + Dan) ▷ #abhishek_autotrain_llms (57 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #clavie_beyond_ragbasics (3 messages):

Link mentioned: Announcing the Vespa ColBERT embedder: Announcing the native Vespa ColBERT embedder in Vespa, enabling explainable semantic search using token-level vector representations


LLM Finetuning (Hamel + Dan) ▷ #jason_improving_rag (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #gradio (6 messages):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #axolotl (7 messages):


LLM Finetuning (Hamel + Dan) ▷ #zach-accelerate (9 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #wing-axolotl (6 messages):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #freddy-gradio (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #charles-modal (12 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #langchain-langsmith (70 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #allaire_inspect_ai (4 messages):


LLM Finetuning (Hamel + Dan) ▷ #credits-questions (12 messages🔥):

Link mentioned: Tweet from Hamel Husain (@HamelHusain): The $3,500 in compute credits end TODAY. We won't be able to give them out after 11:59 PM PST 5/29/2024 Quoting Eugene Yan (@eugeneyan) PSA: Signups for LLM-conf + finetuning workshop close to...


LLM Finetuning (Hamel + Dan) ▷ #west-coast-usa (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #east-coast-usa (4 messages):

Link mentioned: [NYC] Modal Office Hours · Luma: Have questions about your Modal deployment or just want to learn more? Come by our first office hours in NY! Even if you don't have a particular question in…


LLM Finetuning (Hamel + Dan) ▷ #europe-tz (4 messages):


LLM Finetuning (Hamel + Dan) ▷ #announcements (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #predibase (2 messages):

Link mentioned: Request Free Trial: Try Predibase for free today - Sign up for your trial


LLM Finetuning (Hamel + Dan) ▷ #career-questions-and-stories (8 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #openai (1 messages):

rubenamtz: 👀 , credits are still cooking?


HuggingFace ▷ #announcements (10 messages🔥):

Links mentioned:


HuggingFace ▷ #general (415 messages🔥🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (3 messages):

Link mentioned: Hugging Face - Learn: no description found


HuggingFace ▷ #cool-finds (6 messages):

Links mentioned:


HuggingFace ▷ #i-made-this (10 messages🔥):

Links mentioned:


HuggingFace ▷ #reading-group (1 messages):

taha_69513: Thaaaaaaaaaaaaaaaaaaaaaaaaaaaaaanks 🙌


HuggingFace ▷ #computer-vision (3 messages):

Links mentioned:


HuggingFace ▷ #NLP (7 messages):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #general (205 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (9 messages🔥):

Link mentioned: Tweet from Daniel Han (@danielhanchen): My notes from a NVIDIA research talk: 1) NVIDIA has an research inference 4nm chip doing 96 int4 TOPs/Watt vs Blackwell's 20T/W 2) B200's float4 is exponent=2 and mantissa=2? Maybe I mishear...


Unsloth AI (Daniel Han) ▷ #help (150 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #showcase (4 messages):

For further exploration, check out Ghost Alpha on Hugging Face.

Links mentioned:


Perplexity AI ▷ #general (281 messages🔥🔥):

Links mentioned:


Perplexity AI ▷ #sharing (2 messages):


CUDA MODE ▷ #general (11 messages🔥):

Links mentioned:


CUDA MODE ▷ #torch (1 messages):


CUDA MODE ▷ #announcements (1 messages):


CUDA MODE ▷ #beginner (10 messages🔥):

Links mentioned:


CUDA MODE ▷ #pmpp-book (2 messages):

Link mentioned: Join our Cloud HD Video Meeting: Zoom is the leader in modern enterprise video communications, with an easy, reliable cloud platform for video and audio conferencing, chat, and webinars across mobile, desktop, and room systems. Zoom ...


CUDA MODE ▷ #llmdotc (127 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #youtube-watch-party (3 messages):

Link mentioned: Join the PMPP UI lectures timezones Discord Server!: Check out the PMPP UI lectures timezones community on Discord - hang out with 37 other members and enjoy free voice and text chat.


CUDA MODE ▷ #bitnet (4 messages):


Stability.ai (Stable Diffusion) ▷ #general-chat (152 messages🔥🔥):

Links mentioned:


LM Studio ▷ #💬-general (60 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (30 messages🔥):

Links mentioned:


LM Studio ▷ #📝-prompts-discussion-chat (2 messages):


LM Studio ▷ #🎛-hardware-discussion (21 messages🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (11 messages🔥):


LM Studio ▷ #model-announcements (1 messages):


OpenRouter (Alex Atallah) ▷ #announcements (7 messages):

Link mentioned: WizardLM-2 8x22B by microsoft | OpenRouter: WizardLM-2 8x22B is Microsoft AI's most advanced Wizard model. It demonstrates highly competitive performance compared to leading proprietary models, and it consistently outperforms all existing ...


OpenRouter (Alex Atallah) ▷ #app-showcase (1 messages):


OpenRouter (Alex Atallah) ▷ #general (97 messages🔥🔥):

Links mentioned:


OpenAI ▷ #ai-discussions (85 messages🔥🔥):

Links mentioned:


OpenAI ▷ #gpt-4-discussions (10 messages🔥):


OpenAI ▷ #prompt-engineering (3 messages):


OpenAI ▷ #api-discussions (3 messages):


Nous Research AI ▷ #ctx-length-research (1 messages):

moonride303: https://x.com/jaseweston/status/1795978611784089799


Nous Research AI ▷ #off-topic (6 messages):


Nous Research AI ▷ #interesting-links (16 messages🔥):

Link mentioned: Tweet from Angelica Chen (@_angie_chen): New work w/@sadhikamalladi, @lilyhzhang, @xinyichen2, @QiuyiRichardZ, Rajesh Ranganath, @kchonyc: Contrary to conventional wisdom, RLHF/DPO does not produce policies that mostly assign higher likeli...


Nous Research AI ▷ #general (63 messages🔥🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (1 messages):


LlamaIndex ▷ #blog (2 messages):

Links mentioned:


LlamaIndex ▷ #general (72 messages🔥🔥):

Links mentioned:


Eleuther ▷ #general (20 messages🔥):

Links mentioned:


Eleuther ▷ #research (34 messages🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (7 messages):


Eleuther ▷ #lm-thunderdome (9 messages🔥):

Links mentioned:


Eleuther ▷ #gpt-neox-dev (1 messages):

gpantaz: Thank you for the reply 🙂


Modular (Mojo 🔥) ▷ #general (13 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (1 messages):

ModularBot: From Modular: https://twitter.com/Modular/status/1796606227981726168


Modular (Mojo 🔥) ▷ #ai (3 messages):


Modular (Mojo 🔥) ▷ #🔥mojo (28 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #🏎engine (1 messages):


Modular (Mojo 🔥) ▷ #nightly (9 messages🔥):

Links mentioned:


LangChain AI ▷ #general (41 messages🔥):

Links mentioned:


LangChain AI ▷ #langserve (5 messages):


LangChain AI ▷ #langchain-templates (2 messages):


LangChain AI ▷ #share-your-work (2 messages):

Links mentioned:


LAION ▷ #general (48 messages🔥):

Link mentioned: Tweet from Nirit Weiss-Blatt, PhD (@DrTechlash): Eliezer Yudkowsky's institute published its "2024 Communication Strategy" The main goal (as he argued in TIME magazine) is to 🔻shut down🔻 AI development. So, let's take a look at t...


LAION ▷ #research (2 messages):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #general (22 messages🔥):

Link mentioned: Paper page - YUAN 2.0: A Large Language Model with Localized Filtering-based Attention: no description found


OpenAccess AI Collective (axolotl) ▷ #general-help (12 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #datasets (2 messages):


OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (6 messages):

Links mentioned:


OpenAccess AI Collective (axolotl) ▷ #axolotl-phorm-bot (7 messages):

Links mentioned:


DiscoResearch ▷ #general (5 messages):


DiscoResearch ▷ #discolm_german (23 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #news (10 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (7 messages):


Interconnects (Nathan Lambert) ▷ #random (8 messages🔥):


Interconnects (Nathan Lambert) ▷ #retort-podcast (2 messages):

Link mentioned: The Retort AI Podcast | Murky waters in AI policy: Tom and Nate catch up on many AI policy happenings recently. California's


Cohere ▷ #general (11 messages🔥):


Cohere ▷ #project-sharing (2 messages):


Latent Space ▷ #ai-general-chat (12 messages🔥):

Links mentioned:


AI Stack Devs (Yoko Li) ▷ #events (1 messages):

Link mentioned: Tweet from Rosie @ Rosebud AI 🌹 (@Rosebud_AI): Turn your favorite story into a game using AI! 📚 👾 Get ready for our third Game Jam: “Book to Game”. Use Rosebud Game Maker to transform a literary work into an interactive game and bring stories t...


OpenInterpreter ▷ #general (5 messages):


OpenInterpreter ▷ #O1 (6 messages):


Mozilla AI ▷ #llamafile (3 messages):


MLOps @Chipro ▷ #events (2 messages):


Datasette - LLM (@SimonW) ▷ #ai (2 messages):


tinygrad (George Hotz) ▷ #learn-tinygrad (1 messages):

helplesness: Why is tensoflow better than pytorch?




{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

If you enjoyed AInews, please share with a friend! Thanks in advance!

{% endif %}