Frozen AI News archive

Contextual Position Encoding (CoPE)

**Meta AI** researcher **Jason Weston** introduced **CoPE**, a novel positional encoding method for transformers that incorporates *context* to create learnable gates, enabling improved handling of counting and copying tasks and better performance on language modeling and coding. The approach can potentially be extended with external memory for gate calculation. **Google DeepMind** released **Gemini 1.5 Flash** and **Pro** models optimized for fast inference. **Anthropic** announced general availability of tool use for **Claude**, enhancing its ability to orchestrate tools for complex tasks. **Alexandr Wang** launched **SEAL Leaderboards** for private, expert evaluations of frontier models. **Karpathy** reflected on the 4th anniversary of **GPT-3**, emphasizing scaling and practical improvements. **Perplexity AI** launched **Perplexity Pages** to convert research into visually appealing articles, described as an "AI Wikipedia" by **Arav Srinivas**.

Canonical issue URL

AI News for 5/29/2024-5/30/2024. We checked 7 subreddits, 384 Twitters and 29 Discords (391 channels, and 4383 messages) for you. Estimated reading time saved (at 200wpm): 478 minutes.

A quiet day, but the CoPE paper got some buzz: so we're talking about it.

Traditional LLMs have known issues with simple algorithmic tasks like counting and copying. This is likely an artefact of their positional encoding strategy.

Jason Weston of Meta AI released his paper on CoPE, a new positional encoding method for transformers that takes into account context, creating "gates" with learnable indices.

image.png

Using this, a CoPE LLM can:

You could even modify this concept to use external memory, not merely local context, to calculate the gates.

As Lucas Beyer notes, the raft of position encoding variants this year is perhaps a richer source of research because "Linear attention was about removing capacity from the model, which didn’t make sense long term. Position embedding is about adding missing capabilities to the model, which makes a lot more sense."


{% if medium == 'web' %}

Table of Contents

[TOC]

{% else %}

The Table of Contents and Channel Summaries have been moved to the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


AI Twitter Recap

all recaps done by Claude 3 Opus, best of 4 runs. We are working on clustering and flow engineering with Haiku.

New AI Models and Benchmarks

Advancements in AI Applications and Platforms


AI Reddit Recap

Across r/LocalLlama, r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity. Comment crawling works now but has lots to improve!

Technology Developments and Partnerships

Model Benchmarks and Evaluations

AI Hardware & Performance


AI Discord Recap

A summary of Summaries of Summaries

1. New AI Model Releases and Benchmarks:

2. Optimizations and Advancements in AI Systems:

3. AI Model Fine-tuning and Customization:

4. Competitions and Open Initiatives:


PART 1: High level Discord summaries

Perplexity AI Discord

Perplexity Pages Pioneers Prettier Posts: Perplexity AI has unveiled Perplexity Pages, a tool for transforming research into curated articles, creating an AI Wikipedia. The feature is currently available to Pro users, with expectations to open to more users, and elaborated upon in their blog post.

Grok Woes Lead to Search Superiority Strive: Community member sneakyf1shy strives to build an improved model over Grok, aiming to enhance the search functionality within Perplexity's web application. The community also debated the efficacies of existing models, APIs, and indexed data, citing limitations and envisioning enhancements.

Pages Feedback: The Good, the Bad, the Ugly: Users experimenting with Perplexity Pages shared mixed feedback; some praised its utility while others faced issues, such as missing content sections. The community's pulse ranged from skepticism about Perplexity's indexing to excitement about the feature, with a how-to guide circulating for those interested.

API Angst and Google vs. OpenAI Grudge Match: Technical discussions delved into the challenges of user-friendly API scalability and multi-step reasoning improvements. Meanwhile, the Google-OpenAI rivalry captured attention, sparking debate over their strategic AI moves with speculation around AGI progress and market influence.

AI Ethics and Physics Explored by the Curious: The sharing channel highlighted member contributions on the ethical and physical dimensions of perplexing topics. Links to discussions on consciousness, LLM functionalities, and a supposed pro/con analysis indicate a community engaged in substantive and diverse AI-related themes.


LLM Finetuning (Hamel + Dan) Discord


OpenAI Discord


HuggingFace Discord


LM Studio Discord

Codestral Joins the Coding Model Fray: Mistral introduced Codestral-22B-v0.1, capable of dealing with over 80 programming languages, demonstrating impressive performance in tasks like code instruction and Fill in the Middle (FIM). For those interested in testing the model, download and explore Codestral-22B here.

The Never-Ending Context Length Challenge: Engineers highlighted the limitations of models like the llama series, capped at 4096 tokens, and noted RoPE extension allowing a maximum of 16k tokens, with spirited banter about the importance of context size.

Hardware Discussions Heat Up: The RTX 5090 stirred speculation with its purported 448-bit bus and 28 GB GDDR7 memory. Meanwhile, pragmatic comparisons of CPU inference and the pros and cons of GPU setups, such as using multiple 3090 cards, dominated the discussion.

Whisper & Amuse in Spotlight: A technical hiccup was observed with the Whisper models not being compatible with llama.cpp, as well as a broken GitHub link for Amuse. Solutions included utilizing whisper.cpp and accessing Amuse through an available Hugging Face link.

Practical Tips in Adding Inference GPUs: One discussion clarified the reality of adding additional GPUs for inference in LM Studio, stressing the need for appropriate space, power, and correct settings management, proving that juggling hardware is as much art as it is science.


Unsloth AI (Daniel Han) Discord


Stability.ai (Stable Diffusion) Discord


Eleuther Discord


CUDA MODE Discord

These targeted discussions reflect the community's focus on achieving performance improvements, optimizing cost efficiency, and tackling practical issues faced in implementing machine learning models at scale.


LlamaIndex Discord


LAION Discord


Nous Research AI Discord


Modular (Mojo 🔥) Discord


OpenRouter (Alex Atallah) Discord


Cohere Discord


LangChain AI Discord

Memory Lane with ChatMessageHistory: Kapa.ai illustrated the use of LangChain's ChatMessageHistory class for persisting chat conversations, providing a clear example of maintaining context across sessions, with a nod to the LangChain documentation.

Navigating LLM Conversation Complexity: Discussion centered around the difficulties of designing non-linear conversation flows with Large Language Models (LLMs), citing extraction and JSON handling concerns. An experimental approach on GitHub was linked to demonstrate these challenges in action.

Crafting an Analytical Copilot: Engineering dialogue included strategies for pairing LangChain with a PostgreSQL database, offering insight into handling ambiguous SQL query results via few-shot learning.

Hybrid Agents for Enhanced Interactivity: Integration of create_react_agent and create_sql_agent within LangChain was unraveled, detailing steps to avoid common initialization pitfalls and the importance of naming tools correctly for successful operation.

Evolving AI Assistants & Knowledge Graphs: Wave of new releases like Everything-ai v3.0.0 included advancements like integrating llama.cpp and Qdrant-backed vector databases, while a tutorial video shared across channels provided learners with a practical guide to creating bots using Pinecone, LangChain, and OpenAI.


Interconnects (Nathan Lambert) Discord


OpenInterpreter Discord


Latent Space Discord


Mozilla AI Discord

LLM360 Launches Community AMA: Mozilla AI's LLM360 kicks off community engagement with an AMA on their new 65B model and open-source initiatives, fostering knowledge sharing and Q&A with AI enthusiasts.

Bay Area Engineers, Mark Your Calendars: An IRL Open Source Hack Lab event has been scheduled in the Bay Area, inviting local members to collaborate and share their expertise.

Embeddings Insight Session: A community session on utilizing llamafiles for generating embeddings promises a practical learning experience for engineers seeking to apply embeddings in their machine learning projects.

Developer Support Enhanced at Mozilla AI: In the "Amplifying Devs" event, moderator-led discussions will focus on better supporting the development community within Mozilla AI, an essential platform for developer growth and collaboration.

Tackling LlamaFile Puzzles: Engineers report challenges with granile-34b-code-instruct.Q5_0.llamafile when running on M2 Studio and using VectorStoreIndex in Python, with solutions involving correct IP binding and addressing WSL localhost quirks. Interest in LlamaFiles with vision/image capabilities is growing, highlighted by Mozilla's llava-v1.5-7b-llamafile available on Hugging Face, potentially offering image support for creative AI applications.


OpenAccess AI Collective (axolotl) Discord

Fine-Tuning LLMs for Multimedia Tasks: Members are exploring ideal strategies to fine-tune large language models (LLMs), such as Llava, for tasks involving image and video understanding. The benefits and practicality of using Direct Preference Optimization (DPO) as opposed to Supervised Fine-Tuning (SFT) have precipitated a lively debate, particularly regarding the volume of data required for effective DPO.

DPO's Diminished VRAM Appetite: An unexpected reduction in VRAM usage during DPO has piqued the interest of one engineer, sparking speculation on recent updates that might have led to such efficiency gains.

Protobuf Heavyweight Champion Wanted: There’s an open call within the community for experts with a strong background in Google's Protobuf, especially those who can boast reverse engineering, malware analysis, or bug bounty hunting skills.

SDXL Custom Ads Campaign Hits a Snag: Someone's request for expertise in refining SDXL models is still hanging in the ether, as they aim to optimize their models for producing customized product advertisements and have not yet obtained the desired results with LoRA training or ControlNet.

Small Data for Grand Conversations: Curiosity abounds as to whether a small dataset of merely hundreds of samples could possibly suffice for successful DPO, particularly for domains as nuanced as general chitchat. It has been suggested that manually compiling such a dataset could be a practical approach.


AI Stack Devs (Yoko Li) Discord

AI-Powered Literature to Gameplay Transition: Rosebud AI is hosting a Game Jam: "Book to Game" where participants will use Phaser JS to turn books into games on the AI Game Maker platform, competing for a $500 prize with submissions due by July 1st. News of the jam was shared via Rosebud AI's tweet and interested devs can join their Discord community.

Android Access Annoyance: A newcomer to the Discord community described the Android experience as "a bit hard to navigate... Glitchy and buggy" but confirmed they are still able to engage with content. They also inquired about changing their username, expressing a feeling of being an "alien".


tinygrad (George Hotz) Discord


Datasette - LLM (@SimonW) Discord


MLOps @Chipro Discord

Unfortunately, as there is only one message provided and this message lacks sufficient technical content or details relevant to AI Engineers, it is not possible to create a summary as per the given guidelines. If more messages with the appropriate detail are provided, a summary can be generated.


DiscoResearch Discord


The LLM Perf Enthusiasts AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


{% if medium == 'web' %}

PART 2: Detailed by-Channel summaries and links

Perplexity AI ▷ #announcements (1 messages):


Perplexity AI ▷ #general (686 messages🔥🔥🔥):

- **Grok fails to impress; sneakyf1shy builds better search model**: Users discussed their disappointment with Grok and sneakyf1shy mentioned working on a similar project with intentions of enhancement. They aim to surpass Perplexity's web app by creating a comprehensive custom searching pipeline.
- **OpenAI and API enhancements**: Conversations highlighted the challenges of creating user-friendly APIs and scaling them effectively. Some users, such as sneakyf1shy, expressed interest in developing API solutions that could improve multi-step reasoning and integrating own indexing/cache layers.
- **Perplexity Pages gains traction; user experiences varied**: Many users explored Perplexity Pages, sharing their experiences and learnings. Some users encountered issues like missing sections in converted threads, while others found it a valuable addition for documentation and knowledge databases. One user shared a [Perplexity Pages guide](https://www.perplexity.ai/page/How-to-Use-FvLfzZ_ATyqE2n_tAGKk7A).
- **Skepticism and API limitations**: Users expressed skepticism about Perplexity's use of its own index, questioning the true capabilities of their web scraper. Some lamented the inactivity and limited availability of the API, while others discussed alternative models and their efficiencies.
- **Google and OpenAI comparisons stir debate**: Lively debates ensued about Google’s and OpenAI’s AI strategies, resource usage, and effectiveness in comparison to competitors like Nvidia. Users speculated on AGI developments and commercial impacts, especially regarding OpenAI's products and potential future releases.

Links mentioned:


Perplexity AI ▷ #sharing (15 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #general (50 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #workshop-1 (5 messages):

Link mentioned: fine-tuning workshop 1 slides: LLM Fine Tuning For Data Scientists & Software Engineers


LLM Finetuning (Hamel + Dan) ▷ #asia-tz (2 messages):

- **New Member from Sydney Joins the Team**: A new member introduced themselves, noting they are a Senior Manager in Advanced Analytics based in Sydney, Australia. They expressed interest in applying fine-tuning for specific use cases and deploying LLMs using minimal prompting, as well as learning about best practices for hosting and deploying LLMs in production settings.

- **Global AI Hackathon Alert**: An upcoming **Global AI Hackathon** from June 7 to 9 was announced, facilitating events in multiple cities including Singapore, Sydney, and San Francisco. Attendees are encouraged to RSVP via [this link](https://lu.ma/igqisb0e), noting that the hackathon is backed by top AI builders and aims to address "AI for a better world".

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #🟩-modal (74 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #learning-resources (5 messages):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #jarvis-labs (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #replicate (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #langsmith (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #ankurgoyal_textsql_llmevals (2 messages):


LLM Finetuning (Hamel + Dan) ▷ #berryman_prompt_workshop (16 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #workshop-2 (6 messages):

Link mentioned: Fine-tuning workshop 2 slides: Mastering LLMs A Conference For Developers & Data Scientists


LLM Finetuning (Hamel + Dan) ▷ #workshop-3 (18 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #gradio (1 messages):


LLM Finetuning (Hamel + Dan) ▷ #axolotl (56 messages🔥🔥):


LLM Finetuning (Hamel + Dan) ▷ #zach-accelerate (35 messages🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #wing-axolotl (6 messages):


LLM Finetuning (Hamel + Dan) ▷ #freddy-gradio (7 messages):

Link mentioned: Sharing Your App: A Step-by-Step Gradio Tutorial


LLM Finetuning (Hamel + Dan) ▷ #charles-modal (86 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #langchain-langsmith (59 messages🔥🔥):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #simon_cli_llms (1 messages):

imaurer: Simon's newsletter is a great resource: https://simonwillison.net/about/#subscribe


LLM Finetuning (Hamel + Dan) ▷ #allaire_inspect_ai (93 messages🔥🔥):

- **Quarto for Inspect site**: Members discussed the use of **Quarto** for the [Inspect AI site](https://ukgovernmentbeis.github.io/inspect_ai/), with some expressing strong approval, "Quarto is the best."
- **Logs as a unit of reproducibility**: The use of logs as a unit of reproducibility in Inspect AI received praise from several members. One said, "This feels ahead of its time (in a really good way) 👀."
- **Links and resources for Inspect AI**: Multiple important links were shared, including the [Inspect homepage](https://ukgovernmentbeis.github.io/inspect_ai/), [AI Safety Institute](https://www.aisi.gov.uk/), and the [Inspect LLM workshop repository](https://github.com/jjallaire/inspect-llm-workshop).
- **Concerns and feedback on Inspect AI**: Attendees discussed various aspects and suggestions for Inspect AI, including the feature to compare runs in the UI and ideas for future enhancements. "Solvers is amazing," one member remarked, highlighting the tool's flexibility and composability.
- **Recording issues resolved**: There were initial issues with accessing video recordings of sessions, but these were subsequently addressed. "JJ's recording now works for me," a member confirmed after the fixes.

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #credits-questions (42 messages🔥):

Link mentioned: Tweet from Hamel Husain (@HamelHusain): The $3,500 in compute credits end TODAY. We won't be able to give them out after 11:59 PM PST 5/29/2024 Quoting Eugene Yan (@eugeneyan) PSA: Signups for LLM-conf + finetuning workshop close to...


LLM Finetuning (Hamel + Dan) ▷ #west-coast-usa (5 messages):


LLM Finetuning (Hamel + Dan) ▷ #east-coast-usa (16 messages🔥):

Link mentioned: Live from Civic Hall! AI Tinkerers Meetup | NY#TechWeek [AI Tinkerers - New York City] : no description found


LLM Finetuning (Hamel + Dan) ▷ #europe-tz (27 messages🔥):


LLM Finetuning (Hamel + Dan) ▷ #announcements (4 messages):

Links mentioned:


LLM Finetuning (Hamel + Dan) ▷ #predibase (1 messages):

abhay_m: 👋


OpenAI ▷ #annnouncements (3 messages):


OpenAI ▷ #ai-discussions (375 messages🔥🔥):

<ul>
    <li><strong>Clarifications on GPT-4o Availability:</strong> Multiple members asked about GPT-4o availability for free users. It was explained that free users cannot force access and would be automatically switched between GPT-3.5 and GPT-4o based on the system's discretion.</li>
    <li><strong>Concern Over Subscription Value:</strong> A user expressed confusion over continuing to pay for ChatGPT. Responses highlighted advantages like early access to new features, quotas, and additional functionalities exclusive to subscribers.</li>
    <li><strong>Discussion on AI's Analytical Capabilities:</strong> Users debated how well different AI models handle logical reasoning tasks, like the "apples test" and the "susan test." It was noted that AI models often exhibit biases based on training data.</li>
    <li><strong>Code and Model Usage Insights:</strong> Members discussed using various AI models for coding assistance, comparing the performance of tools like GPT-4o, Mistral’s codestral, and Copilot. Speed and accuracy were highlighted as key factors in choosing specific models.</li>
    <li><strong>News and Media Detection AI Idea:</strong> A user discussed an AI concept for detecting fake news and propaganda by assessing posts on social media. Another user suggested it might run into common issues like hallucination and bias in AI's interpretation.</li>
</ul>

Links mentioned:


OpenAI ▷ #gpt-4-discussions (64 messages🔥🔥):


OpenAI ▷ #prompt-engineering (2 messages):


OpenAI ▷ #api-discussions (2 messages):


HuggingFace ▷ #announcements (5 messages):

Link mentioned: HuggingChat: Making the community's best AI chat models available to everyone.


HuggingFace ▷ #general (362 messages🔥🔥):

Links mentioned:


HuggingFace ▷ #today-im-learning (1 messages):

venatic007: ✋🏻


HuggingFace ▷ #cool-finds (12 messages🔥):

Links mentioned:


HuggingFace ▷ #i-made-this (8 messages🔥):

- **Demo Nvidia's embedding model**: A member shared a demo for Nvidia's new embedding model and requested PRs for cool examples or improved functions. *"You can test it out here: [Nvidia Embed V1](https://huggingface.co/spaces/Tonic/Nvidia-Embed-V1/)."*
- **Llama 3 SOLAR recreation attempt**: A user attempted to recreate Upstage's old Solar models using Llama 3. They used datasets like **`llm-wizard/alpaca-gpt4-data`** and [shared the model on HuggingFace](https://huggingface.co/cookinai/Llama-3-SOLAR-v0.2).
- **Codestral-22B quantized version**: Shared a quantized version of Codestral-22B-v0.1, created using llama.cpp, beneficial for code-related tasks. *"More details in the [Blogpost](https://mistral.ai/news/codestral/)."*
- **DuckDB supports Hugging Face datasets on WrenAI**: Announcement about DuckDB supporting the `hf://` path, enabling easy loading and querying of Hugging Face datasets in WrenAI. Learn more [here](https://blog.getwren.ai/how-to-load-huggingface-datasets-into-duckdb-and-query-with-gpt-4o-c2db89519e4d).
- **LLMinator v1.0.3 releases new features**: LLMinator now supports websocket interaction, context-aware chatbots, model conversion, and customized LLM inference parameters. Check out the project on [GitHub](https://github.com/Aesthisia/LLMinator).

Links mentioned:


HuggingFace ▷ #reading-group (3 messages):

Links mentioned:


HuggingFace ▷ #computer-vision (17 messages🔥):

Links mentioned:


HuggingFace ▷ #NLP (9 messages🔥):

Links mentioned:


LM Studio ▷ #💬-general (91 messages🔥🔥):

- **Codestral Model Release and Uses Discussed**: Released the **Codestral-22B-v0.1** model, which handles 80+ programming languages including Python, Java, and JavaScript. The model supports code instruction and Fill in the Middle (FIM) functionalities; [more details in the blogpost](https://mistral.ai/news/codestral/).
- **Concerns about Model Variants**: Members discussed the practicality of different quantization variants, with some noting that **_S variants** are generally too "smoothbrained" and not useful.
- **Code Models and Prompt Formats**: The recommended format for querying Codestral-22B-v0.1-GGUF was discussed, referencing [this GitHub link](https://huggingface.co/bartowski/Codestral-22B-v0.1-GGUF#prompt-format).
- **Loading Issues on Limited Hardware**: A user experienced long loading times on **LM Studio** due to low system specs, suggesting smaller models might work better.
- **Inquiring Business Contact Options**: A member inquired about direct business contact for a project, and was guided to email **[email protected]** for further discussion.

Links mentioned:


LM Studio ▷ #🤖-models-discussion-chat (62 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧠-feedback (1 messages):

cancerous1: thanks for the rocm/windows build 🍻 you doubled my real estate for models


LM Studio ▷ #📝-prompts-discussion-chat (1 messages):

tiltspinner: Thanks!


LM Studio ▷ #⚙-configs-discussion (2 messages):


LM Studio ▷ #🎛-hardware-discussion (207 messages🔥🔥):

Links mentioned:


LM Studio ▷ #🧪-beta-releases-chat (3 messages):


LM Studio ▷ #amd-rocm-tech-preview (2 messages):

Link mentioned: Amuse_v1.3.0.zip · Stackyard-AI/Amuse at main: no description found


LM Studio ▷ #model-announcements (1 messages):


Unsloth AI (Daniel Han) ▷ #general (233 messages🔥🔥):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #random (6 messages):

Links mentioned:


Unsloth AI (Daniel Han) ▷ #help (130 messages🔥🔥):

Links mentioned:


Stability.ai (Stable Diffusion) ▷ #general-chat (351 messages🔥🔥):

Links mentioned:


Eleuther ▷ #general (11 messages🔥):

Links mentioned:


Eleuther ▷ #research (50 messages🔥):

Links mentioned:


Eleuther ▷ #scaling-laws (191 messages🔥🔥):

Links mentioned:


Eleuther ▷ #interpretability-general (1 messages):

Link mentioned: Tweet from Jannik Brinkmann (@BrinkmannJannik): Can we find evidence of latent reasoning and search in language models? Our #acl2024 paper (w/ @abhayesian and @VictorLevoso) reverse-engineers the internal mechanisms of a transformer trained on tre...


Eleuther ▷ #lm-thunderdome (19 messages🔥):

Links mentioned:


Eleuther ▷ #gpt-neox-dev (4 messages):


CUDA MODE ▷ #general (3 messages):

Links mentioned:


CUDA MODE ▷ #triton (9 messages🔥):

Links mentioned:


CUDA MODE ▷ #torch (6 messages):

Link mentioned: Added memory budget to partitioner by Chillee · Pull Request #126320 · pytorch/pytorch: Stack from ghstack (oldest at bottom): #127520 -> #126320 #127446


CUDA MODE ▷ #beginner (11 messages🔥):

Link mentioned: hqq/hqq/utils/generation_hf.py at master · mobiusml/hqq: Official implementation of Half-Quadratic Quantization (HQQ) - mobiusml/hqq


CUDA MODE ▷ #torchao (4 messages):


CUDA MODE ▷ #hqq (1 messages):


CUDA MODE ▷ #triton-viz (3 messages):


CUDA MODE ▷ #llmdotc (122 messages🔥🔥):

Links mentioned:


CUDA MODE ▷ #youtube-watch-party (3 messages):


CUDA MODE ▷ #bitnet (72 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #blog (5 messages):

Link mentioned: Solving the challenges of using LLMs in production with financial services data, Wed, Jun 12, 2024, 6:00 PM | Meetup: If you are building NLP pipelines for processing financial services data, you will know how hard it can be to manage vector databases in production, reliably process large


LlamaIndex ▷ #general (89 messages🔥🔥):

Links mentioned:


LlamaIndex ▷ #ai-discussion (1 messages):


LAION ▷ #general (52 messages🔥):

Links mentioned:


LAION ▷ #announcements (1 messages):

Links mentioned:


LAION ▷ #research (33 messages🔥):

Links mentioned:


Nous Research AI ▷ #ctx-length-research (11 messages🔥):


Nous Research AI ▷ #off-topic (5 messages):

Link mentioned: Codestral Mistral AI's first-ever code model: Codestral, is Mistal's first-ever code model. Codestral is an open-weight generative AI model explicitly designed for code generation tasks. It helps develop...


Nous Research AI ▷ #interesting-links (3 messages):

Link mentioned: SEAL leaderboards: no description found


Nous Research AI ▷ #general (41 messages🔥):

Links mentioned:


Nous Research AI ▷ #ask-about-llms (7 messages):


Nous Research AI ▷ #rag-dataset (14 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #general (30 messages🔥):

Links mentioned:


Modular (Mojo 🔥) ▷ #💬︱twitter (1 messages):

ModularBot: From Modular: https://twitter.com/Modular/status/1796232248678883347


Modular (Mojo 🔥) ▷ #📺︱youtube (1 messages):

Link mentioned: Speed up K-Means clustering by porting Python implementation to Mojo🔥: In this video we'll share a step-by-step guide to porting kmeans clustering from Python+NumPy to pure Mojo for huge (250x) speedup! How? Mojo is Pythonic in ...


Modular (Mojo 🔥) ▷ #🔥mojo (5 messages):


Modular (Mojo 🔥) ▷ #performance-and-benchmarks (6 messages):

Links mentioned:


Modular (Mojo 🔥) ▷ #nightly (22 messages🔥):

Link mentioned: Build software better, together: GitHub is where people build software. More than 100 million people use GitHub to discover, fork, and contribute to over 420 million projects.


OpenRouter (Alex Atallah) ▷ #app-showcase (5 messages):


OpenRouter (Alex Atallah) ▷ #general (45 messages🔥):

Links mentioned:


Cohere ▷ #general (37 messages🔥):


Cohere ▷ #project-sharing (1 messages):

sssandra: hi, let me give you some cohere credits! dming


LangChain AI ▷ #general (34 messages🔥):

Links mentioned:


LangChain AI ▷ #share-your-work (2 messages):

Link mentioned: How to build chat with your data using Pinecone, LangChain and OpenAI: I show step by step how to build a Chatbot using Pinecone, LangChain and OpenAI in this easy to follow tutorial for beginners.I ingest my entire blog full of...


LangChain AI ▷ #tutorials (1 messages):

zackproser: https://www.youtube.com/watch?v=Bxj4btI3TzY


Interconnects (Nathan Lambert) ▷ #news (23 messages🔥):

Links mentioned:


Interconnects (Nathan Lambert) ▷ #ml-drama (9 messages🔥):

Link mentioned: OpenAI board members respond to a warning by former members: no description found


OpenInterpreter ▷ #general (21 messages🔥):

Links mentioned:


OpenInterpreter ▷ #O1 (9 messages🔥):

Link mentioned: 2Noise/ChatTTS · Hugging Face: no description found


Latent Space ▷ #ai-general-chat (17 messages🔥):

Links mentioned:


Latent Space ▷ #ai-announcements (5 messages):

Link mentioned: Tweet from Latent Space Podcast (@latentspacepod): 🆕 pod: How to train a Million Context LLM! @ylecun says we should publish, or perish. We asked @markatgradient to spill ALL the beans on how his team extended Llama-3 to 1M+ context with ~perfect @G...


Latent Space ▷ #llm-paper-club-west (2 messages):

<ul>
    <li><strong>No messages to summarize</strong>: The channel "llm-paper-club-west" currently holds no substantial messages that can be summarized. Only placeholders are present without any actual content to analyze.</li>
</ul>

Mozilla AI ▷ #announcements (1 messages):


Mozilla AI ▷ #llamafile (19 messages🔥):

Link mentioned: Mozilla/llava-v1.5-7b-llamafile at main: no description found


OpenAccess AI Collective (axolotl) ▷ #general (10 messages🔥):


OpenAccess AI Collective (axolotl) ▷ #axolotl-dev (4 messages):


OpenAccess AI Collective (axolotl) ▷ #general-help (1 messages):


AI Stack Devs (Yoko Li) ▷ #events (1 messages):

Link mentioned: Tweet from Rosie @ Rosebud AI 🌹 (@Rosebud_AI): Turn your favorite story into a game using AI! 📚 👾 Get ready for our third Game Jam: “Book to Game”. Use Rosebud Game Maker to transform a literary work into an interactive game and bring stories t...


AI Stack Devs (Yoko Li) ▷ #ai-town-discuss (9 messages🔥):


tinygrad (George Hotz) ▷ #general (3 messages):


tinygrad (George Hotz) ▷ #learn-tinygrad (6 messages):


Datasette - LLM (@SimonW) ▷ #ai (2 messages):


MLOps @Chipro ▷ #events (1 messages):

_awesomewaffle: Will be at the PRS event at Netflix tomorrow . Anyone else attending this event?


DiscoResearch ▷ #general (1 messages):

Link mentioned: Tweet from LAION (@laion_ai): Help us build an open GPT-4-Omni! With this blog post we show promising directions (including data sets and tutorials) https://laion.ai/notes/open-gpt-4-o/

{% else %}

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: [{{ email.subject }}]({{ email_url }})!

{% endif %}


If you enjoyed AInews, please share with a friend! Thanks in advance!