This splits up the langchain repo to be more maintainable and scalable, an inevitable part of every integration-heavy open source framework:
Itâs all backwards compatible so no big rush.
In other news, Mistral has a new Discord (weâll be adding to our tracker) and Anthropic is rumored to be raising another $3b.
[TOC]
OpenAI Discord Summary
- Discussion on Information Leakage in AI Training and Mixture of Experts Models:
@moonkingyt
queried about information leakage during AI conversations and expressed curiosity about a model called âmixtral 8x7bâ. It was clarified that this model is a type of Mixture of Experts (MoE) model. - Advanced Use Cases of AI: Users explored various potential AI use cases, such as modifying a companyâs logo for specific events, explaining complex concepts in large classes at universities, and potential applications for local language models.
- Technical Challenges with OpenAI: Multiple users shared issues they have experienced with ChatGPT, such as server unresponsiveness, slow response times, and difficulty with image uploads. Users also discussed the feasibility of training an AI model using a Discord server and HTML text files, and how to go about doing it.
- Concerns about ChatGPTâs Performance and Behavior: User frustrations included perceived degradation in response quality/nuance, poor memory retention, and poor customer support. Some users reported that GPT is not answering questions as before and instead provides references to search answers on the internet.
- Discussion on Accessing OpenAI GPT-4 API: Users discussed the access and cost of GPT-4âs API, the usage limit of the ChatGPT Plus, how to access it, and the process of reaching human support. An issue about incorrect highlighting in diff markdowns was found and reported.
- Prompt Engineering Techniques: Users discussed various advanced prompt engineering tactics, with an emphasis on clear prompt setting for obtaining desired outputs. Techniques such as outlining for longer written outputs, iterative proofreading, and implementation of a dynamic model system were proposed.
- Oauth2 Issues and Solutions: Oauth2 authentication issues were solved in collaboration among users. Discussion regarding minor bugs and the process of reporting them also took place.
- Using AI for Games: Users shared their attempts at getting AI to play games like hangman. A potential feature of toggling the Python display in the interface for such engagements was proposed.
- Request for Detailed 2D Illustration: A user requested detailed 2D illustration, and was guided to use DALL·E 3 on Bingâs image creator or through ChatGPT Plus! An issue of not being able to edit custom GPTâs was also reported.
OpenAI Channel Summaries
â· #ai-discussions (42 messagesđ„):
- Information Leakage in AI Training:
@moonkingyt
queried about cases where chatbots leak information during conversations.@xiaoqianwx
responded affirmatively without further elaborating the context. - Discussion on Mixture of Experts Models:
@moonkingyt
expressed curiosity about a term called âmixtral 8x7bâ.@xiaoqianwx
clarified that itâs a type of Mixture of Experts (MoE) model. - Logo Generation with AI:
@julien1310
sought guidance on whether an AI can modify a companyâs existing logo for a special event like a 20th anniversary.@elektronisade
suggested using specific features like inpainting from StyleGAN or its derivatives. - Usage of AI in Educational Institutions:
@offline
mentioned that some universities donât permit AI assistance, hinting at the potential of AI like GPT-3 to explain complex concepts, often in a more detailed manner than instructors in large classes. - Question about Local Language Models:
@Shunrai
asked the group for recommendations on good local language models. It spurred subsequent discussions, but no specific models were recommended. - Microsoftâs New Language Model:
@ăïœïœ ïœïœïœïœïœïœïœïœïœ ă
shared news about a small language model named Phi-2 being released by Microsoft. Further details or a link to more information were not provided. - AGI (Artificial General Intelligence) Debates: A side discussion was sparked by
@chief_executive
questioning the term AGI if it keeps getting ânerfedâ every step. Meanwhile,@ăïœïœ ïœïœïœïœïœïœïœïœïœ ă
criticized constant shifting of AGIâs goalpost.
â· #openai-chatter (337 messagesđ„đ„):
- Issues with OpenAI ChatGPT: Users
@ruhi9194
,@stupididiotguylol
,@rjkmelb
,@yyc_
,@eskcanta
,@toutclaquer
,@gingerai
,@ivy7k
,@millicom
,@ayymoss
,@knightalfa
,@loschess
,@primordialblanc
reported multiple technical issues with ChatGPT including server unresponsiveness, slow response times, difficulty with image uploads, and error messages such as âHmmâŠsomething seems to have gone wrong.â - Discussion on Using Custom GPT Files: User
@maxdipper
asked about the possibility of training a model off a Discord server and whether HTML text files can be utilized for feeding data to the model. - Confusion Over Image Analysis in ChatGPT: User
@toutclaquer
shared an issue where image analysis in ChatGPT was not working. Error messages were displayed when attempting to analyze images. - Concerns about ChatGPTâs Behavior and Performance: Users
@prncsgateau
,@the_time_being
,@.nasalspray
,@one_too_many_griz
,@fefe_95868
shared their frustrations with ChatGPTâs change in behavior, perceived degradation in response quality/nuance, poor memory retention, and poor customer support. - Questions about Accessing GPT-4 API: User
@moonkingyt
inquired about the access and cost of GPT-4âs API.@elektronisade
provided a link to OpenAIâs documentation outlining the access conditions and@offline
further clarified the requirement of being a Pay-As-You-Go customer with a minimum spend of $1.@DawidM
raised concerns about degradation of service for premium users.
â· #openai-questions (170 messagesđ„đ„):
- ChatGPT Troubleshooting and Questions: Many users such as
@mustard1978
,@yyc_
,@blinko0440
,@jonsnov
, among others, reported various issues with the platform, with common challenges being inconsistency of responses, error messages, issues in uploading images, and problems when the system has reached a usage limit.@solbus
provided extensive troubleshooting advice, suggesting solutions such as checking VPN status, trying different browsers, and clearing browser cache and cookies. - Access Termination Issues: User
@mouse9005
shared an issue around account termination and appealed for help.@rjkmelb
advised them to contact OpenAI support via their help center. - ChatGPT Plus Subscription and Usage Limit: A discussion occurred centered on the usage limit of the ChatGPT Plus, involving
@lunaticspirit
,@rjkmelb
, and@DkBoss
. The limit for ChatGPT Plus is 40 messages per 3 hours. A reference for the discussion was provided in the form of this link. - Accessing Tools and Plugins in the New ChatGPT Plus UI:
@_.gojo
inquired about how to access the new GPT-4 model and plugins in the playground with the ChatGPT Plus subscription.@solbus
provides a link to how one can access GPT-4 API and also explains how to enable plugins from the account settings. - Payment Issues: Users like
@Ugibugi
,@michaelyungkk
, and@tardis77b
have encountered challenges with the payment process for subscribing to the service.@solbus
and@eskcanta
recommended contacting OpenAI help center for assistance.
â· #gpt-4-discussions (74 messagesđ„đ„):
- Oauth2 Issues and Solutions: Users
@jdo300
and@draennir
had discussions about issues they were facing with Oauth2 authentication for their API servers. They worked collaboratively to solve their issues and@draennir
eventually solved his issue by addingapplication/x-www-form-urlencoded
among the parsers in Express. - Dynamic model suggestion for GPT: User
@offline
suggested the implementation of a dynamic model system that would decide which GPT model to use based on the requirements of the userâs request. The idea is further discussed and supported by@solbus
. - Minor bugs & Bug reporting: An issue of incorrect highlighting in diff markdowns was found and reported by user
@offline
. They discussed about the bug reporting process with@solbus
. - GPT Behaviour Issues: User
@maypride
reported an issue with GPT not answering questions like before and instead it was providing references to search answers on the internet.@cris_cezar
reported an inability to edit one of their custom GPTâs, even after trying multiple browsers and clearing cache. - User Interface & Support:
@cris_cezar
expressed confusion about the Discord interface, comparing it to an aviation cockpit. The user was guided to the OpenAI staff support via the help.openai.com chat.@solbus
and@eskcanta
discussed the process of reaching human support through the help.openai.com chat interface.
â· #prompt-engineering (32 messagesđ„):
-
ChatGPT and Writing Specific Word Count:
@davidmajoulian
mentioned that ChatGPT underproduces the number of words when asked to write an article of a specific length, to which@rjkmelb
responded that ChatGPT doesnât know how to count words and itâs not realistic to expect it to produce an article of a specific length. The two users also discussed the possibility of using tools that add such functionality to OpenAI models. -
Exploration of Advanced Prompt Engineering:
@yautja_cetanu
initiated a discussion on advanced prompt engineering and the challenge in finding before-and-after examples of prompt adjustment that significantly improved the modelâs performance. Several users affirmed that given the advancements in models like ChatGPT, many traditional prompt engineering techniques seem less important, with the quality of prompt output largely leaning on the clarity and specificity of the instructions given to the AI. -
Prompt Engineering Techniques:
@eskcanta
and@cat.hemlock
emphasized on the importance of clear, specific, error-free prompt setting for obtaining desired results from the AI, with the latter suggesting the use of outlining for longer written outputs and iterative proofreading. -
Using AI for Games:
@eskcanta
shared a failed attempt at getting the AI to play hangman and@thepitviper
responded with an example where they managed to get the AI to generate a working Python game for hangman. There was also a suggestion for the ability to toggle Python display on and off for this kind of engagement. -
Detailed 2D Illustration Request:
@yolimorr
requested for a detailed 2D illustration, and@solbus
mentioned that image generation is not supported on the server and directed the user to use DALL·E 3 on Bingâs image creator or with ChatGPT Plus!.
â· #api-discussions (32 messagesđ„):
-
Prompt Length with ChatGPT:
@davidmajoulian
asked about the issue of ChatGPT producing less content than requested in his prompts. He stated that when he requested a 1,500-word article, it returned only roughly 800 words.@rjkmelb
clarified that ChatGPT doesnât know how to count words and suggested using the tool to assist in constructing an article rather than expecting a set word length.@rjkmelb
also hinted at building other tools using the OpenAI GPT-4 API for advanced features. -
Advanced Prompt Engineering:
@yautja_cetanu
initiated a discussion on advanced prompt engineering, expressing difficulty in providing pre and post examples for his meetup talk because of how good chatGPT already is.@tryharder0569
suggested focusing more on how specific prompts enhance an applicationâs functionality.@madame_architect
recommended considering step-back prompting or thinking about output quality in terms of higher value or more helpfulness instead of a working vs non-working dichotomy. -
Prompt Output Quality Control:
@cat.hemlock
emphasized the importance of being specific and deliberate in instructing the AI to avoid generating âAI slopâ (garbage output). They suggested using ChatGPT to generate an article outline first and refining it before expanding it into full content, followed by a final proofread. -
ChatGPT Capability Limitations: Certain limitations of ChatGPT were also mentioned.
@eskcanta
provided an example of the model failing to correctly play a game of Hangman even with a known word, known order of letters, in-conversation correction, and step by step examples to follow.@thepitviper
demonstrated the use of Python with ChatGPT to run a slow, yet accurate Hangman game. -
Python Tool Display Toggle: A potential feature of toggling the Python display in the interface was discussed between
@thepitviper
and@eskcanta
which could improve the user experience, though this would need to satisfactorily address situations where the system is performing an action to prevent undesired outcomes.
Nous Research AI Discord Summary
- Summary of extensive discussions on the SOLAR 10.7B AI model across multiple channels involving benchmark results, performance comparison, and validity of its claims on outperforming other models. The model showcased significant gains in the AGIEval, with interesting skepticism and eventual recognition of its performance.
- âfrom first look, itâs a flopâ -
@teknium
- âTLDR: Good but dont deserve so much attâ -
@artificialguybr
- âfrom first look, itâs a flopâ -
- Interaction on technical challenges in various areas: Base AI model described as an âuntamed beastâ, the query on GPU performance leading to a discussion on cooling methods for Lambdaâs Vector One workstations, and queries on optimizing the inference speed with HuggingFace transformers.
- Dialogue on OpenHermes 2.5âs ability for function calling and the complexities involved with shared experiences and resources. Microsoftâs Phi-2 was mentioned for reportedly outperforming Mistral 7b and Llama 7b on benchmarks.
- Ongoing discourse on various Mistral models, the potential of fine-tuning models on Apple Silicon. Resources and discussion threads shared for further exploration.
- Community collaboration and assistance sought in areas like AI repositories needing frontend interface development, advice on choosing an orchestration library for developing an LLM-based application, with recommendations against LangChain favoring Llama-Index or custom platforms.
- The release announcement of the UNA-Neural-Chat-v3-3-Phase 1 model, described as outperforming the original in initial tests.
- Miscellaneous points: Humorous observations on working at Microsoft, involvement in gene editing technologies, grammar and spelling problem-solving using bots, sharing content of interest across different channels in the form of tweets and video links.
Nous Research AI Channel Summaries
â· #ctx-length-research (1 messages):
teknium: https://fxtwitter.com/abacaj/status/1734353703031673200
â· #off-topic (18 messagesđ„):
- Concerns about Base Model:
@nagaraj_arvind
described the base AI model as an âuntamed beast,â expressing some challenges or difficulties with it. - AI Repos Assistance:
@httpslinus
sought out any AI-related repos that need help with frontend interface building or extension, intending to assist with their current skill set. - Workplace Woes:
@airpods69
shared their experience of leaving a workplace due to dissatisfaction and stress caused by the practices of a senior developer. - DeciLM-7B Discussion:
@pradeep1148
shared a YouTube video about DeciLM-7B, a 7 billion-parameter language model. However,@teknium
expressed disappointment, stating that DeciLM-7Bâs performance is driven by its gsm8k score and is worse at mmlu than Mistral. - Vector One Workstations Analysis:
@erichallahan
shared a Twitter post providing analysis on Lambdaâs Vector One workstations. This led to a discussion on cooling methods, with@fullstack6209
expressing a desire for air-cooled systems, and@erichallahan
agreeing with this preference. - GPU Utilization Query:
@everyoneisgross
asked about the GPU performance when running an LLM on GPU VRAM, using python to max out CPU, and multi-screening YouTube videos at the same time, alluding to possible visual glitches or artifacts.
â· #benchmarks-log (15 messagesđ„):
- SOLAR 10.7B Benchmarks Shared:
@teknium
posted benchmark scores for the SOLAR 10.7B AI model on various tasks, including:truthfulqa_mc
: Value: 0.3182 (mc1), 0.4565 (mc2) andarc_challenge
: Value: 0.5247 (acc), 0.5708 (acc_norm) - Comparative Evaluation with Other Models:
@teknium
compared the performance of SOLAR 10.7B with that of Mistral 7b. SOLAR 10.7B seemed to achieve a notable gain in AGIEval (39% vs 30.65%) but demonstrated similarities in other evaluations (72% vs 71.16% in gpt4all, 45% vs 42.5% in truthfulqa). - Remarks on SOLAR 10.7Bâs Performance: Despite the initial perception that SOLAR 10.7Bâs performance was not impressive (
"from first look, it's a flop"
and"good but probably not as good as yi like it claims"
), the comparative results led@teknium
to reassess the modelâs performance, significantly in AGIEval. - Final Thoughts on SOLAR 10.7B:
@artificialguybr
summarised the discussion by saying that while SOLAR 10.7B had decent performance, it did not seem to warrant significant attention ("TLDR: Good but dont deserve so much att"
).
â· #interesting-links (42 messagesđ„):
- SOLAR-10.7B Model Discussion:
@metaldragon01
linked to a Hugging Face post that introduced the first 10.7 billion parameter model, SOLAR-10.7B. - Claims of Outperforming Other Models: The SOLAR team claims their model outperforms others, including Mixtral. Users
@n8programs
and@teknium
expressed skepticism about the claims. - Evaluating the SOLAR-10.7B Model:
@teknium
began benchmarking the new SOLAR model and planned to compare the results with Mixtralâs scores. - SOLAR-10.7B Characterization: In discussion, users concluded that SOLAR is likely pre-trained but may not be a base model, based on its similarity to a past iteration of SOLAR, which was not a base model.
- Tweet on Mixtral and 3090 Setup:
@fullstack6209
asked about the optimal setup for Mixtral and a 3090, to which@lightningralf
responded with a Twitter link possibly containing the answer.
â· #bots (48 messagesđ„):
- Gene Editing Technologies: User
@nagaraj_arvind
asked about advances in gene editing technologies since 2015. The bot replied but the details were not elaborated in the messages. - Backward Sentences Completion:
@crainmaker
seeked the bot to complete a string of backward sentences. - Questions in Portuguese:
@f3l1p3_lv
asked a series of questions in Portuguese, most of which are related to basic math problems, day of the week calculations, analogy formation, and appropriate pronoun selection for sentences. The bot@gpt4
and@compbot
helped to solve most of these questions. - Grammar and Spelling Problems:
@f3l1p3_lv
also posted multiple spelling and grammar problems for the bot to solve, such as filling the blanks with specific letters or groups of letters, selecting pronouns for sentences, and choosing correctly written sentences. The bot@gpt4
provided the solutions and explanations.
â· #general (337 messagesđ„đ„):
-
Discussion on Function Calling Capabilities: User
@realsedlyf
inquired about OpenHermes 2.5âs ability to do function calling, other users, including@.beowulfbr
and@tobowers
, shared their experiences with function calling with various models.@.beowulfbr
shared a link to a HuggingFace Model with function calling capabilities, whereas@tobowers
shared his implementation of function calling using the SocialAGI code from GitHub. -
Announcement of UNA-Neural-Chat-v3-3-Phase 1 Release: User
@fblgit
announced the release of UNA-Neural-Chat-v3-3-Phase 1, which in initial tests outperforms the original model. -
Discussion on Optimization of Inference Speed w/ hf transformers:
@lhl
shared his experiences on optimizing the inference speed with HuggingFace transformers, detailing how he achieved better results. -
Discussion on Usefulness and Performance of Small Scale Models:
@a.asif
asked about the best performers in the realm of small scale models that could be run on laptops. Plugyy raised a question about the possibility of running the Mixtral MoE model on limited hardware with the help of quantization and llama.cpp. -
Links shared:
- A tweet by
@nruaif
discussing a new development in AI modeling techniques. - A tweet shared by
@tokenbender
discussing Real SoTA work in function calling. - A tweet by
@weyaxi
discussing Marcoroni models. - Microsoftâs Phi-2 was mentioned in the discussion for reportedly outperforming Mistral 7b and Llama 7b on benchmarks.
- User
@benxh
has a dataset to upload which would lead to >100B high quality tokens if properly processed.
- A tweet by
â· #ask-about-llms (63 messagesđ„đ„):
-
Mistral Model Discussion: Users
@agcobra1
,@teknium
,@gabriel_syme
, and@giftedgummybee
discussed about various Mistral models, including base and instruct versions. -
Library for LLM-based Application: User
@coco.py
asked for advice on choosing an orchestration library between LangChain, Haystack, and Llama-Index for developing a LLM-based application. Users@.beowulfbr
and@decruz
expressed their problems with LangChain and recommended trying LLama-Index or building their own platforms. -
Fine-tuning Discussion: Users
@n8programs
,@teknium
,@youngphlo
, and@eas2535
discussed the potential of fine-tuning models on Apple Silicon and brought up some resources including Github discussion threads https://github.com/ggerganov/llama.cpp/issues/466 and Reddit posts Fine tuning on Apple Silicon and Fine tuning with GGML and Quantization. -
Benchmarking Models:
@brace1
asked for recommendations for benchmarking open-source models for a specific text extraction task.@night_w0lf
suggested models like Mamba, Mistral0.2, Mixtral, OpenHermes etc, and also directed to a leaderboard on Hugging Face.@n8programs
cautioned about the limitations of benchmarks and recommended checking the modelsâ ELO on another leaderboard. -
Parameter Impact on Speed: User
@n8programs
raised a query on why âstable diffusionâ, despite having around 3 billion parameters, was significantly slower than a 3 billion parameter LLM.
â· #memes (1 messages):
- Discussion on Microsoft Depiction: User
@fullstack6209
shared a humorous observation about working at Microsoft, also questioning the absence of a certain group in their representation.
DiscoResearch Discord Summary
-
Release of Mixtral and Model optimization: Mistral released Mixtral 8x7B, a high-performance SMoE model with open weights. The model was praised for its efficiency and speed, with
@aortega_cyberg
emphasizing that itâs 4 times faster and performs better for coding compared toGoliath
. Users also noted improvements inHermes 2.5
overHermes 2
. The availability and usage ofMistral-Medium
model were also brought up. Read the full release details and pricing in Mistralâs blog post. -
Quantization, Software Integration and Challenges: Discussions about quantization methods including
"INT3 quantization"
and"AWQ"
were highlighted by@vince62s
and@casper_ai
. A need for exllamav2 support for contexts over 4k was also mentioned, and software limitations in LM Studio and exllama were brought up by@aiui
. -
Model Efficiency and Function-Calling: The modelâs efficiency is theorized to originate from the âorthogonalityâ of its experts, as per
@fernando.fernandes
. A minimum-p routing system for expert evaluations was suggested which could enable scalable speed optimization. Strategies to mitigate similar problems and enrich Mixtral training data, such as including function-calling examples, were addressed with relevant resources shared, among them the Glaive Function-Calling v2 dataset on HuggingFace. -
Benchmarks, Scoring Models, and Emotional Intelligence: A new benchmark EQ-Bench, focusing on the emotional intelligence of LLMs, was introduced by
@.calytrix
who claimed its strong correlations with other benchmarks. Concerns about possible bias due to GPT-4 usage and potential improvements were discussed. GSM8K was recommended as a SOTA math benchmark. Score variations between mixtral instruct and other models on EQ-Bench stirred up conversations about potential causes. -
Resources and New Development: An explainer from Hugging Face on the Mixture of Experts (MoEs) model was shared here. A new model âYi_34B_Chat_2bitâ by Minami-su on HuggingFace was introduced here. Corrected pricing page link was shared here and Mistralâs Discord server link is here.
DiscoResearch Channel Summaries
â· #mixtral_implementation (170 messagesđ„đ„):
-
Mixtral Model Performance and Optimization: Users expressed a positive reception of the Mixtral modelâs performance and efficiency.
@Makya
stated that"Hermes 2.5"
performs better than"Hermes 2"
after adding code instruction examples. Regarding theMixtral-Instruct
model,@aortega_cyberg
highlighted that althoughGoliath
is better for creative writing, Mixtral performs better for coding and is 4x faster.@goldkron
pointed out repetition loops in Poeâs Mixtral implementation. -
Quantization and Software Integration: Participants discussed different quantization methods, including 1-bit and 3-bit options.
@vince62s
discussed the need for"INT3 quantization"
with"AWQ"
.@casper_ai
responded by stating that Mixtral is not yet ready for"AWQ"
but anticipated that it should offer faster performance than 12 tokens per second once it is. In the meantime,@vince62s
suggested that finetuning with six active experts might allow the Mixtral model to run on a single 24GB card. Furthermore, users discussed the need for exllamav2 support for contexts over 4k, with@aiui
pointing out current limitations in LM Studio and exllama. -
Importance of Experts in Model Efficiency:
@fernando.fernandes
theorized that the efficiency of Mixtral potentially results from the âorthogonalityâ of its experts, leading to maximal knowledge compression.@someone13574
proposed a minimum-p routing system where experts are evaluated based on their softmax score compared to the top expert. This would potentially enable scalable speed optimization by controlling the minimum-p level. -
Using Function Call Examples in Training:
@fernando.fernandes
suggested enriching the Mixtral training data with function-calling examples for diversity and to mitigate similar problems. He shared the Glaive Function-Calling v2 dataset on HuggingFace as a potential resource. Users also shared other potential datasets for this purpose. -
Availability of the Mistral-Medium Model: Users mentioned that the
Mistral-Medium
model is now available through the Mistral API. They speculated that this version could be around 70 billion parameters, though the modelâs exact size is as yet unconfirmed.
â· #general (9 messagesđ„):
- Mixtral Release and Pricing:
@bjoernp
shared the official announcement about the release of Mixtral 8x7B, a high-quality sparse mixture of experts model (SMoE) with open weights, by Mistral. He noted that it offers the fastest performance at the lowest priceâ up to 100 tokens/s for $0.0006/1K tokens on the Together Inference Platform. Its optimized versions are available for inferences on Togetherâs platform. Full details in Mistralâs blog post. - Mixture of Experts Explainer:
@nembal
shared a link to an explainer from HuggingFace about the Mixture of Experts (MoEs) model. The post dives into the building blocks of MoEs, training methods, and trade-offs to consider for inferences. Check out the full explainer here. - Pricing Page Issue:
@peterg0093
reported that the Mistralâs pricing page link responded with a 404 error. However,@_jp1_
provided the correct link to Mistralâs pricing information. - Mistral Discord Link:
@le_mess
asked for the Mistralâs Discord server link, which@nembal
successfully provided. Join their server here. - New Model by Minami-su:
@2012456373
shared a new model by Minami-su on HuggingFace named âYi_34B_Chat_2bitâ. The model is optimized to run on an 11GB memory GPU with a weights-only quantization method called QuIP# to achieve near fp16 performance using only 2 bits per weight. Detailed information can be found here.
â· #benchmark_dev (15 messagesđ„):
- EQ-Bench for Emotional Intelligence:
@.calytrix
introduced his paper and benchmark on emotional intelligence for LLMs, EQ-Bench, claiming that it has strong correlations with other benchmarks (r=0.97 correlation with MMLU, for example). Mentioning that EQ-Bench seems to be less sensitive to fine-tuning for scoring well on other benchmarks. Notably, the parameter size seems to count for more as nuanced understanding and interpretations required for emotional intelligence seem more strongly emergent at larger param sizes. He also shared the Paper Link and the Github Repository. - Potential Improvements and Limitations to EQ-Bench:
@onuralp.
suggested potential improvements including incorporating item response theory and reporting per-subject MMLU score correlations. The benefit of exploring model responses for well-structured scenarios involving agreeableness and negotiation was also suggested. A concern about potential bias due to choosing GPT-4 as the generator was brought up. - EQ-Bench creatorâs response:
@.calytrix
shared that the decision to use GPT-4 was based on resource constraints but confirmed that it did not generate the questions and answers. Defending the EQ-Bench, he argued that it genuinely measures some deep aspects of cognitive capabilities and offers useful discriminatory power. - Math Benchmark queries: In response to a request from
@mr.userbox020
for references about SOTA math benchmarks,@.calytrix
recommended GSM8K for its focus on reasoning over raw calculation. He also shared a Paper Link exploring the use of the inference space as a scratchpad for working out problems using left-to-right heuristics. - Score Disparities between models on EQ-Bench: Users
@bjoernp
,@rtyax
, and@.calytrix
engaged in a discussion about the lower scores mixtral instruct gets on EQ-Bench compared to its scores on other benchmarks. They showed curiosity about the disparity, suggesting possible reasons including a quick and dirty fine tune and inherent limitations of using 7b models as the base for MoE.
OpenAccess AI Collective (axolotl) Discord Summary
-
Engaged discussion around Mixtral, dealing with TypeError during loading mixtral via vllm, training on single A100 using qlora, LM Studioâs support, and code updates including support for Mixtral. A notable mention of a wandb project showing possible training set up was shared by
@le_mess
here. -
Technical dialogues surround Axolotl, including activating âoutput_router_logitsâ setting in the model configuration for auxiliary balancing loss, and issues with loading multiple datasets with hyphenated names.
-
Inquiry about current quantization technics led to the mention of Hugging Faceâs APIs, GenUtils and autogptq for choice-based inference.
-
Heartened interaction regarding training models, particularly around VRAM minimum requirements for training and fine-tuning AI models, issues and solutions related to stuck Mixtral training, discovering and sharing of wandb project results. Users shared their experiences while setting and solving problems in Axolotl setup and debugging with Docker, and discussed issues, views and workarounds for Axolotl training issues related to huggingface/transformers. A question related to loss spikes in pretraining remains open, without a common consensus.
-
Dataset-specific chat involved the potentials and cons of training AI on YouTube debate transcripts, interest in the local search repository, personal recommendation and experiences with tools for PDF conversion to markdown, specifically mentioning Marker. The discourse broadened to alternative solutions for document processing, but budget constraints negated their use for some users.
OpenAccess AI Collective (axolotl) Channel Summaries
â· #general (48 messagesđ„):
- Issues loading mixtral via vllm:
@papr_airplane
reported TypeError when trying to load mixtral via vllm using âptâ format.@caseus_
announced the official mixtral modeling code release now supports mixtral. - Impact of Rope on Training: User
@kaltcit
sparked a discussion on the effect of enabling rope on VRAM usage and training sequences.@nanobitz
clarified that rope itself doesnât affect VRAM, but increasing the sequence length will. - Loading mixtral gguf onto LM studio:
@papr_airplane
asked if anyone has loaded mixtral gguf onto LM studio, but@nruaif
replied that LM studio doesnât support it yet, while@faldore
claimed to achieve it using ollama. - Training Mixtral on a single a100 using qlora:
@le_mess
shared that training Mixtral using qlora on a single a100 seems possible, linking to a wandb project. - Quality of Life Improvement in Docker Images:
@caseus_
announced a new update addingevals_per_epoch
andsaves_per_epoch
parameters that would be available in docker images soon. The update is designed to support quality of life improvement by eliminating the need to calculate total steps or back-calculate from the number of total epochs.
â· #axolotl-dev (11 messagesđ„):
- Updated Modeling Code:
@bjoernp
mentioned that the updated modeling code for Mixtral has been pushed. - Auxiliary Balancing Loss for the Router:
@bjoernp
inquired about the activation of the auxiliary balancing loss for the router. According to@caseus_
, enabling this feature is done through the configuration.@caseus_
provided a sample code snippet, which indicates that it should be possible to enable this feature in Axolotl by adding"output_router_logits: true"
undermodel_config
. - Loading Multiple Datasets:
@le_mess
reported an issue with loading multiple datasets that have-
in their names, and asked if anyone else had encountered this problem. According to@nanobitz
and@noobmaster29
, the loading should work fine if using quotes or if they have not updated to the latest mixtral update for Axolotl.
â· #other-llms (3 messages):
- Discussion on Quantization Techniques: User
@jovial_lynx_74856
asked about the current technology being used for quantization and inference, suggesting Hugging Faceâs APIs and GenUtils. In response,@nanobitz
mentioned autogptq for quantization, along with advising the choice-based inference.
â· #general-help (70 messagesđ„đ„):
-
Training Models with Different VRAMs: Users discussed about the minimum requirements of VRAM for training AI models.
@nruaif
mentioned that the minimum is 1 A6000, which equates to 48 GB of VRAM.@gameveloster
asked if 24 GB or 48 GB VRAM is enough to fine-tune a 34b model, to which@le_mess
confirmed that 48 GB should be enough, but there could be problems with 24 GB. -
Mixtral Training:
@le_mess
shared issues about stuck Mixtral training with no drop in loss. Other users suggested enablingoutput_router_logits
, increasing learning rate by a factor of 10, and provided references to similar training issues.@le_mess
tracked his experiment using Weights & Biases. Link to result 1, Link to result 2.@caseus_
suggested reference training with higher learning rate Link -
Axolotl Setup and Debugging with Docker:
@jovial_lynx_74856
posted about issues encountered while setting up Axolotl with CUDA version 12.0 on an 80 GB A100 server and running the example command.@le_mess
suggested running the setup in a Docker environment on Ubuntu and CentOS. -
Axolotl Training Issues:
@jovial_lynx_74856
reported errors encountered during Axolotl training.@caseus_
offered some insights and shared a link for reference on relevant changes introduced by huggingface/transformers Link. The issue seems to down to falling back to SDPA (Scaled Dot Product Attention) instead of the eager version and@jovial_lynx_74856
found a workaround by settingsample_packing
tofalse
in the YML file. -
Loss Spike in Pretraining:
@jinwon_k
asked about loss spikes in pretraining. The users did not provide solutions or pointers. This question remains open.
â· #datasets (20 messagesđ„):
- Training AI on Debate YouTube Transcripts:
@papr_airplane
suggested training AI on YouTube transcripts of debate competitions to improve reasoning skills. However,@faldore
countered that it risks teaching the AI to always argue with the user. - Access to Local Search Repository:
@emrgnt_cmplxty
offered to give early access to a new repository after@papr_airplane
expressed interest. - Discussion on PDF to Markdown Conversion Tools:
@visuallyadequate
shared personal experiences with multiple tools and libraries for PDF to markdown conversion. They shared the link to Marker as a strong candidate, although noted its limitations with tables and tables of contents. They concluded some PDFs fare better than others while being converted. - Alternative Solutions for Document Processing: Other users suggested alternatives for document processing.
@lightningralf
mentioned using Llama Index Hub, while@joshuasundance
mentioned using Azure Document Intelligence with Langchain. However, these alternatives werenât very suitable for@visuallyadequate
given budget constraints.
HuggingFace Discord Discord Summary
- Technical Outage: A reported issue with all models not working, acknowledged by user
@harrison_2k
. - Query about starting study in NLP with a suggested exploration into transformer models as an initial step.
- Question about fine-tuning BERT Model on a Turkish Q/A dataset for better performance and accuracy.
- Conversation on creating an EA for XMâs standard account for FX Automated trading and speculations about it.
- Users participated in a discussion on a Neural Network Distillation paper, with a link to the paper provided in both the general and reading group channels, supporting the exchange of ideas and further understandings.
- Discussion around the runtime error in the Nougat Transformers space with a suggestion to try running it in a different space provided by user
@osanseviero
. - Introduction of a new model âDeepsex-34bâ by user
@databoosee_55130
and several shared links associated with the model. - Shared interest in the value of an Object, Property, Value list (OPV) in the digital field.
- Concern about the layout of the required dataset for a research paper in the field.
- Sharing of an article on a neuromorphic supercomputer that simulates the human brain.
- Promotion of Learning Machine Learning for Mathematicians paper for more mathematicians to apply machine learning techniques.
- Presentation of a variety of educational videos and articles, including a Retrieval Augmented Generation (RAG) video, XTTS2 Installation guide video, and a breakthrough in language models - MOE-8x7b.
- Sharing of newly created models on HuggingFace, such as the SetFit Polarity Model with sentence-transformers/paraphrase-mpnet-base-v2 and the Go Bruins V2 Model, with requests for guides on how to create similar models like the ABSA Model.
- Discussions primarily held on Wednesdays in the Reading Group channel in addition to the introduction to the paper Distilling the Knowledge in a Neural Network.
- Inquiry on creating separate bounding boxes for each news article in a newspaper.
- Discussion around the best Encoder/Decoder for a RAG system, with BERT being the current model in use, and a high interest in knowing more about model rankings for multilingual or code-specific use cases. The Palm model is frequently recommended.
HuggingFace Discord Channel Summaries
â· #general (21 messagesđ„):
- Fault in Models:
@noone346
reported that all models appear to be down, a concern that was acknowledged by@harrison_2k
. - Initiating NLP Study:
@bonashio
sought advice on starting with NLP, already having a basic understanding of Deep Learning.@jskafer
suggested an exploration into transformer models first. - Fine-tuning BERT Model on Turkish Q-A:
@emirdagli.
asked for suggestions on fine-tuning a changeable BERT model on Turkish Q-A with a dataset of ~800 questions and ~60-70 answers. - Inquiry about MT4 and FX Automated Trading:
@ronny28717797
mentioned their intent to create an EA for performing on XMâs standard account, detailing specific performance expectations and preferences. - Discussion on Distillation Paper:
@murtazanazir
initiated a discussion on a distillation paper, providing a link to the paper on arXiv. - Nougat Transformers Space Runtime Error:
@pier1337
reported a runtime error on@697163495170375891
âs Nougat Transformers space, asking if it can be run on a 16GB RAM CPU only.@osanseviero
provided a link to try it in a different space. - Introduction of Deepsex-34b Model:
@databoosee_55130
introduced a new model, âDeepsex-34bâ and shared multiple links related to it. - Object, Property, Value (OPV) List:
@noonething
sparked a discussion by questioning the value of an Object, Property, Value list in the field. - Dataset Layout Query:
@largedick
asked about the layout of a dataset required for a paper. - Neuromorphic Supercomputer Article Share:
@bread browser
shared a New Scientist article about a supercomputer that simulates the entire human brain.
â· #today-im-learning (1 messages):
- Machine Learning for Mathematicians:
@alexmath
shared a link to an arXiv paper encouraging more mathematicians to learn to apply machine learning techniques.
â· #cool-finds (3 messages):
- Retrieval Augmented Generation (RAG) Video: User
@abelcorreadias
shared a video from Umar Jamil that explains the entire Retrieval Augmented Generation pipeline. - XTTS Installation Guide:
@devspot
posted a link to a YouTube video that explains how to install XTTS2, a popular Text-To-Speech AI model, locally using Python. - Breakthrough in Language Models - MOE-8x7b:
@tokey72420
shared a link to an article on MarkTechPost about Mistral AIâs new breakthrough in language models with its MOE-8x7b release.
â· #i-made-this (5 messages):
- SetFit Polarity Model Creation:
@andysingal
shared the link to his newly created model on HuggingFace, the SetFit Polarity Model with sentence-transformers/paraphrase-mpnet-base-v2, and encouraged others to try Aspect-Based Sentiment Analysis (ABSA) with it. - Guide Request:
@fredipy
asked for a guide on how to create models similar to@andysingal
âs ABSA Model. - Go Bruins V2 Model Showcase:
@rwitz_
showcased his fine-tuned language model, Go Bruins V2, on the HuggingFace platform.
â· #reading-group (3 messages):
- Meeting Schedules:
@chad_in_the_house
confirmed that meetings usually occur on Wednesdays and are often held in a dedicated thread. - Discussion on Distillation Paper:
@murtazanazir
initiated a discussion on the paper Distilling the Knowledge in a Neural Network, inviting others to join in for an elaborate understanding.
â· #computer-vision (2 messages):
- Bounding Box for Newspaper Segmentation: User
@navinaananthan
asked for suggestions on models or methodologies to create a separate bounding box for each news article when uploading a newspaper. The user didnât get answers yet.
â· #NLP (5 messages):
- Best Encoder/Decoder for RAG System: User
@woodenrobot
inquired about the most suitable encoder/decoder for a RAG system. They currently use BERT, but are open to other suggestions as they are unsure about BERTâs ability to scale up as data grows over time. - Rankings for Multilingual/Code:
@woodenrobot
expressed an interest in seeing model rankings specifically for multilingual/code use cases. - Palm as a Top Model Suggestion:
@woodenrobot
noted, based on interactions with bard, that Palm appeared frequently as the recommended model.
LangChain AI Discord Summary
- A new architecture for LangChain, namely
langchain-core
andlangchain-community
, was introduced by@hwchase17
. This updated architecture will facilitate the development of context-aware reasoning applications using LLMs. Questions about this change are welcomed. - Various technical inquiries and issues:
- New techniques for tracking token consumption per LLM call and improving SQL query result summarization were requested.
- An issue with LangSmith API maintaining a log of LLM calls over the last 45 minutes was reported.
- Community members sought advice on using LangChain for specific task completion, employing open-source alternatives for OpenAIEmbeddings, and enhancing chatbots to show similar records from a database.
- Discussion about using LangChain to improve Azure Search Serviceâs multi-query retrieval and fusion RAG capabilities.
- In the langserve channel:
- Users faced challenges with callback managers in Llama.cpp and the integration of langserve, langchain, and Hugging Face pipelines.
- A proposal was made to create an issue about challenges with Langserve and possible adjustments to use RunnableGenerators over RunnableLambdas.
- A request for exposing the
agent_executor
in langserve.
- A free 9-part code-along series on building AI systems using OpenAI & LangChain was shared. The first part specifically covers sentiment analysis with GPT and LangChain, MRKL prompts, and building a simple AI agent. The course can be found on this link and a corresponding YouTube Session is also available.
LangChain AI Channel Summaries
â· #announcements (1 messages):
- New LangChain Architecture:
@hwchase17
shared a blog post detailing the newlangchain-core
andlangchain-community
. The goal of this re-architecture is to make the development of context-aware reasoning applications with LLMs easier. This change is a response to LangChainâs growth and community feedback, and implemented in a completely backwards-compatible way.@hwchase17
also offered to answer questions regarding this change.
â· #general (12 messagesđ„):
-
Tracking Token Consumption Per LLM Call:
@daii3696
inquired about a way to track token consumption for each LLM call in JavaScript. In response,@pferioli
suggested using a callback and provided a code snippet, dealing specifically with thehandleLLMEnd
callback, which logs the token usage details. -
Improve SQL Query Result Summarization:
@jose_55264
asked about possible methods to expedite the summarization of results from SQL query execution. No specific solutions were provided in the given messages. -
Using LangChain for Task Completion:
@sampson7786
expressed a desire to utilize LangChain for completing a task with a specific procedure, seeking assistance on the platform for this issue. -
LangSmith API Issues:
@daii3696
raised concerns about an apparent issue with the LangSmith API as they were unable to trace their LLM calls for the past 45 minutes. -
Open Source Alternatives for OpenAIEmbeddings in YouTube Assistant Project:
@infinityexists.
is working on a YouTube assistant project and asked how HuggingFace can be used as an open-source alternative to OpenAIEmbeddings. They provided a link to the GitHub code used in their project. -
Increasing Odds in Vector Search:
@rez0
queried about the name of a function that splits retrieval into three queries in vector search to improve the chances of getting the wanted results. -
Enhanced Search Capabilities in Chatbot:
@hamza_sarwar_
is interested in enhancing their chatbotâs capabilities to display similar records from a database (e.g., vehicle info) when there are no precise matches to a userâs query. -
Azure Search Service Multi-query Retrieval/Fusion RAG with LangChain:
@hyder7605
is working with Azure Search Service and wishes to integrate multi-query retrieval or fusion RAG abilities with Azure Cognitive Search using LangChain. They also aim to include advanced features like hybrid and semantic search in their query process but are unsure about defining filters and search parameters with LangChain.
â· #langserve (5 messages):
- Issues with Callback Managers in Llama.cpp: User
@jbdzd
raised a challenge with callback managers in Llama.cpp during their use of langserve. They got a RuntimeWarning indicating a coroutine âAsyncCallbackManagerForLLMRun.on_llm_new_tokenâ was never awaited. - Struggles with Implementing Streaming in Langserve:
@fortune1813
expressed trouble with the integration of langserve, langchain, and hugging face pipelines, especially with respect to streaming. They have investigated the notebooks but requested further clarification on proper streaming implementation. - Proposal to Create Issue in Langserve:
@veryboldbagel
suggested that@fortune1813
create an issue about their challenge in Langserve and share the full server script. They also brought to attention that RunnableGenerators should be used instead of RunnableLambdas for common operations, noting that it has been poorly documented. - Request for Agent Executor Exposure in Langserve:
@robertoshimizu
queried about exposingagent_executor
in langserve and shared a code snippet with an example. However, they are struggling as the input seems different when invoked in a Python script.
â· #tutorials (1 messages):
- AI Code-Along Series: User
@datarhys
shared a free 9-part code-along series on AI, with a focus on âBuilding AI Systems with OpenAI & LangChainâ, which was released by DataCamp. This series aims to guide learners from basic to more advanced topics in AI and LangChain.- The first code-along covers:
- Performing sentiment analysis with GPT and LangChain
- Learning about MRKL prompts used to help LLMs reason
- Building a simple AI agent
- The first code-along covers:
- The instructor, Olivier Mertens, is praised for his ability to present complex topics in an accessible manner.
- To start this code-along, follow the link provided: Code Along
- A YouTube session for the code-along is also available: YouTube Session
Alignment Lab AI Discord Summary
- Conversation about a Mixtral-based OpenOrca Test initiated by
@lightningralf
, with the reference to a related fxtwitter post from the OpenOrcaâs development team. - Speculation on the speed of the machine learning process, proposed solution includes using server 72 8h100 to enhance performance.
@teknium's
declaration of testing an unidentified model and the need for further clarification of the said model.- Inquiry from
@mister_poodle
on ways to extend or fine-tune Mistral-OpenOrca for specific tasks, namely boosting NER task performance using their datasets and generating JSON outputs.
Alignment Lab AI Channel Summaries
â· #oo (5 messages):
- Discussion about a Mixtral-based OpenOrca Test:
@lightningralf
asked@387972437901312000
if they tested Mixtral based on OpenOrca, linking a fxtwitter post - Question about Process Speed:
@nanobitz
expressed surprise about the speed of the process, with@lightningralf
suggesting the use of server 72 8h100. - Unidentified Model Testing:
@teknium
mentioned testing some model, but being uncertain about which one.
â· #open-orca-community-chat (1 messages):
- Extending/Fine-tuning Mistral-OpenOrca for Specific Tasks: User
@mister_poodle
expressed interest in using their datasets to boost Mistral-OpenOrcaâs performance on an NER task with JSON outputs. They sought examples or suggestions for extending or fine-tuning Mistral-OpenOrca to achieve this goal.
Latent Space Discord Summary
- Discussion about fine-tuning open source models with recommendations for platforms such as Replicate.ai with the ability to run models with a single line of code shared by
@henriqueln7
. - Introduction of an all-in-one LLMOps solution, Agenta-AI, including prompt management, evaluation, human feedback, and deployment by
@swizec
. - Recognition of the popularity of daily AI newsletters highlighted through a Twitter post by Andrej Karpathy shared by
@swyxio
. - Inquiry about the ideal size of language models for RAG applications by
@henriqueln7
, focusing the debate on balancing world knowledge and reasoning capacity. - Sharing of a YouTube video presenting insights into the future of AI, titled âAI and Everything Else - Benedict Evans | Slush 2023â by
@stealthgnome
. - Request for access to Smol Newsletter Discord text and API made by
@yikesawjeez
with the goal of displaying content as embeds or creating a daily digest. - Development progress update by
@yikesawjeez
on a SK plugin, planning to complete this before catching up on backlog items.
Latent Space Channel Summaries
â· #ai-general-chat (11 messagesđ„):
-
Fine-tuning Open Source Models: User
@henriqueln7
asked for recommendations on platforms to fine-tune open source models and presented https://replicate.ai/ as a suggestion. This platform allows one to run open-source models with one single line of code. -
All-in-one LLMOps platform: User
@swizec
shared a link to the GitHub repository for Agenta-AI. This platform provides an all-in-one LLMOps solution, including prompt management, evaluation, human feedback, and deployment. -
Daily AI newsletters:
@swyxio
commented on the popularity of daily AI newsletters, putting forward a Twitter post by Andrej Karpathy as an example. -
Size of Language Models for RAG Applications:
@henriqueln7
proposed a question about the ideal size of language models for RAG Applications. The question aimed to clarify whether a smaller or larger model would be better, considering the smaller modelâs limited world knowledge and the larger modelâs superior reasoning abilities. -
AI Overview for 2023: User
@stealthgnome
shared a YouTube link to the âAI and Everything Else - Benedict Evans | Slush 2023â video, offering insights into the future of AI.
â· #llm-paper-club (2 messages):
- Requesting Access to Smol Newsletter Discord Text and API: User
@yikesawjeez
requests@272654283919458306
for an API to access the raw text or .md files from the smol newsletter discord. An API would enable them to display the content as embeds using another user, Red, or create a daily digest. - Development on SK Plugin:
@yikesawjeez
is currently working on a sk plugin and plans on finishing that before going through their backlog.
LLM Perf Enthusiasts AI Discord Summary
- Conversation around Anthropic Fundraising with user
@res6969
mentioning a rumor about Anthropic raising an additional $3 billion, and jesting about function calling issues, suggesting that more funds might solve them. - Discussion about Fine-tuning for Email Parsing initiated by
@robhaisfield
, regarding the number of examples needed for creating a JSONL file to parse email strings into structured objects. - Share by
@robotums
of a Microsoft Research blog article about Phi 2: The Surprising Power of Small Language Models. - User
@lhl
shared their experience with Inference Engine Performance, claiming a 50X performance boost after replacing some code with vLLM. They also compared transformers with other engines and shared a GitHub repository containing detailed results. - Dialogue on prompting techniques including the MedPrompt method and DeepMindâs Uncertainty Routed CoT (Cooperative Output Transformations). This discussion also touched on OCR (Optical Character Recognition) usage and MS Researchâs achievements. All topics were introduced and discussed by users
@robotums
and@dhruv1
.
LLM Perf Enthusiasts AI Channel Summaries
â· #claude (3 messages):
- Anthropic Fundraising: User
@res6969
commented about a rumor that Anthropic is raising an additional $3 billion. - Function Calling Issues:
@res6969
made a jesting remark that maybe a few more billion dollars will get function calling to function properly.
â· #finetuning (2 messages):
- Fine-tuning for Email Parsing: User
@robhaisfield
queried about the number of examples needed for creating a JSONL file to fine-tune a model to parse an email string into a structured object with nested replies. They asked if 30 examples would be sufficient.
â· #opensource (1 messages):
- Phi 2: The Surprising Power of Small Language Models:
@robotums
shared a link to a Microsoft Research blog article addressing the potential of smaller language models, accompanied by a list of contributors to the content.
â· #speed (1 messages):
- Inference Engine Performance: User
@lhl
detailed their experience with different inference engines. They noted that after replacing some existing code with vLLM, they observed a 50X performance boost. They also compared various inference options from transformers with other engines and shared their findings through a GitHub repository. The repo contains detailed results of the inferencing tests.
â· #prompting (5 messages):
- Understanding When to Use OCR:
@robotums
expressed curiosity about how to determine when OCR (Optical Character Recognition) is necessary for a page, asking if anyone knows how chatDOC accomplishes it. - Inquiring About the MedPrompt Technique:
@dhruv1
asked if anyone has used the MedPrompt technique in their application. - Achievement by MS Research Using MedPrompt:
@dhruv1
shared that MS Research has written a post about using the MedPrompt technique to surpass Geminiâs performance on MMLU (Multiple-choice Machine Learning dataset from University of Minnesota). - DeepMindâs Uncertainty Routed CoT Technique:
@dhruv1
informed the channel that DeepMind has revealed a new technique called the uncertainty routed CoT (Cooperative Output Transformations) that outperforms GPT on MMLU. - Sharing the CoT Technique:
@dhruv1
promised to share more about the CoT technique.
AI Engineer Foundation Discord Summary
Only 1 channel had activity, so no need to summarizeâŠ
- Event Timing Mix-up:
@yikesawjeez
inquired about the eventâs time being marked at 9:30 PM, which they had expected to be at 8:00 AM.@._z
clarified that the event is actually scheduled to be at 9:30 AM PST. There was a minor confusion due to the time being marked incorrectly as PM instead of AM.
Skunkworks AI Discord Summary
- Mention of an individualâs influence on mistral licensing in the #moe-main channel, with a comment from user â.mrfooâ: âInfluencing mistral licensing I see. Nice!â.
- In the #off-topic channel, user âpradeep1148â shared a YouTube link: https://www.youtube.com/watch?v=wjBsjcgsjGg. The content of the video was not discussed.
Skunkworks AI Channel Summaries
â· #moe-main (1 messages):
.mrfoo: <@1117586410774470818> : Influencing mistral licensing I see. Nice!
â· #off-topic (1 messages):
pradeep1148: https://www.youtube.com/watch?v=wjBsjcgsjGg
MLOps @Chipro Discord Summary
Only 1 channel had activity, so no need to summarizeâŠ
- The Complete Data Product Lifecycle:
@viv2668
shared links to two parts of an enhanced end-to-end guide on practical Data Products.
The Ontocord (MDEL discord) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The Perplexity AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The YAIG (a16z Infra) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.