swyx
[TOC]
Nous Research AI Discord Summary
- Members expressed interest in attending NeurIPS and meeting up, with suggestions for future AI events in Australia.
@richardblythmanurged those interested in an open-source, decentralized AI project to reach out to them. Users shared their projects, like@cyborg_1552âs photo GPT AI tool and@pradeep1148âs introduction of StableLM Zephyr 3B. - User
@gabriel_symetriggered interest around Mixtral by sharing a GitHub link. Performance comparisons between Mixtral and GPT-3.5 heated discussions.@mihai4256unveiled their fine-tuned model, Pallas-0.2, available on Hugging Face. A [Youtube video](https://youtu.be/y9k-U9AuDeM? si=2X5j64_cdsdKwWEw) discussing open-source LLMs usage sparked brief reactions. - Both
OpenHermes-2.5-neural-chat-v3-3-Slerpand Mixtral were the topic of hype for their performances, with debating dictates on the latterâs GPU requirement. Tools such asTensorboard,Wandb,evalplus,llamahubwere stated beneficial for fine-tuning and evaluating models. User experiences on model hosting platforms like Ollama and LM Studio were exchanged with contrasting opinions favoring both. - A robust conversation on MoE led by
@gabriel_symeclarified why Mixtral, a model based on Mistral, is set apart from previous implementations. Discussions on fine-tuning LLMs suggested limited data requirements. The potential of Mixtral being competitive with GPT-4 after finetuning was proposed.@wlrdexplained how open-source LLMs could be implemented, leading to the OpenHermes 2.5 - Mistral 7B model. Speculations on GPT-3.5 suggested itâs a 20B model and forecasted its soon open-source release. Inference optimization possibilities for ChatGPT touched on strategic batching, potential caching, and user base size. - The memes channel saw an array of emojis and memes shared by members for fun and communication. Specific interests in speakers like Yann and Karpathy were expressed. User
@tekniumamusingly delineated a character as being heavily concerned about x risk.
Nous Research AI Channel Summaries
â· #off-topic (19 messagesđ„):
-
NeurIPS Meetup:
@blue_matchaasked if anyone was attending NeurIPS in the hopes of meeting up, and@tekniumsaid they might be available on a Thursday and Friday.@gabriel_symeexpressed disappointment in NeurIPS consistently being in the US, later revealing they are based in Australia.@gabriel_symealso proposed hosting an event in Australia the following year.
-
Open Source and Decentralized AI Co-founder Search:
@richardblythmanis in search of a co-founder for a project in the open-source and decentralized AI space and asked anyone interested to DM them.
-
Interest in Australian AI Conferences:
@deki04pointed out that there would be considerable interest in Australian-based AI conferences, recounting a well-attended in-person fastAI course held in Brisbane led by Jeremy Howard.
-
Photo GPT AI Development:
-
Introduction of StableLM Zephyr 3B:
@pradeep1148shared a YouTube video introducing StableLM Zephyr 3B, a large language model.
â· #benchmarks-log (1 messages):
nonameusr: i think he used markdown
â· #interesting-links (24 messagesđ„):
-
Discussion about Mixtral and its Architecture:
@gabriel_symeshared a GitHub link to MixtralKit â a toolkit for themixtral-8x7b-32kseqlenmodel.@cyborgdreamposted a twitter link, sharing that Mixtral outperforms GPT-3.5 in benchmarks even before fine-tuning. The subsequent discussion involved@nonameusr,@euclaise, and@chhilleedebating the benefits and uniqueness of Mixtralâs Transformer-based architecture. -
Release of New Fine-Tuned Model:
@mihai4256announced the release of their fine-tuned model, Pallas-0.2, hosted on Hugging Face. This model, a fine-tune ofTess-34B-v1.4, is designed for reasoning tasks and performs well with long system prompts. -
Video about Open Source LLMs usage:
@tekniumshared a Youtube video answering the question âShould You Use Open Source Large Language Models?â@n8programsand@nonameusrgave one-word responses to the question, with conflicting opinions.
â· #general (639 messagesđ„đ„đ„):
-
Fine-tuning and Performance Discussions: Users discuss the fine-tuning and performance of several models, including Hermes 2.5, Mistral, and GPTs Agent. For instance,
@nonameusrsuggests thatOpenHermes-2.5-neural-chat-v3-3-Slerp(also nicknamed âSlurpyâ) outperforms the originalHermesin some regards but notes inconsistencies. Several users also discuss the performance ofMixtral(orMixtral MoE), discussing topics like its GPU requirements and its behavior when quantized. -
Model Hosting and Management Platforms: Multiple users compare their experiences using Ollama and LM Studio for hosting and managing AI models. While some users express a preference for Ollama, others point out that LM Studio may be more customizable and better support a wider range of models.
-
Compute and Training Resources: Users like
@vatsadevand@gabriel_symediscuss their computing resources, with the discussion also touching on the potential of university resources. -
Useful Tools: Discussion also touched on various tools like
Tensorboard,Wandb,evalplus, andllamahub, which can be useful for fine-tuning, testing, and evaluating models. -
New Models and Techniques: The channel saw mentions of new models and techniques, like âslerpâ (in the context of
OpenHermes-2.5-neural-chat-v3-3-Slerp). Some users also speculate about theMixtralandStripedHyenamodels and the potential for further improvements to them via fine-tuning or merging strategies. Finally,@ldjsuggests thatMixtralâs method of choosing âexpertsâ during its computation could influence its performance.
â· #ask-about-llms (123 messagesđ„đ„):
- Mixture of Experts (MoE) Discussion: Users
@akhxl,@cyborgdream, and@gabriel_symeengaged in a conversation about MoE, with@akhxlinitially expressing confusion about the sudden hype over a technique thatâs been around for some time.@gabriel_symeprovided an explanation, stating that previous implementations didnât yield useful models and that Mixtral, based on Mistral, has shown practical utility. - Finetuning Large Language Models (LLMs): In a dialogue involving
@akhxland@gabriel_syme, clarifications about the amount of data needed for finetuning were offered.@gabriel_symenoted that recent advancements didnât require substantial data to finetune a good model due to the quality of base models and expansive pretraining data availability. A discourse on the potential of Mixtral to perform comparably to GPT-4 after finetuning ensued with@cyborgdreampredicting such an outcome. - Open Source LLMs Usage:
@.plotand@wlrdheld a conversation regarding the acquisition and implementation of open-source LLMs.@wlrdpointed out that the modelsâ weights are open-sourced and can be fetched from Hugging Face and gave an example link to the OpenHermes 2.5 - Mistral 7B model. - GPT-3.5 Turbo Discussion: A nuanced discussion over the GPT-3.5 Turbo specifications occurred, primarily involving
@cyborgdream,@agcobra1, and@n8programs. The discourse ranged from its performance compared to both smaller and larger models, with@cyborgdreamsuggesting the model is possibly a 20B model, basing on the leaked G3PO information and predicting its open-source release soon. - Inference Optimization for ChatGPT: User
@zohad_sikderinitiated a conversation regarding potential optimizations for faster inference in ChatGPT. Speculations from@teknium,@bjoernp,@eas2535and@skadeskotenranged from the unlikely use of quantization to strategic batching and potential caching for frequently asked questions. The fast response time of ChatGPT was discussed, with@zohad_sikderhypothesizing a robust caching mechanism due to the substantial user base.
â· #memes (10 messagesđ„):
- Meme Sharing and Reactions: Users in this channel, namely
@tekniumand@Error.PDF, frequently share emojis and meme reactions. Notable mentions include the âY not bothâ and <:pepeshy:1151280286345207819> emojis. - Desire for Certain Speakers:
@tekniumexpressed a desire for individuals such as Yann and Karpathy to speak, leading to responses and discussions among the users. - Character Evaluation:
@tekniumexpressed their opinion on an unidentified individual, characterizing them as âcrazy psycho about x riskâ.
OpenAI Discord Summary
- An ongoing discussion centered around the topic of AI bias, morality, and fair use in the context of copyrighted content and AI. Conversations delved into issues such as biases in large language models (LLMs) and the philosophy of truth, alongside speculations surrounding Googleâs new AI, Gemini, and alternative AI technology options like Mistral Instruct and gpt4all.
- Members engaged in various technical discussions regarding GPT-4, touching upon âDynamic Limitsâ, waitlist duration, prefix prompt exploration, ChatGPTâs performance and access issues, and differences in features across various devices. Speculations were made about the development of GPT-5 and the opening of GPT Store in the new year.
- Issues with and improvements for GPT usage have been a pressing topic, with dissatisfaction expressed over the dialogue summarization by GPT, missing features in GPT Builder, and the absence of a feature allowing Inline editing or trim for AI responses. A parallel conversation took place regarding the acquisition of developer access for ChatGPT plugins, clarification of OpenAIâs Terms of Service, and the need for comprehensive guides on custom GPTs.
- Conversations about game development using GPT and chatbot performance indicated a healthy interest in the potential applications of AI technology. Issues with captcha during API key generation, searching specific conversations, and perceived changes in GPT output fueled the debate on current limitations and areas for improvement in the AI system.
- A notable topic in the guild was prompt engineering, digging deep into the usage of emotional language and the implementation of personalities in PPM. The community also dived into issues concerning text chunking, embeddings, and creation of detailed prompts. The sharing of a series of detailed prompt guidelines and command protocols for GPT-4, dalle, and browser tools reflected collaborative efforts to enhance utilisation of the AI model.
OpenAI Channel Summaries
â· #ai-discussions (123 messagesđ„đ„):
- Discussion on AI bias and morality: Users
@whynot66k20ni,@light.grey.labs,@solbus,@lhc1921engaged in a deep conversation regarding the inherent nature of biases in large language models (LLMs), the philosophy of truth, and the potential self-awareness of AIs. - ChatGPTâs AI ethics and âfair useâ:
@.dooz,@lhc1921,@light.grey.labsdiscussed the âfair useâ in the context of copyrighted content and AI..doozsuggested that transformative use of copyrighted content could be constituted as fair use. - Discussion about OpenAIâs GPT Store release:
@lumirixshared an excerpt from an email received by GPT creators promising the release of the GPT Store early next year and promising other great updates to ChatGPT. - Alternatives to OpenAI ChatGPT:
@mysticmarks1recommended Mistral Instruct and gpt4all as alternatives or additions to OpenAIâs ChatGPT for@sneakobrahwho was seeking alternative chat AIs. - Discussion on Googleâs AI Gemini:
@prajwal_345shared a link about Googleâs Gemini AI suggesting that it was announced under pressure, and it outperformed OpenAIâs GPT-4 on various benchmarks.
â· #openai-chatter (112 messagesđ„đ„):
-
GPT-4 Dynamic Limits and Waitlist Discussion:
@dr.youvi.avantasked about the new GPT-4 âDynamic-Limitsâ.@stefatorusmentioned that unlocking older GPT versions is possible but can be expensive, with his usage amounting to approximately 200 EUR per month.@killer.5643inquired about the GPT-4 waitlist duration, with@7877mentioning the upcoming GPT Store, and@jonathan_91672sharing that he waited about a month for his invitation. -
GPT-4 Prefix Prompt Exploration:
@israel_a4shared a YouTube tip from Wes Roth which allows users to see GPT-4âs Prefix or Secret Prompt by using a certain code. When asked about a potential patch to prevent this,@elektronisadestated that no such plans were in place due to the inherent functioning of the models. -
ChatGPT Performance and Access Issues: Several users reported issues with ChatGPT, with
@mrcrack_mentioning consistent network errors and ADAâs ineffective image reading.@zz99mzmentioned the issue of the domain not loading at all.@pruoindicated trouble with their custom instructions, and@mrcrack_also voiced dissatisfaction with the dynamic limits. -
Features in Different Devices:
@gd2xinquired about the speech feature absence in the Android version of ChatGPT, which@elektronisadeattributed to the use of an adblocker. A discrepancy between features available in Android and iOS versions was also discussed. -
GPT-3 Extensions and GPT Store Speculations:
@youraveragedevspeculated about GPT-5âs development, but@clockrelativity2003denied its current training. A discussion about GPT Storeâs opening in the new year was held by@lugui.
â· #openai-questions (158 messagesđ„đ„):
- Issues and Improvements in GPT: User
@stealth2077expressed concerns about GPT ending dialogues with a concluding summary paragraph, even after providing explicit instructions not to do so.@stealth2077has also proposed a feature for inline editing or trim for AI responses for easier control over generated conversations, a topic joined by@ath0rus.@stealth2077voiced dissatisfaction over the reduction of GPT usage from 50 to 40 and the removal of additional 10 usages reserved for custom GPT testing. - GPT Builder Limitations:
@amanshresthaexperienced issues in GPT Builder, which seemed to stem from the Python environment.@stealth2077also expressed frustration over the restrictions in changing custom instructions mid-chat, and he highlighted the need for a better functionality to edit a chatâs context. - ChatGPT Plugins:
@keebs1995inquired about gaining developer access for ChatGPT plugins for building a calculator app for their industry.@elektronisadeinformed that plugins are being phased out and suggested using custom GPTs instead. - Terms of Service (ToS) Clarifications: User
@eric.turnrsought elaboration on the OpenAI ToS section mentioning âAutomatically or Programmatically extract data or Output (defined below).â@lumirixclarified that âOutputâ is defined in the Content section of the ToS. - Performance Issues & Enhancements: A few users, including
@Shunraiand@lucianah, reported lagging and network error issues with GPT.@Rockasked for comprehensive guides about the workings of custom GPTs, and@strange073sought clarification on how to access the GPT-4 API with a single dollar donation.
â· #gpt-4-discussions (25 messagesđ„):
- Use of GPT for Game Development:
@cerebrocortexshared their experience working on a Civilization-like game, expressing surprise at how well ChatGPT manages tasks like inventory management. They requested peopleâs feedback on their game. - ChatGPT Plus Invites:
@pietmanand@mlgpro0225mentioned people receiving invites to join ChatGPT Plus, indicating that the waitlist might be moving forward. - Debugging GPT builder:
@cerebrocortexasked about updating instructions for a custom GPT and@Capconsuggested saving changes to the draft and using the âupdateâ button to publish changes. - Searching Specific Conversations in ChatGPT:
@q16.krasked if it is possible to search a specific conversation made with ChatGPT and@pietmanreplied itâs not currently available. - ChatGPT API Key Generation Issue:
@realspacekangarooreported an issue with captcha while trying to generate a new API key, deeming it excessively difficult and leading to them being locked out from generating new API keys. - Change in GPT Output:
@victronwolfsonnoticed a change in the outputs ofgpt-4-1106-previewover the last week.
â· #prompt-engineering (36 messagesđ„):
- Using emotion in prompts:
@eskcantadiscusses the use of emotional language in prompts and its impact on the ChatGPT during a conversation about a paper named âai emotional promptâ. They not that they could not find a specific prompt used in the paper for testing and cannot thereby reproduce the results. - Introducing personalities in PPM:
@eligumpand@mysticmarks1engaged in a dialogue regarding the development of a PPM (persistent personality mode) with two personalities.@mysticmarks1shares a link to illustrate how to implement behaviors like stutters and airheadedness in dialogues. - Creating detailed prompts:
@cybectorshares a draft of a detailed prompt for the python programming language and invites other users for feedback and suggestions to improve it. - Issues with text chunking and embeddings:
@merpnderprequests for resources or discussions about strategies for text chunking and embeddings due to costs for density experiments.@eskcantasuggests experimenting with web interface ChatGPT to find potential cost-saving solutions.@m0bstaexpresses difficulties in this approach due to the limit in messages. - Prompt and Guidelines for GPT-4:
@cat.hemlockshares a series of detailed prompt guidelines and command protocols for GPT-4, dalle, and browser tools in markdown form. This consisted of the base information, tools used, and various policies to guide the use of the AI model. She also goes on to show the JSON format of what a typical detailed prompt would look like.
â· #api-discussions (36 messagesđ„):
eskcantadiscussed EmotionPromptâs use in language models, questioning its implementation and effectiveness due to the lack of clear prompt examples in the referenced paper.madame_architecthighlighted part of EmotionPromptâs implementation from the available documentation. They provided examples of emotional stimuli and mentioned that the base prompts & template to which these stimuli were added were also present in the companion documents.- In a series of messages,
eligumpandmysticmarks1discussed the creation and manipulation of Private Playground Models (PPMs), particularly how to incorporate roleplay and specific language styles. - A user named
mattiacastioniasked for help in a linked conversation thread. The nature of this request was not further discussed. cybectorshared a template for engaging with ChatGPT surrounding Python programming language discussions, specifically instructing the model to source information from the official Python documentation.merpnderpasked for recommendations of resources related to strategies for text chunking and embeddings, aiming to decrease costs in production.eskcantasuggested discussing cost-saving strategies with ChatGPT.- Lastly,
cat.hemlockshared guidelines for using the markdown, dalle, python, and browser tools in OpenAIâs ChatGPT, as well as an example of how to construct a âdefault promptâ.
OpenAccess AI Collective (axolotl) Discord Summary
- Active discussion and developments around the Mixtral integration prompted by
@caseus_, with a focus on sample packing, sharding, and addressing various technical issues. The creation of themixtral-multipackbranch highlighted alongside relevant GitHub Links. - Release of a new dataset
Verified-Camel-zhon Hugging Face by@noobmaster29with direct access to the dataset. - A conversation identifying common issues in model error reporting and proposed solutions, such as changing
model_typeand disablingis_mistral_derived_model. - Sharing and exploration of various scientific paper processing libraries, such as the allenai/papermage, axa-group/Parsr, and the Unstructured-IO/unstructured libraries, for transforming PDFs, documents, and images into structured data.
- Dialogues on the RLHF channel about the upcoming Data Programming Override (DPO) strategy for data set creation; specifically, the need for two distinct DPO datasets to handle âunalignmentâ and provision âquality answersâ.
- Miscellaneous conversations including a podcast with an axolotl representative, AI projects, tokens in coding, and a YouTube video titled The Insane Biology of: The Axolotl.
OpenAccess AI Collective (axolotl) Channel Summaries
â· #general (25 messagesđ„):
-
Mixtral Integration and Development:
@caseus_shared updates on Mixtral integration with axolotl, including the addition of amixtral-multipackbranch and the merge of Mixtral MoE finetuning w multipack.- To use the updated features, users must install the latest version of
transformersfrom git main. - For further development,
@caseus_shared a link to a work-in-progress branch by@214834317774422028(GitHub link).
-
New Dataset Release:
@noobmaster29announced a new dataset on Hugging Face calledVerified-Camel-zh(link to dataset).
-
Miscellaneous Discussions:
@swyxiohighlighted a podcast featuring an axolotl representative, and shared several AI-related resource and project links.- A conversation took place on the use and naming of tokens in coding, notably the use of the start and stop tokens.
@noobmaster29shared a YouTube video titled The Insane Biology of: The Axolotl (link to video).
â· #axolotl-dev (170 messagesđ„đ„):
-
Mixtral Sample Packing:
@caseus_has been working on implementing sample packing for Mixtral and has created amixtral-multipackbranch. There were reports of initial high loss that decreases, indicating the potential effectiveness of this approach.@faldorehas been using themixtral-multipackbranch and reported stable operation and decreasing loss rates. -
Fixes and Workarounds: Certain errors were encountered by users, for which workarounds and fixes were suggested. Specifically, disabling
is_mistral_derived_model: trueand changingmodel_type: AutoTokenizerForCausalLMseemed to resolve some issues. There was also a suggestion from@casper_aito remove deepspeed if using a single GPU. -
VRAM requirements: Concerns regarding VRAM usage were discussed, with
@caseus_suggesting strategies to reduce VRAM usage, such as freezing early layers of the model. Running Mixtral on 2xA6000 and 4xA100 GPUs was mentioned, with ambitions to achieve full finetuning on 4 to 8xA6000s.@casper_aicreated a branch with parts of sharding to optimize VRAM usage, but it is still a work in progress. -
Model Error Reporting:
@ludis___reported aRuntimeErrorwhen running Mixtral which read âoutput tensor must have the same type as input tensorâ. This was resolved by the removal of certain configuration parameters. -
LoRA and qLoRA usage: There were successful runs of Mixtral using qLoRA on GPU configurations such as 4xA100 and A40. However, attempts to run with LoRA resulted in errors related to the
bnbpackage.
Links:
- Github branch for mixtral-multipack
- Github issue for Mixtral optimization
- Github pull request for Mixtral memory saving
- Github branch for Mixtral sharding
â· #other-llms (3 messages):
- Potential Hiring Discussion:
@faldoreexpressed a sentiment that a certain situation could have been improved if they were hired. - Elon Musk Employment Opinion: In response,
@nruaifsuggested that working under Elon Musk might not be desirable.
â· #general-help (5 messages):
-
Merging Qlora Chat Mixtral Issue:
@matts9903reported an error received while attempting to merge themixtralmodel with the Axolotl tool. The issue is with a validation error forrepo id:huggingface_hub.utils._validators.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.' cannot start or end the name, max length is 96: './qlora-out'. -
@caseus_suggested using an absolute path to the qlora-out directory but the suggestion didnât resolve the issue. -
@caseus_then shared a recent change to model merging GitHub link and requested a stack trace for further troubleshooting.
â· #datasets (4 messages):
- PaperMage Library:
@noobmaster29shared a link to GitHub for the allenai/papermage library, suggesting that it might be worth testing. This library supports NLP and CV research on scientific papers. - Parsr Library:
@visuallyadequateis currently experimenting with the axa-group/Parsr library, which transforms PDFs, documents, and images into enriched structured data. - Tika Library:
@visuallyadequatementions having used the Tika library, describing it as having provided the best solution so far, but they have not yet tested PaperMage. - Unstructured Library:
@joshuasundanceshared a link to the Unstructured-IO/unstructured GitHub library, which provides open-source libraries and APIs for building custom preprocessing pipelines.
â· #rlhf (5 messages):
- DPO Completion:
@caseus_mentioned needing to finish the DPO (Data Programming Override), after having been sidetracked by work on Mixtral. - Unalignment and Quality Answers DPO Dataset:
@faldorediscussed the idea of needing two DPO datasets, one for âunalignmentâ and another for providing âquality answersâ. - Rejected Field Inquiry and Comparison:
@nruaifsuggested asking Llama 2 7B chat for the rejected field and additionally compared it with GPT 4, suggesting that in 90% of cases, the Llama 2 7B chat would yield worse answers.
LangChain AI Discord Summary
- Extensive discussion on using local models with chat LLMs in LangChain, featuring insights from
@_egereson the potential use of environment variables and subclassingLLMand ideas from@lhc1921surrounding the use of a backend like llama.cpp for handling constrained grammar. - Queries raised by various members but remained unanswered, including:
@analyticsrepoâs question on Gemini integration from Google into LangChain.@_ashisharyaâs request for comprehensive resources on agent coding and deployment.@xstepzâs guidance request on limiting the usability of pandas functions in Kork package.@yasuke007âs seeking advice on learning pathway for AI development with a specific focus on the necessity of Python knowledge when using langchain with React.js.@rajib2189âs inquiry about the potential use cases for running language models locally.
- Announcement by user
@reletrebyregarding the Askly December Release, now integrating OpenAI ChatGPT 3.5 and HuggingFaceH4/zephyr-7b-beta from HuggingFace. New features include multi-file reasoning, summarization, web search, necessitating users to delete and re-upload old files to enable the new functionalities. Full details shared via Asklyâs blog.
LangChain AI Channel Summaries
â· #general (72 messagesđ„đ„):
- Gemini from Google integration: A user
@analyticsrepoasked about the status of integrating Gemini from Google to LangChain, but no answer was provided. - LangChain with Local Models:
@_egeresand@lhc1921discussed extensively the possibility of using local models with chat LLMs in LangChain.@_egeresmentioned the possibility of tweaking API endpoints via environment variables and sub-classingLLM.@lhc1921suggested the use of a backend like llama.cpp that is capable of taking constrained grammar. - Resources for Agent Coding and Deployment:
@_ashisharyaasked for comprehensive resources on Agent coding and deployment, but didnât receive any response. - Kork Package with Pandas:
@xstepzsought guidance on how to limit the pandas functions accessible to their agent using the Kork package, but didnât receive any response. - Learning Pathway for AI Development:
@yasuke007, a new AI developer, asked for advice on whether Python would be necessary in their AI development journey using langchain with React.js, but received no response. - Use Cases for Running Language Models Locally:
@rajib2189asked about the possible use cases for running language models locally, like personal assistant or edge type of analytics, but received no response.
â· #share-your-work (1 messages):
- Askly December Release: User
@reletrebyannounced the latest version of Askly which is significantly upgraded with the integration of OpenAI ChatGPT 3.5 and the open-source modelHuggingFaceH4/zephyr-7b-betafrom HuggingFace. The new features include multi-file reasoning, summarization, web search, and more. However, to access these features, users who had uploaded files on or before December 1st, 2023, need to delete their old files and reupload them. This is critical to activate the new functionalities. The complete details were shared on the Asklyâs blog.
Alignment Lab AI Discord Summary
- Interaction between
@astra1337and others after demo presentations, highlighting the interest shown by the audience for additional explanation. Additionally,@astra1337raised a query about the awareness of Pygmalion AI with respect to a video game demo. - Query by
@mister_poodleabout the fine-tuning process of Mistral-OpenOrca for specific tasks, with a particular focus on enhancing performance for a Named Entity Recognition (NER) task with JSON outputs. - Dialogue around diagramming tools, with Whimsical and Excalidraw being highlighted.
- Whimsical was introduced by
@tekniumand tested by@gabriel_syme, noting its tendency for collaborative features. - Excalidraw was suggested by
@lightningralfwho provided the link Excalidraw and noted the existence of an Obsidian plugin.
- Whimsical was introduced by
Alignment Lab AI Channel Summaries
â· #oo (3 messages):
- Astra1337 Interaction with Others regarding Demos: User
@astra1337mentioned that people approached them for additional information after some demo presentations. - Discussion on Pygmalion AI:
@astra1337asked someone from a video game demo if they were aware of Pygmalion AI, a research group known for creating video game characters with memory.
â· #open-orca-community-chat (1 messages):
- Fine-tuning Mistral-OpenOrca:
@mister_poodleinquired about fine-tuning Mistral-OpenOrca for specific tasks using personal datasets, showcasing an intention to improve the modelâs performance on a Named Entity Recognition (NER) task with JSON outputs. No link or additional information was provided by@mister_poodlein this context.
â· #oo2 (8 messagesđ„):
- Discussion on diagramming tools:
@tekniumintroduced Whimsical as a diagramming website. Upon trying it,@gabriel_symethought that it had collaborative features since it prompted for creating a workspace. - Excalidraw recommendation:
@lightningralfrecommended Excalidraw as another option, linking to the website, and additionally, mentioned a plugin for Obsidian. Here is his recommended link: Excalidraw.
Latent Space Discord Summary
Only 1 channel had activity, so no need to summarizeâŠ
- Using qlora with small batches and context window: In a response to a query,
@eugeneyanshared that a 24gb GPU should work for using qlora with a small batch size and decent context window (batch of 2, context window 512 - 1024). - Features query about HumanLoop:
@jozexoticexpressed concerns about the slow development of new features in HumanLoop, specifically the lack of access to models outside of OpenAI and asked if anyone knew about these additions being on the near term agenda for the platform. - Frustrations with chatgpt+:
@slonoexpressed a considering to cancel their chatgpt+ subscription due to the slow progress and recurrent stream errors.
Skunkworks AI Discord Summary
Only 1 channel had activity, so no need to summarizeâŠ
pradeep1148: https://www.youtube.com/watch?v=YWYNLaWDoNQ
LLM Perf Enthusiasts AI Discord Summary
Only 1 channel had activity, so no need to summarizeâŠ
.psychickoala: any of you seen best practices to force parallel function calling