All tags
Topic: "instruction-tuning"
not much happened today
deepseek-r1 deepseek-v3 coder-v2 prover deepseek hugging-face dell openai instruction-tuning performance-benchmarks model-deployment training-costs hardware-scalability ai-safety risk-mitigation ethical-ai open-source gpu-utilization yann-lecun yoshua-bengio francois-chollet giffman
DeepSeek-R1 and DeepSeek-V3 models have made significant advancements, trained on an instruction-tuning dataset of 1.5M samples with 600,000 reasoning and 200,000 non-reasoning SFT data. The models demonstrate strong performance benchmarks and are deployed on-premise via collaborations with Dell and Hugging Face. Training costs are estimated around $5.5M to $6M, with efficient hardware utilization on 8xH100 servers. The International AI Safety Report highlights risks such as malicious use, malfunctions, and systemic risks including AI-driven cyberattacks. Industry leaders like Yann LeCun and Yoshua Bengio provide insights on market reactions, AI safety, and ethical considerations, with emphasis on AI's role in creativity and economic incentives.
Too Cheap To Meter: AI prices cut 50-70% in last 30 days
gpt-4o gpt-4o-mini llama-3-1-405b mistral-large-2 gemini-1.5-flash deepseek-v2 sonnet-3.5 exaone-3.0 minicpm-v-2.6 claude-3.5 gpt-4o-2024-08-06 llamaindex together-ai deepinfra deepseek-ai mistral-ai google-deepmind lg-ai-research llamaindex llamaindex llamaindex price-cuts context-caching instruction-tuning vision benchmarks pytorch attention-mechanisms reinforcement-learning-from-human-feedback compute-optimal-scaling rohanpaul_ai akhaliq mervenoyann sophiamyang chhillee karpathy
Gemini 1.5 Flash has cut prices by approximately 70%, offering a highly competitive free tier of 1 million tokens per minute at $0.075/mtok, intensifying the AI model price war. Other significant price reductions include GPT-4o (~50% cut to $2.50/mtok), GPT-4o mini (70-98.5% cut to $0.15/mtok), Llama 3.1 405b (46% cut to $2.7/mtok), and Mistral Large 2 (62% cut to $3/mtok). Deepseek v2 introduced context caching, reducing input token costs by up to 90% to $0.014/mtok. New model releases include Llama 3.1 405b, Sonnet 3.5, EXAONE-3.0 (7.8B instruction-tuned by LG AI Research), and MiniCPM V 2.6 (vision-language model combining SigLIP 400M and Qwen2-7B). Benchmarks show Mistral Large performing well on ZebraLogic and Claude-3.5 leading LiveBench. FlexAttention, a new PyTorch API, simplifies and optimizes attention mechanisms. Andrej Karpathy analyzed RLHF, highlighting its limitations compared to traditional reinforcement learning. Google DeepMind research on compute-optimal scaling was also summarized.