All tags  
  Topic: "apache-license"
 not much happened today 
   glm-4.5  glm-4.5-air  qwen3-coder  qwen3-235b  kimi-k2  wan-2.2  grok-imagine  smollm3  figure-01  figure-02  vitpose++   zhipu-ai  alibaba  moonshot-ai  x-ai  ideogram  figure  smollm  openai   model-releases  moe  model-benchmarking  image-generation  video-generation  pose-estimation  robotics  training-code-release  apache-license   yuchenj_uw  corbtt  cline  reach_vb  ollama  deeplearningai  ostrisai  hojonathanho  adcock_brett  skalskip92  loubnabenallal1  
 Chinese labs have released a wave of powerful, permissively licensed models in July, including Zhipu AI's GLM-4.5 and GLM-4.5-Air, Alibaba's Qwen3 Coder and Qwen3-235B, and Moonshot AI's Kimi K2. These models feature large-scale Mixture of Experts architectures with active parameters ranging from 3B to 32B and context windows up to 256K tokens. Zhipu AI's GLM-4.5 competes with Claude 4 Opus and Gemini 2.5 Pro in benchmarks. Moonshot AI's Kimi K2 is a 1 trillion-parameter MoE model surpassing other open-weight models on LiveCodeBench and AceBench. In video and image generation, xAI launched Grok Imagine, and Wan2.2 impressed with its Image-to-Video approach. Ideogram released a character consistency model. Robotics advances include Figure's Figure-01 and Figure-02 humanoid robots and ViTPose++ for pose estimation in basketball analysis. The SmolLM3 training and evaluation code was fully released under an Apache 2.0 license. "Orgs avoiding these Chinese open-source models are at a significant competitive disadvantage," noted by @corbtt.
  LlamaCon: Meta AI gets into the Llama API platform business 
   llama-4  qwen3  qwen3-235b-a22b  qwen3-30b-a3b  qwen3-4b  qwen2-5-72b-instruct  o3-mini   meta-ai-fair  cerebras  groq  alibaba  vllm  ollama  llamaindex  hugging-face  llama-cpp   model-release  fine-tuning  reinforcement-learning  moe  multilingual-models  model-optimization  model-deployment  coding  benchmarking  apache-license   reach_vb  huybery  teortaxestex  awnihannun  thezachmueller  
 Meta celebrated progress in the Llama ecosystem at LlamaCon, launching an AI Developer platform with finetuning and fast inference powered by Cerebras and Groq hardware, though it remains waitlisted. Meanwhile, Alibaba released the Qwen3 family of large language models, including two MoE models and six dense models ranging from 0.6B to 235B parameters, with the flagship Qwen3-235B-A22B achieving competitive benchmark results and supporting 119 languages and dialects. The Qwen3 models are optimized for coding and agentic capabilities, are Apache 2.0 licensed, and have broad deployment support including local usage with tools like vLLM, Ollama, and llama.cpp. Community feedback highlights Qwen3's scalable performance and superiority over models like OpenAI's o3-mini.