Qwen
19 articles tagged with this topic
阿里 Qwen 3.6 Max 悄悄上线,中国模型榜单第一——但开源还是闭源,这才是真正的问题
Alibaba's Qwen 3.6 Max quietly launched in preview, scoring highest among Chinese models — but its open-source status remains undecided.
CrewAI 装了跑不起来?一篇部署指南背后,是 AI 多智能体工具门槛还没降下来的现实
A 3,000-word Cre wAI setup guide went viral on Juejin—proof that multi -agent frameworks are hot, but nowhere near enterprise-ready.
有人开始用国产开源模型替换 Claude 做日常编程助手 — 性能差距正在缩小到「够用」
Developers on Reddit are seriously evaluating Alibaba's Qwen-35B-A3B as a local replacement for Claude Opus 4. 7 in daily coding workflows.
Qwen 3.6 35B Runs "Browser OS" Locally — Open- Source Models Are Closing the Gap
A developer ran Alibaba's Qwen 3.6 35B locally to achieve "Browser OS" — AI orchest rating a browser like an OS, no cloud needed.
一台消费级显卡,AI 帮用 户重写了整个记账软件——阿里 Qwen 新模型让「本地运行」开 始变得真实
Alibaba's Qwen3.6-35B-A3B rewrote a full accounting app on a single RTX 5070 Ti in under an hour—where older models failed.
Alibaba Releases Qwen3.6-35B-A3B Mixture-of-Experts Model
Alibaba's Qwen team releases Qwen3.6-35B-A3B, a 35B-parameter MoE model activating 3B parameters per token.
Qwen3.6-35B-A3B released!
Alibaba's Qwen team releases a 35B sparse MoE model with only 3B active params under Apache 2.0.
Fine-Tune Qwen 2.5 for Tool Calling with SageMaker RLVR
AWS SageMaker serverless RLVR fine-tuning improved Qwen 2.5 7B tool-call accuracy by 57% without GPU management.
37 LLMs Benchmarked on MacBook Air M5 32GB: Full Speed Results
Community benchmark of 37 local LLMs on M5 Air 32GB using llama-bench reveals MoE models as clear winners for speed-to-quality ratio.
OpenCode + Local LLMs: Which Models Work Best for Solo Dev Tasks
A hands-on benchmark of OpenCode with 6+ self-hosted LLMs on an RTX 4080 for real coding tasks.
Agent Swarms + Continuous Batching Cut LLM Task Time 36x
Running 50 parallel agents on Qwen 27B drops a 42-minute research job to 70 seconds using continuous batching.
AI Saturation: What It Means for the Future of Work
Brookings examines AI saturation trends and workforce displacement risks for knowledge workers globally.
Marc Andreessen's AI Coworker Vision: What It Means for Dev Teams
a16z's Andreessen outlines autonomous AI agents as future coworkers, with real implications for how small teams hire and build.
Qwen 3.6 Spotted in Official App Alongside 3.5 Max Preview
A Reddit user spotted Qwen 3.6 inside the official Qwen app, suggesting an imminent public release beyond API access.
Local Inference vs Distributed Training: Where the Real Gap Is
Indie devs run models locally, but training still requires datacenter scale. Can distributed training ever close that gap?
Chinese AI Labs Delay Open-Source Releases: What Solo Builders Should Do Now
Qwen, GLM, MiniMax all stalling open-weight releases. Here's how solopreneurs should hedge their model stack.
TurboQuant KV Cache Quantization Beats Baselines on Gemma 4 and Qwen
Community benchmarks show TurboQuant KV quantization achieves near-zero accuracy loss at 3.1 bits on Gemma 4 with 34% long-context speedup.
vLLM PagedAttention: From Memory Management to Production Deployment
vLLM's PagedAttention raises GPU memory utilization from 60% to 95%+ using OS paging concepts for LLM inference.
Why Programmers Should Ignore the AI Replacement Anxiety Hype
AI lowers the floor for output but not the ceiling for quality. Here is why effort still compounds.