Back to home

Qwen

19 articles tagged with this topic

QwenAlibaba

阿里 Qwen 3.6 Max 悄悄上线,中国模型榜单第一——但开源还是闭源,这才是真正的问题

Alibaba's Qwen 3.6 Max quietly launched in preview, scoring highest among Chinese models — but its open-source status remains undecided.

Apr 202 min read
CrewAIQwen

CrewAI 装了跑不起来?一篇部署指南背后,是 AI 多智能体工具门槛还没降下来的现实

A 3,000-word Cre wAI setup guide went viral on Juejin—proof that multi -agent frameworks are hot, but nowhere near enterprise-ready.

Apr 203 min read
QwenClaude

有人开始用国产开源模型替换 Claude 做日常编程助手 — 性能差距正在缩小到「够用」

Developers on Reddit are seriously evaluating Alibaba's Qwen-35B-A3B as a local replacement for Claude Opus 4. 7 in daily coding workflows.

Apr 203 min read
QwenAlibaba

Qwen 3.6 35B Runs "Browser OS" Locally — Open- Source Models Are Closing the Gap

A developer ran Alibaba's Qwen 3.6 35B locally to achieve "Browser OS" — AI orchest rating a browser like an OS, no cloud needed.

Apr 191 min read
AlibabaQwen

一台消费级显卡,AI 帮用 户重写了整个记账软件——阿里 Qwen 新模型让「本地运行」开 始变得真实

Alibaba's Qwen3.6-35B-A3B rewrote a full accounting app on a single RTX 5070 Ti in under an hour—where older models failed.

Apr 183 min read
QwenAlib aba

Alibaba Releases Qwen3.6-35B-A3B Mixture-of-Experts Model

Alibaba's Qwen team releases Qwen3.6-35B-A3B, a 35B-parameter MoE model activating 3B parameters per token.

Apr 162 min read
QwenQwen3.6- 35B-A3B

Qwen3.6-35B-A3B released!

Alibaba's Qwen team releases a 35B sparse MoE model with only 3B active params under Apache 2.0.

Apr 163 min read
QwenAmazon SageMaker

Fine-Tune Qwen 2.5 for Tool Calling with SageMaker RLVR

AWS SageMaker serverless RLVR fine-tuning improved Qwen 2.5 7B tool-call accuracy by 57% without GPU management.

Apr 72 min read
llama.cppQwen

37 LLMs Benchmarked on MacBook Air M5 32GB: Full Speed Results

Community benchmark of 37 local LLMs on M5 Air 32GB using llama-bench reveals MoE models as clear winners for speed-to-quality ratio.

Apr 62 min read
OpenCodellama-server

OpenCode + Local LLMs: Which Models Work Best for Solo Dev Tasks

A hands-on benchmark of OpenCode with 6+ self-hosted LLMs on an RTX 4080 for real coding tasks.

Apr 62 min read
QwenvLLM

Agent Swarms + Continuous Batching Cut LLM Task Time 36x

Running 50 parallel agents on Qwen 27B drops a 42-minute research job to 70 seconds using continuous batching.

Apr 62 min read
BrookingsWorkforce Automation

AI Saturation: What It Means for the Future of Work

Brookings examines AI saturation trends and workforce displacement risks for knowledge workers globally.

Apr 52 min read
a16zAutoGen

Marc Andreessen's AI Coworker Vision: What It Means for Dev Teams

a16z's Andreessen outlines autonomous AI agents as future coworkers, with real implications for how small teams hire and build.

Apr 52 min read
QwenAlibaba Cloud

Qwen 3.6 Spotted in Official App Alongside 3.5 Max Preview

A Reddit user spotted Qwen 3.6 inside the official Qwen app, suggesting an imminent public release beyond API access.

Apr 51 min read
llama.cppDistributed Training

Local Inference vs Distributed Training: Where the Real Gap Is

Indie devs run models locally, but training still requires datacenter scale. Can distributed training ever close that gap?

Apr 52 min read
open-source-llmQwen

Chinese AI Labs Delay Open-Source Releases: What Solo Builders Should Do Now

Qwen, GLM, MiniMax all stalling open-weight releases. Here's how solopreneurs should hedge their model stack.

Apr 52 min read
llama.cppTurboQuant

TurboQuant KV Cache Quantization Beats Baselines on Gemma 4 and Qwen

Community benchmarks show TurboQuant KV quantization achieves near-zero accuracy loss at 3.1 bits on Gemma 4 with 34% long-context speedup.

Apr 52 min read
vLLMPagedAttention

vLLM PagedAttention: From Memory Management to Production Deployment

vLLM's PagedAttention raises GPU memory utilization from 60% to 95%+ using OS paging concepts for LLM inference.

Apr 52 min read
ChatGPTARC-AGI

Why Programmers Should Ignore the AI Replacement Anxiety Hype

AI lowers the floor for output but not the ceiling for quality. Here is why effort still compounds.

Apr 42 min read