Top Builders

Explore the top contributors showcasing the highest number of app submissions within our community.

Qwen3

Qwen3 is the third-generation text model family from Alibaba Cloud's Qwen team, released on April 28, 2025. It covers six dense sizes (0.6B to 32B) and two MoE variants, all trained on approximately 36 trillion tokens across 119 languages. A key design choice is a unified thinking and non-thinking mode in every model, so developers can choose between step-by-step reasoning and fast single-pass responses without switching models.

General
Release date28 Apr 2025
DeveloperQwen / Alibaba Cloud
TypeOpen-weight text LLM family
LicenseApache 2.0
GitHubQwenLM/Qwen3
Hugging Facehuggingface.co/Qwen
Documentationqwenlm.github.io/blog/qwen3

Core Features

  • Thinking/non-thinking mode: every model supports both step-by-step chain-of-thought reasoning and direct response generation within a single checkpoint.
  • Thinking budget: developers can set a token budget for the reasoning phase, allowing inference cost to be tuned per request.
  • Long context: models at 4B and above support 131,072-token context windows; 0.6B and 1.7B support 32,768 tokens.
  • Multilingual: pretrained on 119 languages and dialects.
  • Apache 2.0: all weights are released for commercial use, fine-tuning, and redistribution.

Model Variants

VariantTotal ParamsActive ParamsContextBest for
Qwen3-0.6B0.6B0.6B32KEdge and on-device
Qwen3-1.7B1.7B1.7B32KLightweight inference
Qwen3-4B4B4B128KBalanced performance
Qwen3-8B8B8B128KGeneral tasks
Qwen3-14B14B14B128KHigher accuracy
Qwen3-32B32B32B128KStrong reasoning
Qwen3-30B-A3B30B3B128KEfficient MoE
Qwen3-235B-A22B235B22B128KFlagship MoE

Benchmarks

The flagship Qwen3-235B-A22B model scores:

  • AIME '24: 85.7
  • AIME '25: 81.5
  • LiveCodeBench v5: 70.7
  • BFCL v3: 70.8

Tools and Resources


Ecosystem and Integrations

  • Available on Hugging Face Hub in both standard and GGUF formats.
  • Accessible via Alibaba Cloud DashScope using an OpenAI-compatible endpoint.
  • Supported by Ollama, LM Studio, and major inference frameworks including vLLM and llama.cpp.
  • All sizes available for fine-tuning using standard supervised fine-tuning and RL pipelines.

Qwen3 weights are available immediately on Hugging Face. To access via API, generate a key on the Qwen API Platform and follow the Model Studio documentation.

Qwen Qwen3 AI technology Hackathon projects

Discover innovative solutions crafted with Qwen Qwen3 AI technology, developed by our community members during our engaging hackathons.

Boundary Forge

Boundary Forge

Boundary Forge is a model-agnostic AI safety pipeline that helps enterprises deploy LLMs with measurable confidence. Instead of relying on manual red-teaming or hoping a system prompt is enough, Boundary Forge automatically attacks a model, identifies where it behaves unsafely or inconsistently, and converts those discovered failures into runtime guardrails. For this hackathon, we demonstrated Boundary Forge using Qwen 2.5-72B on AMD Developer Cloud with AMD MI300X. Qwen powered the adversarial red-team workflow and was also the model under test, allowing the system to expose real behavioral failure boundaries such as jailbreak attempts, policy drift, unsafe financial guidance, KYC bypass, fraud patterns, coercion signals, asset concealment, and inconsistent refusals. The pipeline works in five stages: generate adversarial probes, run high-throughput model inference, mathematically detect boundary failures, compile those failures into semantic safety rules, and enforce them through middleware before risky prompts reach the LLM. This creates a practical enterprise safety layer that can block, flag, or ask for clarification in real time. The important point is that Boundary Forge is not tied to one model. Qwen 2.5-72B was used to demonstrate the system, but the architecture can benchmark and harden other open-source or proprietary models as well. The goal is to improve models exactly where they fail and make model evaluation repeatable across different deployments. In our AMD Cloud production run with Qwen 2.5-72B, Boundary Forge generated 1,009 unique adversarial probes, fired 4,036 total inferences, discovered 25 boundary failures, and compiled 15 semantic safety rules. The middleware intercepted 68% of known attacks and reduced the effective failure rate from 2.48% to 0.79%. Boundary Forge turns AI safety into an automated engineering workflow: attack, measure, learn, protect, and benchmark again.

Thor v2 — RAG-Free Fitness Intelligence

Thor v2 — RAG-Free Fitness Intelligence

Thor v2 is a domain-expert fitness AI built on a single fine-tuned Qwen3-8B model trained on 7,118 carefully constructed instruction-response pairs spanning exercise science, nutrition, programming, injury screening, and population-specific guidance. Unlike RAG-based fitness apps that retrieve documents at query time, Thor v2 encodes knowledge directly into model weights during supervised fine-tuning on AMD MI300X hardware using ROCm. Evidence is referenced through compact citation keys — e.g. [CITE:NSCA_HYPERTROPHY_VOLUME] — that the model emits inline. A lightweight citation resolver validates these keys against a locked registry and surfaces the source document on demand. If the model emits an unknown key, it is rejected at runtime. Hallucinated citations are structurally impossible. The dataset covers 113 unique citation keys from 9 authoritative organisations — NSCA, ACSM, ISSN, NASM, HHS, USDA, NIH, CDC, and ExRx — with 80 exercise technique entries and 14 population profiles including senior, postpartum, teen, vegan, rehab return, and competitive athlete. Six conversational style variants (casual, research-nerd, anxious, skeptical, verbose, follow-up-first) are baked into training so the model adapts tone naturally without prompt engineering. Training results: 100% JSON contract pass rate across all eval prompts. Coach gating behavior confirmed — model asks clarifying questions before prescribing when context is missing, rather than giving generic advice. All responses emit valid citation_keys, follow_up_questions, and safety_notes fields. Adapter size: <350MB on top of a frozen 8B base. Built entirely on AMD MI300X (192GB HBM3, ROCm 6.3) using HuggingFace PEFT + TRL. One model. No retrieval. No vector database. The model knows. The resolver proves.

TempoGraph: Local Multimodal Video Analysis

TempoGraph: Local Multimodal Video Analysis

TempoGraph is a fully-local, privacy-preserving multimodal video analysis system that turns raw video files into rich structured outputs — entities, behaviors, transcripts, timelines, and interactive knowledge graphs — without sending a single frame to the cloud. Stage 1 — Frame Selection: Motion-aware sampling with static, moving, and auto camera modes. For moving cameras it estimates homography to separate object motion from camera movement, then identifies keyframes where motion peaks exceed a configurable sigma threshold. Stage 1.5 — Audio Transcription: Whisper.cpp running on Vulkan transcribes the full audio track to millisecond-accurate segments. Stage 2 — YOLO Detection: YOLO26 runs on 2nd GPU over every sampled frame, outputting normalized bounding boxes, class names, track IDs, and confidence scores. Stage 3 — Depth Estimation: Depth Anything V2 via HuggingFace Transformers adds per-detection mean depth to every bounding box, giving 3D spatial context to 2D detections. Stage 4 — Frame Scoring: Picks which frames the VLM actually sees. In keyframes mode, only motion-peak frames are forwarded. In scored mode, FrameScorer ranks all YOLO-scanned frames using a weighted combination of motion delta, new YOLO class appearances, tracked object churn, and IoU drop between frames — then fills the VLM budget with the highest-signal frames. Keyframes are always pinned in first regardless of mode. Stage 5 — VLM Captioning: Qwen3.5-VL-9B served by a custom llama.cpp build compiled for AMD ROCm/HIP, running on an AMD RX 9070 XT with a 100k-token context window. Frames are chunked and sent to the model alongside YOLO-derived annotations. Each chunk's summary seeds the next prompt for narrative continuity across the video. Stage 6 — Aggregation: A final text-only LLM call synthesizes all per-chunk captions and the audio transcript into a structured JSON with entities, visual events, audio events, and multimodal correlations linking what was said to what was seen.