• ukuina 3 hours ago

    [June 2025]

    • p0w3n3d 35 minutes ago

      Yeah I've been using qwen3 on mlx in July already

      • NiekvdMaas 2 hours ago
        • littlestymaar an hour ago

          The unfortunate thing with their Qwen3-next naming is that it doesn't reflect on the fact that the architecture is completely different from Qwen3. Much more different than the difference between Qwen2 and Qwen3 even.

          So support is likely to take quite some time because it's not just regular transformer blocks stacked on each other, but a brand new hybrid architecture using SSM.

          • NiekvdMaas an hour ago

            From https://github.com/ggml-org/llama.cpp/issues/15940#issuecomm...:

            > This is a massive task, likely 2-3 months of full-time work for a highly specialized engineer. Until the Qwen team contributes the implementation, there are no quick fixes.

            • veber-alex 32 minutes ago

              It's already supported in vLLM, SGLang and MLX.