Most active commenters
  • diggan(5)
  • echelon(4)
  • rushingcreek(3)

←back to thread

221 points lnyan | 32 comments | | HN request time: 1.311s | source | bottom
1. rushingcreek ◴[] No.44397235[source]
It doesn't seem to have open weights, which is unfortunate. One of Qwen's strengths historically has been their open-weights strategy, and it would have been great to have a true open-weights competitor to 4o's autoregressive image gen. There are so many interesting research directions that are only possible if we can get access to the weights.

If Qwen is concerned about recouping its development costs, I suggest looking at BFL's Flux Kontext Dev release from the other day as a model: let researchers and individuals get the weights for free and let startups pay for a reasonably-priced license for commercial use.

replies(4): >>44397843 #>>44397858 #>>44397893 #>>44398602 #
2. Jackson__ ◴[] No.44397843[source]
It's also very clearly trained on OAI outputs, which you can tell from the orange tint to the images[0]. Did they even attempt to come up with their own data?

So it is trained off OAI, as closed off as OAI and most importantly: worse than OAI. What a bizarre strategy to gate-keep this behind an API.

[0]

https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen-VLo/cas...

https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen-VLo/cas...

https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen-VLo/cas...

replies(5): >>44397961 #>>44398084 #>>44398731 #>>44401456 #>>44418786 #
3. diggan ◴[] No.44397858[source]
> One of Qwen's strengths historically has been their open-weights strategy [...] let researchers and individuals get the weights for free and let startups pay for a reasonably-priced license for commercial use.

But if you're suggesting they should do open weights, doesn't that mean people should be able to use it freely?

You're effectively suggesting "trial-weights", "shareware-weights", "academic-weights" or something like that rather than "open weights", which to me would make it seem like you can use them for whatever you want, just like with "open source" software. But if it misses a large part of what makes "open source" open source, like "use it for whatever you want", then it kind of gives the wrong idea.

replies(1): >>44397891 #
4. rushingcreek ◴[] No.44397891[source]
I am personally in favor of true open source (e.g. Apache 2 license), but the reality is that these model are expensive to develop and many developers are choosing not to release their model weights at all.

I think that releasing the weights openly but with this type of dual-license (hence open weights, but not true open source) is an acceptable tradeoff to get more model developers to release models openly.

replies(1): >>44398288 #
5. echelon ◴[] No.44397893[source]
The era of open weights from China appears to be over for some reason. It's all of a sudden and seems to be coordinated.

Alibaba just shut off the Qwen releases

Tencent just shut off the Hunyuan releases

Bytedance just released Seedream, but it's closed

It's seems like it's over.

They're still clearly training on Western outputs, though.

I still suspect that the strategic thing to do would be to become 100% open and sell infra/service.

replies(6): >>44397943 #>>44398085 #>>44398090 #>>44399651 #>>44401386 #>>44403837 #
6. pxc ◴[] No.44397943[source]
Why? And can we really say that already? Wasn't the Qwen3 release still very recent?
7. echelon ◴[] No.44397961[source]
The way they win is to be open. I don't get why China is shutting down open source. It was a knife at the jugular of US tech dominance.

Both Alibaba and Tencent championed open source (Qwen family of models, Hunyuan family of models), but now they've shut off the releases.

There's totally a play where models become loss-leader for SaaS/PaaS/IaaS and where they extinguish your closed competition.

Imagine spreading your model so widely then making the terms: "do not use in conjunction with closed source models".

replies(3): >>44398211 #>>44399036 #>>44401766 #
8. vachina ◴[] No.44398084[source]
Huh, so orange tint = openAI output? Maybe their training process ended up causing the model to prefer that color balance.
replies(1): >>44398560 #
9. natrys ◴[] No.44398085[source]
> Alibaba just shut off the Qwen releases

Alibaba from beginning had some series of models that are always closed-weights (*-max, *-plus, *-turbo etc. but also QvQ), It's not a new development, nor does it prevent their open models. And the VL models are opened after 2-3 months of GA in API.

> Tencent just shut off the Hunyuan releases

Literally released one today: https://huggingface.co/tencent/Hunyuan-A13B-Instruct

replies(1): >>44399326 #
10. logicchains ◴[] No.44398090[source]
What do you mean Tencent just shut off the Hunyuan releases? There was another open weights release just today: https://huggingface.co/tencent/Hunyuan-A13B-Instruct . And the latest Qwen and DeepSeek open weight releases were under 2 months ago, there hasn't been enough time for them to finish a new version since then.
replies(1): >>44399335 #
11. diggan ◴[] No.44398211{3}[source]
> I don't get why China is shutting down open source [...] now they've shut off the releases

What are you talking about? Feels like a very strong claim considering there are ongoing weight releases, wasn't there one just today or yesterday from a Chinese company?

12. diggan ◴[] No.44398288{3}[source]
> but the reality is that these model are expensive to develop and many developers are choosing not to release their model weights at all.

But isn't that true for software too? Software is expensive to develop, and lots of developers/companies are choosing not to make their code public for free. Does that mean you also feel like it would be OK to call software "open source" although it doesn't allow usage for any purpose? That would then lead to more "open source" software being released, at least for individuals and researchers?

replies(2): >>44398604 #>>44399878 #
13. Jackson__ ◴[] No.44398560{3}[source]
Here's an extreme example that shows how it continually adds more orange: https://old.reddit.com/r/ChatGPT/comments/1kawcng/i_went_wit...

It's really too close to be anything but a model trained on these outputs, the whole vibe just screams OAI.

replies(1): >>44399782 #
14. dheera ◴[] No.44398602[source]
> One of Qwen's strengths historically has been their open-weights strategy

> let researchers and individuals get the weights for free and let startups pay for a reasonably-priced license for commercial use

I'm personally doubtful companies can recoup tens of millions of dollars in investment, GPU hours, and engineering salaries from image generation fees.

15. rushingcreek ◴[] No.44398604{4}[source]
Yes, I think the same analogy applies. Given a binary choice of a developer not releasing any code at all or releasing code under this type of binary "open-code" license, I'd always take the latter.
replies(1): >>44398701 #
16. diggan ◴[] No.44398701{5}[source]
> Given a binary choice of a developer not releasing any code at all

I mean it wasn't binary earlier, it was "to get more model developers to release", so not a binary choice, but a gradient I suppose. Would you still make the same call for software as you do for ML models and weights?

17. VladVladikoff ◴[] No.44398731[source]
What would be the approximate cost of doing this? How many million API requests must be made? How many tokens in total?
replies(1): >>44399468 #
18. yorwba ◴[] No.44399036{3}[source]
The problem with giving away weights for free while also offering a hosted API is that once the weights are out there, anyone else can also offer it as a hosted API with similar operating costs, but only the releasing company had the initial capital outlay of training the model. So everyone else is more profitable! That's not a good business strategy.

New entrants may keep releasing weights as a marketing strategy to gain name recognition, but once they have established themselves (and investors start getting antsy about ROI) making subsequent releases closed is the logical next step.

replies(1): >>44401469 #
19. echelon ◴[] No.44399326{3}[source]
Hunyuan Image 2.0, which is of Flux quality but has ~20 milliseconds of inference time, is being withheld.

Hunyuan 3D 2.5, which is an order of magnitude better than Hunyuan 3D 2.1, is also being withheld.

I suspect that now that they feel these models are superior to Western releases in several categories, they no longer have a need to release these weights.

replies(1): >>44399796 #
20. echelon ◴[] No.44399335{3}[source]
Hunyuan Image 2.0 and Hunyuan 3D 2.5 are not being released. They're being put into a closed source web-based offering.
21. refulgentis ◴[] No.44399468{3}[source]
Most pedantically correct answer is "mu", because the answers are both derivable quantitively from "How many images do you want to train on?", which is answered by a qualitative question that doesn't admit numbers ("How high quality do you want it to be?")

Let's say it's 100 images because you're doing a quick LoRA. That'd be about $5.00 at medium quality (~$0.05/image) or $1 at low. ~($0.01/image)

Let's say you're training a standalone image model. OOM of input images is ~1B, so $10M at low and $50M at high.

250 tokens / image for low, ~1000 for medium, which gets us to:

Fastest LoRA? $1-$4. 25,000 - 100,000 tokens output. All the training data for a new image model? $10M-$50M, 2.5B - 10B tokens out.

22. jacooper ◴[] No.44399651[source]
Deepseek R1 0528, the flagship Chinese model is open source. Qwen3 is open source. HIdream models are also open source
23. acheong08 ◴[] No.44399782{4}[source]
That form of collapse might just be inherent to the methodology. Releasing the weights would be nice so people can figure out why
24. natrys ◴[] No.44399796{4}[source]
> I suspect that now that they feel these models are superior to Western releases in several categories, they no longer have a need to release these weights.

Yes that I can totally believe. Standard corporation behaviour (Chinese or otherwise).

I do think DeepSeek would be an exception to this though. But they lack diversity in focus (not even multimodal yet).

25. hmottestad ◴[] No.44399878{4}[source]
I wouldn't equate model weights with source code. You can run software on your own machine without source code, but you can't run an LLM on your own machine without model weights.

Though, you could still sell the model weights for local use. Not sure if we are there yet that I myself could buy model weights, but of course if you are a very big company or a very big country then I guess most AI companies would consider selling you their model weights so you can run them on your own machine.

26. amelius ◴[] No.44401386[source]
era -> fluke
27. roenxi ◴[] No.44401456[source]
There seem to be a lot of AI images on the web these days and it might have become the single most dominant style given that AI has created more images than any individual human artist. So they might have trained on them implicitly rather than in a synthetic way.

Although theory is not practice. If I were an AI company I'd try to leverage other AI company APIs.

28. roenxi ◴[] No.44401469{4}[source]
That is also how open source works in other contexts. Initially closed source is dominant, then over time other market entrants use OSS solutions to break down the incumbent advantage.

In this case I'm expecting people with huge pools of capital (the big cloud providers) to push out open models because the weights are a commodity then people will rent their servers to multiply them together.

replies(1): >>44402420 #
29. rfv6723 ◴[] No.44401766{3}[source]
If you have worked or lived in China, you will know that Chinese open-source software industry is a totally shitshow.

The law in China offers little protection for open-source software. Lots of companies use open-source code in production without proper license, and there is no consequence.

Western internet influencers hype up Chinese open-source software industry for clicks while Chinese open-source developers are struggling.

These open-weight model series are planed as free-trial from the start, there is no commitment to open-source.

replies(1): >>44403779 #
30. yorwba ◴[] No.44402420{5}[source]
Even for a big cloud provider, putting out model weights and hoping that people host with them is unlikely to be as profitable as gating it behind an API that guarantees that people using the model are using their hosted version. How many people self-hosting Qwen models are doing so on Aliyun?
31. diggan ◴[] No.44403779{4}[source]
> Western internet influencers hype up Chinese open-source software industry for clicks while Chinese open-source developers are struggling.

That kind of downplays that Chinese open weights are basically the only option for high quality weights you can run yourself, together with Mistral. It's not just influencers who are "hyping up Chinese open-source" but people go where the options are.

> there is no commitment to open-source

Welcome to open source all around the world! Plenty of non-Chinese projects start as FOSS and then slowly move into either fully proprietary or some hybrid-model, that isn't exactly new nor not expected, Western software industry even pioneered a new license (BSL - https://en.wikipedia.org/wiki/Business_Source_License) that tries to look as open source as possible while not actually being open source.

32. WiSaGaN ◴[] No.44403837[source]
Deepseek and Alibaba just published their froniter models in open weights weeks ago. And they happen to be the leading open weights models in the world. What are you talking about?