←back to thread

361 points mseri | 9 comments | | HN request time: 0.31s | source | bottom
1. tcsenpai ◴[] No.46002872[source]
Also: I simply asked to the 7B model "hi, who are u":

Thought for 1 seconds Okay, the user is asking "Who are u?" with a casual tone ("u" instead of "you", exclamation mark). Let me break this down.

First, I need to address the question itself. They want to know who I am as an AI assistant. My standard response is that I'm DeepSeek R1, designed to be helpful and handle reasoning step by step. But maybe they're not familiar with AI terms in general. The "u" part might just be their way of speaking, so I should reply politely using "you" form.

Next, considering the context. They started with an HTML tag " "

----

And it hung there without a reply. It might be an OpenWebUI bug but not so promising

replies(5): >>46002961 #>>46004196 #>>46005255 #>>46005278 #>>46007584 #
2. hasperdi ◴[] No.46002961[source]
I just tried that on their playground:

7B:Hi! I'm Olmo 3, an AI assistant created by the non-profit organization Ai2. I'm here to help with questions, ideas, or tasks you have—just let me know what you need! How can I assist you today? Rawr!

32B: Hi! I'm Olmo, a helpful AI assistant built by the Allen Institute for AI (Ai2). My knowledge is up to December 2024, and I'm designed to assist with a wide range of tasks. How can I help you today?

3. embedding-shape ◴[] No.46004196[source]
Every time a new model is released, there are a bunch of reports or written experiences about people using the model with software that seemingly doesn't support it.

GPT-OSS really made that clear, where 90% of the ecosystem declared it broken, but most people were using dumb quants and software that didn't properly support it. Guess we'll repeat the same thing with OLMo now.

replies(2): >>46004552 #>>46010529 #
4. andy99 ◴[] No.46004552[source]
There are a bunch (currently 3) of examples of people getting funny output, two of which saying it’s in LM studio (I don’t know what that is). It does seem likely that it’s somehow being misused here and the results aren’t representative.
replies(1): >>46004978 #
5. embedding-shape ◴[] No.46004978{3}[source]
Definitely. Usually I'd wait 2-3 weeks for the ecosystem to catch up and iron out the kinks, or do what I did for GPT-OSS, fix it in the places where it's broken, then judge it when I'm sure it's actually used correctly.

Otherwise, in that early period of time, only use the provided scripts/tools from the people releasing the model itself, which is probably the only way in those 2-3 weeks to be sure you're actually getting the expected responses.

6. fnbr ◴[] No.46005255[source]
(I’m a researcher on the post-training team at Ai2.)

Where did you try this? On the Ai2 playground?

7. latexr ◴[] No.46005278[source]
Reminds me of an old joke where a guy is walking down the street and another person says “good morning”. The guy starts deconstructing what “good morning” means until he finally reaches the conclusion “that bastard was calling me an asshole”.
8. red2awn ◴[] No.46007584[source]
Check the maximum number of completion tokens allowed.
9. Sabinus ◴[] No.46010529[source]
I'm really glad to read this, as this was my experience in LM studio with olmo. Worked for the first message but got progressively more unstable. Also doesn't seem to reset model state for a new conversation, every response following the model load gets progressively worse, even in new chats.