Most active commenters
  • TeMPOraL(3)

←back to thread

586 points mizzao | 26 comments | | HN request time: 1.762s | source | bottom
1. k__ ◴[] No.40666893[source]
I played around with Amazon Q and while setting it up, I needed to create an IAM identity center.

Never did this before, so I was asking Q in the AWS docs how to do it.

It refused to help, as it didn't answer security related questions.

thank.

replies(7): >>40666950 #>>40667091 #>>40667339 #>>40669069 #>>40669289 #>>40669327 #>>40671251 #
2. arianvanp ◴[] No.40666950[source]
This limitation is new. And it's so annoying. 95% of the time my questions I have surrounding AWS are IAM or security related and this thing refuses to answer anything. It's so annoying.
replies(1): >>40666979 #
3. el_benhameen ◴[] No.40666979[source]
It’s an absolute disaster. It wouldn’t answer something along the lines of “what is IAM” when I asked increasingly simple “security” related questions. Very little chance I’ll try an aws AI offering again any time soon.
4. menacingly ◴[] No.40667091[source]
it’s similar asking the gemini-1.5 models about coding questions that involve auth

one of my questions about a login form also tripped a harassment flag

replies(1): >>40667279 #
5. michaelt ◴[] No.40667279[source]
I suspect the refusal to answer questions about auth aren't a matter of hacking or offensive material.

I suspect instead the people training these models have identified areas of questioning where their model is 99% right, but because the 1% wrong is incredibly costly they dodge the entire question.

Would you want your LLM to give out any legal advice, or medical advice, or can-I-eat-this-mushroom advice, if you knew due to imperfections in your training process, it sometimes recommended people put glue in their pizza sauce?

replies(1): >>40667649 #
6. lhl ◴[] No.40667339[source]
I believe Amazon Q is running on Amazon's own Titan G1 model. I recently ran the "Premier" version (their highest end one) through my personal vibecheck test and was quite surprised by its RL. It was the only non-Chinese model I've tested to refuse to answer about Tiananmen Square and the only model I believe I've tested with this eval (over 50 at this point) that refused to answer about the LA riots. It also scored an impressive 0/6 on my reasoning/basic world understanding tests (underperforming most 3B models) but that's more capabilities than RL...

Amazon claims the Titan model is suitable for: "Supported use cases: RAG, agents, chat, chain of thought, open-ended text generation, brainstorming, summarization, code generation, table creation, data formatting, paraphrasing, rewriting, extraction, and Q&A." (it is not, lol)

replies(1): >>40668902 #
7. TeMPOraL ◴[] No.40667649{3}[source]
"If you can't take a little bloody nose, maybe you ought to go back home and crawl under your bed. It's not safe out here. It's wondrous, with treasures to satiate desires both subtle and gross... but it's not for the timid."

So sure, the LLM occasionally pranks someone, in a way similar to how random Internet posts do; it is confidently wrong, in a way similar to how most text on the Internet is confidently wrong because content marketers don't give a damn about correctness, that's not what the text is there for. As much as this state of things pains me, general population has mostly adapted.

Meanwhile, people who would appreciate a model that's 99% right on things where the 1% is costly, rightfully continue to ignore Gemini and other models by companies too afraid to play in the field for real.

replies(2): >>40667683 #>>40667933 #
8. rockskon ◴[] No.40667683{4}[source]
AI is not like some random person posting on the Internet.

A random person on the Internet often has surrounding context to help discern trustworthiness. A researcher can also query multiple sources to determine how much there is concensus about.

You can't do that with LLMs.

I cannot stress strongly enough that direct comparisons between LLMs and experts on the Internet are inappropriate.

replies(2): >>40667739 #>>40667813 #
9. TeMPOraL ◴[] No.40667739{5}[source]
> I cannot stress strongly enough that direct comparisons between LLMs and experts on the Internet are inappropriate.

In this context, I very much agree. But I'd like to stress that "experts on the Internet" is not what 99% of the users read 99% of the time, because that's not what search engines surface by default. When you make e.g. food or law or health-related queries, what you get back isn't written by experts - it's written by content marketers. Never confuse the two.

> A researcher can also query multiple sources to determine how much there is concensus about.

> You can't do that with LLMs.

A person like that will know LLMs hallucinate, and query multiple sources and/or their own knowledge, and/or even re-query the LLM several times. Such people are not in danger - but very much annoyed when perfectly reasonable queries get rejected on the grounds of "safety".

10. Y_Y ◴[] No.40667813{5}[source]
Why can't you estimate the trustworthiness of an LLM? I happen to think that you can, and that the above analogy was fine. You don't need to read someone's forum history to know you shouldn't to trust them on something high-stakes. Maybe instead of strongly stressing you should present a convincing argument.
replies(1): >>40673476 #
11. pjc50 ◴[] No.40667933{4}[source]
The only underlying question here is "who is liable for the output of the LLM?"

I just don't think the "nobody is" current solution is going to last in the current litigious environment.

replies(2): >>40668091 #>>40670649 #
12. TeMPOraL ◴[] No.40668091{5}[source]
Good point. Since LLM isn't a person, this leaves only the vendor and the user as liable parties. That's one less legal person than in regular search, where you have the user, the search engine vendor, and the author/publisher of the content involved in a harm scenario.

What is the consensus on liability in case of regular web search? Your comment made me realize that I never thought much about it in 20+ years of using the Internet; I kind of always assumed it's all on the user.

replies(1): >>40668370 #
13. pjc50 ◴[] No.40668370{6}[source]
> What is the consensus on liability in case of regular web search? Your comment made me realize that I never thought much about it in 20+ years of using the Internet

Have you never noticed those "google has removed some results to comply with the DMCA" notices?

replies(2): >>40668715 #>>40670689 #
14. voxic11 ◴[] No.40668715{7}[source]
But the reason we "needed" the DMCA is because they wouldn't have been liable under existing law, and the DMCA only covers copyright violations.
15. malfist ◴[] No.40668902[source]
It is Titian under the hood. And it's absolutely crap.

Also fun fact, Titan's image generator will refuse any prompt that references Bezos because it "violates content policy"

If you want to do something useful on bedrock use Claude

replies(1): >>40669165 #
16. chuckadams ◴[] No.40669069[source]
I once asked Q to help me fix a broken policy (turns out we were using the wrong thing for the resource name). It gave me some completely unrelated documentation about setting up Cogito. I've never seen an AI as laughably bad as Q.
17. lhl ◴[] No.40669165{3}[source]
I've been poking around this week and there's actually quite a few useful models on Bedrock (this is region dependent!) https://docs.aws.amazon.com/bedrock/latest/userguide/models-...

Claude Opus is supposedly only available in us-west-2, but is listed as "Unavailable" for me (Sonnet and Haiku are available). Cohere's Command R+ is also available and while less capable, for instruction following, I believe its superior to Anthropic's models. There's also Llama 3 70B Instruct and Mistral Large, both which are good for general tasks.

For those that haven't been closely following/testing the models available, I think Artificial Analysis' Quality vs Price charts isn't too bad a place to start https://artificialanalysis.ai/models although if you have specific tasks, it's best to eval some models are surprisingly good/bad at specific things.

Titan appears to be bad at everything though.

replies(1): >>40670561 #
18. gverrilla ◴[] No.40669289[source]
Tried Amazon Q a few times, it was NEVER able to provide any help. Why do they keep that crap?
19. ◴[] No.40669327[source]
20. spmurrayzzz ◴[] No.40670561{4}[source]
> cohere's Command R+ is also available and while less capable, for instruction following, I believe its superior to Anthropic's models

My experience recently is that its actually noticeably better for instruction following than Claude, but can be finicky if you're not careful about adhering to the prompt template. But between the RAG and multi-step tool use capabilities, even if it was slightly worse on the instruction-following side of things I'd still say, as you do, thats its much better than Claude on average.

Agree on titan as well. I recently was forced into a meeting with our AWS TAM, and they kept shoehorning Q into every conversation. I held my tongue knowing that titan was the model powering it under the hood.

21. raxxorraxor ◴[] No.40670649{5}[source]
The person who prompts would be responsible. Everything else doesn't really make sense. This is usually the trivial solution for any form of tool we use.
replies(1): >>40671186 #
22. realusername ◴[] No.40670689{7}[source]
The DMCA is the copyright industry's response to "nobody is liable for results" which was the statu quo before.
23. wumbo ◴[] No.40671186{6}[source]
If there’s going to be a lawsuit, go after Colt before Anthropic.
24. DonsDiscountGas ◴[] No.40671251[source]
In fairness to Amazon Q, the AWS docs are pretty confusing. Maybe it was just embarrassed and made an excuse. (Sidenote to Amazon and others: an LLM is a supplement to good documentation, not a replacement)
replies(1): >>40679038 #
25. rockskon ◴[] No.40673476{6}[source]
Because if I already knew the answer then I wouldn't be asking the LLM?
26. throwaway48476 ◴[] No.40679038[source]
The LLM requires good documentation to train on.