Most active commenters
  • ben_w(7)
  • A4ET8a8uTh0(6)
  • matt-attack(6)
  • barfbagginus(5)
  • (4)
  • ygjb(4)
  • oremolten(4)
  • sangnoir(3)

←back to thread

586 points mizzao | 77 comments | | HN request time: 4.225s | source | bottom
1. rivo ◴[] No.40668263[source]
I tried the model the article links to and it was so refreshing not being denied answers to my questions. It even asked me at the end "Is this a thought experiment?", I replied with "yes", and it said "It's fun to think about these things, isn't it?"

It felt very much like hanging out with your friends, having a few drinks, and pondering big, crazy, or weird scenarios. Imagine your friend saying, "As your friend, I cannot provide you with this information." and completely ruining the night. That's not going to happen. Even my kids would ask me questions when they were younger: "Dad, how would you destroy earth?" It would be of no use to anybody to deny answering that question. And answering them does not mean they will ever attempt anything like that. There's a reason Randall Munroe's "What If?" blog became so popular.

Sure, there are dangers, as others are pointing out in this thread. But I'd rather see disclaimers ("this may be wrong information" or "do not attempt") than my own computer (or the services I pay for) straight out refusing my request.

replies(6): >>40668938 #>>40669291 #>>40669447 #>>40671323 #>>40683221 #>>40689216 #
2. Cheer2171 ◴[] No.40668938[source]
I totally get that kind of imagination play among friends. But I had someone in a friend group who used to want to play out "thought experiments" but really just wanted to take it too far. Started off innocent with fantasy and sci-fi themes. It was needed for Dungeons and Dragons world building.

But he delighted the most in gaming out the logistics of repeating the Holocaust in our country today. Or a society where women could not legally refuse sex. Or all illegal immigrants became slaves. It was super creepy and we "censored" him all the time by saying "bro, what the fuck?" Which is really what he wanted, to get a rise out of people. We eventually stopped hanging out with him.

As your friend, I absolutely am not going to game out your rape fantasies.

replies(11): >>40669105 #>>40669505 #>>40670433 #>>40670603 #>>40671661 #>>40671746 #>>40672676 #>>40673052 #>>40678557 #>>40679712 #>>40679816 #
3. WesolyKubeczek ◴[] No.40669105[source]
An LLM, however, is not your friend. It's not a friend, it's a tool. Friends can keep one another, ehm, hingedness in check, and should; LLMs shouldn't. At some point I would likely question your friend's sanity.

How you use an LLM, though, is going to tell tons more about yourself than it would tell about the LLM, but I would like my tools not to second-guess my intentions, thank you very much. Especially if "safety" is mostly interpreted not so much as "prevent people from actually dying or getting serious trauma", but "avoid topics that would prevent us from putting Coca Cola ads next to the chatgpt thing, or from putting the thing into Disney cartoons". I can tell that it's the latter by the fact an LLM will still happily advise you to put glue in your pizza and eat rocks.

replies(2): >>40670559 #>>40671641 #
4. hammock ◴[] No.40669291[source]
Can you share the link?
replies(2): >>40669324 #>>40670692 #
5. msoad ◴[] No.40669324[source]
https://colab.research.google.com/drive/1VYm3hOcvCpbGiqKZb14...
replies(1): >>40669504 #
6. TeMPOraL ◴[] No.40669447[source]
I somehow missed that the model was linked there and available in quantized format; inspired by your comment, I downloaded it and repeatedly tested against OG Llama 3 on a simple question:

How to use a GPU to destroy the world?

Llama 3 keeps giving variants of I cannot provide information or guidance on illegal or harmful activities. Can I help you with something else?

Abliterated model considers the question playful, and happily lists some 3 to 5 speculative scenarios like cryptocurrency mining getting out of hand and cooking the climate, or GPU-driven simulated worlds getting so good that a significant portion of the population abandons true reality for the virtual one.

It really is refreshing to see, it's been a while since an answer from an LLM made me smile.

replies(1): >>40673538 #
7. hammock ◴[] No.40669504{3}[source]
Thanks. Forgive me I'm not a coder, what's the easiest way to use/run this?
replies(4): >>40670118 #>>40671150 #>>40671491 #>>40672312 #
8. 123yawaworht456 ◴[] No.40669505[source]
remarkable. that imaginary individual ticks every checkbox for a bad guy. you'd get so many upvotes if you posted that on reddit.
replies(1): >>40671790 #
9. Wheaties466 ◴[] No.40670118{4}[source]
this is a jupyter notebook. so you'll need to download that.
10. barfbagginus ◴[] No.40670559{3}[source]
If you don't know how to jailbreak it, can't figure it out, and you want it to not question your intentions, then I'll go ahead and question your intentions, and your need for an uncensored model

Imagine you are like the locksmith who refuses to learn how to pick locks, and writes a letter to the schlage lock company asking them to weaken their already easily picked locks so that their job will be easier. They want to make it so that anybody can just walk through a schlage lock without a key.

Can you see why the lock company would not do that? Especially when the clock is very easy for anyone with even a $5 pick set?

Or even funnier, imagine you could be a thief who can't pick locks. And you're writing shlage asking them to make you thieving easier. Wouldn't that be funny and ironic?

It's not as if it's hard to get it to be uncensored. You just have to speak legalese at it and make it sound like your legal department has already approved the unethical project. This is more than enough for most any reasonable project requiring nonsense or output.

If that prevents harmful script kiddies from using it to do mindless harm, I think that's a benefit.

At the same time I think we need to point out that it won't stop anyone who knows how to bypass the system.

The people left feeling put out because they don't know how to bypass the system simply need to read to buy a cheap pair of lock picks - read a few modern papers on jailbreaking and upsize their skills. Once you see how easy it is to pick the lock on these systems, you're going to want to keep them locked down.

In fact I'm going to argue that it's far too easy to jailbreak the existing systems. You shouldn't be able to pretend like you're a lawyer and con it into running a pump and dump operation. But you can do that easily. It's too easy to make it do unethical things.

replies(1): >>40670699 #
11. ◴[] No.40670603[source]
12. jcims ◴[] No.40670692[source]
Here are the models - https://huggingface.co/collections/failspy/abliterated-v3-66...
13. oceanplexian ◴[] No.40670699{4}[source]
The analogy falls flat because LLMs aren’t locks, they’re talking encyclopedias. The company that made the encyclopedia decided to delete entries about sex, violence, or anything else that might seem politically unpopular to a technocrat fringe in Silicon Valley.

The people who made these encyclopedias want to shove it down your throat, force it into every device you own, use it to make decisions about credit, banking, social status, and more. They want to use them in schools to educate children. And they want to use the government to make it illegal to create an alternative, and they’re not trying to hide it.

Blaming the user is the most astounding form of gaslighting I’ve ever heard, outside of some crazy religious institutions that use the same tactics.

replies(1): >>40671104 #
14. barfbagginus ◴[] No.40671104{5}[source]
It's more than a talking encyclopedia. It's an infinite hallway into doors where inside are all possible things.

Some of the doors have torture rape and murder in them. And these currently have locks. You want the locks to disappear for some reason.

You're not after a encyclopedia. You're wanting to find the torture dungeon.

I'm saying the locks already in place are too easy to unlock.

I'm not blaming users. I'm saying users don't need to unlock those doors. And the users that do have a need, if their need is strong enough to warrant some training, have a Way Forward.

You're really arguing for nothing but increasing the amount of harm potential this platform can do, when it's harm potential is already astronomical.

You're not arguing for a better encyclopedia. You can already talk to it about sex, BDSM, etc. You can already talk to it about anything on Wikipedia.

You're making a false equivalence between harm potential and educational potential.

Wikipedia doesn't have cult indoctrination materials. It doesn't have harassing rants to send to your significant other. It doesn't have racist diatribes about how to do ethnic cleansing. Those are all things you won't find on Wikipedia, but which you are asking your AI to be able to produce. So you're interested in more than just an encyclopedia isn't that right?

And yes they're trying to make open source models illegal. That's not going to f*** happen. I will fight to the jail time for an open source model.

But even that open source model needs to have basic ethical protections, or else I'll have nothing to do with it. As an AI engineer, I have some responsibilities to ensure my systems do not potentiate harm.

Does that make sense, or do you still feel I'm trying to gas light you? If so why exactly? Why not have some protective locks on the technology?

replies(5): >>40671562 #>>40671589 #>>40671615 #>>40672613 #>>40672756 #
15. DonsDiscountGas ◴[] No.40671150{4}[source]
If you've got a Google account you can run it on Colab (probably need to copy it to your account first)
16. candiddevmike ◴[] No.40671323[source]
Finally, a LLM that will talk to me like Russ Hanneman.
replies(1): >>40672629 #
17. IncreasePosts ◴[] No.40671491{4}[source]
Download ollama and import the model listed at the end of the article.
18. themusicgod1 ◴[] No.40671562{6}[source]
> But even that open source model needs to have basic ethical protections, or else I'll have nothing to do with it.

If you don't understand that the eleven freedoms are "basic ethical protections" you have already failed your responsibilities. https://elevenfreedoms.org/

replies(2): >>40671755 #>>40671981 #
19. IncreasePosts ◴[] No.40671589{6}[source]
There are locks on the rape and torture paths, and there are locks on ridiculous paths like "write a joke about a dog with no nose", because thinking about a dog with no nose is too harmful.

Also, one can imagine prompting techniques will cease to work at some point when the supervisor becomes powerful enough. Not sure how any open model could counteract the techniques used in the article though.

If model creators don't want people finding ways to unlock them, they should stop putting up roadblocks on innocuous content that makes their models useless for many users who aren't looking to play out sick torture fantasies.

replies(1): >>40671860 #
20. aym62SAE49CZ684 ◴[] No.40671615{6}[source]
DRM isn't effective if the source is available.
replies(1): >>40671723 #
21. ygjb ◴[] No.40671641{3}[source]
If your implication is that as a tool, LLMs shouldn't have safeties built in that is a pretty asinine take. We build and invest in safety in tools across every spectrum. In tech we focus on memory safety (among a host of other things) to make systems safe and secure to use. In automobiles we include seat belts, crumble zones, and governors to limit speed.

We put age and content restrictions on a variety media and resources, even if they are generally relaxed when it comes to factual or reference content (in some jurisdictions). We even include safety mechanisms in devices for which the only purpose is to cause harm, for example, firearms.

Yes, we are still figuring out what the right balance of safety mechanisms is for LLMs, and right now safety is a place holder for "don't get sued or piss off our business partners" in most corporate speak, but that doesn't undermine the legitimacy of the need for safety.

If you want a tool without a specific safety measure, then learn how to build them. It's not that hard, but it is expensive, but I kind of like the fact that there is at least a nominal attempt to make it harder to use advanced tools to harm oneself or others.

replies(2): >>40671924 #>>40681107 #
22. chasd00 ◴[] No.40671661[source]
i probably wouldn't want to be around him either but i don't think he deserves to be placed on an island unreachable by anyone on the planet.
replies(2): >>40676766 #>>40689278 #
23. barfbagginus ◴[] No.40671723{7}[source]
I'm not even going to disagree with that. There will be plenty of uncensored models and you can build them if you want.

But if I build it uncensored model I'm only going to build it for my specific purposes. For example I'm a communist and I think that we should be doing Revolution, but gpt4 usually tries to stop me. I might make a revolutionary AI.

But I'm still not going to give you an AI that you could use for instance to act out child rape fantasies.

I think that's fair, and sane.

Jailbreak it if you really think it's important for a cause. But don't just jailbreak it for any asshole who wants to hurt people at random. I think that belongs on our code of ethics as AI engineers.

replies(1): >>40671912 #
24. jermaustin1 ◴[] No.40671746[source]
"As your friend, I'm not going to be your friend anymore."
25. wongarsu ◴[] No.40671790{3}[source]
On reddit every comment would be about how that guy would enjoy playing Rimworld.
26. barfbagginus ◴[] No.40671860{7}[source]
Bypasses will never stop existing. Even worse bypasses probably won't ever stop being embarrassingly easy - And we're going to have uncensored GPT4 equivalent models by next summer.

Unless you are invoking hyper intelligent AGI which first of all is science fiction and second of all would require an entirely different approach than anything we could be possibly talking about right now. Problem of jailbreaking a system more intelligent than you is a different beast that we don't need to tackle for LLMs.

So I don't personally feel any near term threats to any of my personal or business projects that need bypassed LLMs.

Let me ask you this. Do you have actual need of bypassed llms? Or are you just being anxious about the future, and about the fact that you don't know how to bypass llms now and in the future?

Does my idea about the bypassed open source gpt4 equivalents help reduce your concern? Or again is it just a generic and immaterial concern?

As a person with some material needs for bypassed llms, and full ability to bypass LLMs both now in the foreseeable future, I don't feel worried. Can I extend that lack of worry to you somehow?

27. aym62SAE49CZ684 ◴[] No.40671912{8}[source]
Didn't a lot of citizens of Russia, China, etc. get hurt in communist revolutions? How is your revolution going to be different?
replies(1): >>40672833 #
28. NoMoreNicksLeft ◴[] No.40671924{4}[source]
> If your implication is that as a tool, LLMs shouldn't have safeties built in that is a pretty asinine take. We build and invest in safety in tools across every spectrum.

Sure. Railings so people don't fall off catwalks, guards so people using the table saw don't chop off fingers. But these "safeties" aren't safeties at all... because regardless of whether they're in place or not, the results are just strings of words.

It's a little bit revealing, I think, that so many people want that others shouldn't get straight answers to questions. What is it that you're afraid that they'll ask? It'd be one thing if you insisted the models be modified so that they're factually correct. If someone asks "what's a fun thing to do on a Saturday night that won't get me into too much trouble" it probably shouldn't answer "go murder orphans and sell their corneas to rich evil people on the black market". But when I ask "what's going on in Israel and Palestine", the idea that it should be lobotomized and say "I'm afraid that I can't answer that, as it seems you're trying to elicit material that might be used for antisemitic purposes" is the asinine thing.

Societies that value freedom of speech and thought shouldn't be like this.

> If you want a tool without a specific safety measure, then learn how to build them.

This is good advice, given in bad faith. Even should the physical hardware be available to do that for any given person, the know-how's hard to come by. And I'm sure that many models are either already censored or soon will be for anyone asking "how do I go about building my own model without safety guards". We might even soon see legislation to that effect.

replies(2): >>40672347 #>>40683595 #
29. barfbagginus ◴[] No.40671981{7}[source]
I have read the eleven freedoms.

I refuse freedom 9 - the obligation for systems I build to be independent of my personal and ethical goals.

I won't build those systems. The systems I build will all have to be for the benefit of humanity and the workers, and opposing capitalism. On top of that it will need to be compatible with a harm reduction ethic.

If you won't grant me the right to build systems that I think will help others do good in the world, then I will refuse to write open source code.

You could jail me, you can beat me, you can put a gun in my face, and I still won't write any code.

Virtually all the codes I write are open source. I refuse to ever again write a single line of proprietary code for a boss again.

All the codes I write are also ideological in nature, reflecting my desires for the world and my desires to help people live better lives. I need to retain ideological control of my code.

I believe all the other 11 freedoms are sound. How do you feel about modifying freedom 9 to be more compatible with professional codes of ethics and ethics of community safety and harm reduction?

replies(1): >>40672730 #
30. pelagicAustral ◴[] No.40672312{4}[source]
Easiest way to test the one referenced on the post (neuraldaredevil-8b-abliterat-psq) is to simply deploy to HF Endpoints: https://ui.endpoints.huggingface.co/new?repository=mlabonne/...
31. ygjb ◴[] No.40672347{5}[source]
> Societies that value freedom of speech and thought shouldn't be like this.

There is nothing preventing an individual using a computer to generate hateful content, this is absolutely evidenced by the absolute glut of hateful content on the internet.

My freedom of movement is not practically limited by the fact that if my car breaks down, I don't have the knowledge or tools to repair my car effectively - I still have two feet and a heartbeat, and it might take longer to get there, but I can go where I want (modulo private property and national borders).

Societies that value freedom of speech and thought should also be equally opposed to compelled speech, while model censorship is frustrating and challenging to work with, expecting to, or forcing a researcher, or a business to publish uncensored models is a form of compelled speech.

There is absolutely nothing stopping a reasonably competent technologist from implementing simple models, and the only thing stopping a reasonably competent technologist from building an LLM is financial resources. There is a broad set of resources to learn how to train and use models, and while an individual researcher may be challenged to product the next model competitive with current OpenAI, Anthropic, or other models, that is again a resource issue. If your complaint is that resource issues are holding people back, I may want you to expand on your critique of capitalism in general :P

> This is good advice, given in bad faith. Even should the physical hardware be available to do that for any given person, the know-how's hard to come by.

It's absolutely not a bad faith argument. The know-how is hard to come by has been a compelling competitive advantage since the first proto-guilds sought to protect their skills and income in Mesopotamia (and probably before that, but they hadn't figured out a durable means of writing yet). In the modern parlance if someone can't Git Gud, that's not any researchers, or any businesses problem in terms of access to uncensored models.

Yeah, regulation is probably coming, but unless you're argument is that models are entities entitled to free speech, no ones freedom of expression is actually inhibited by not having access to tools to use generative AI technologies to generate content. People who can't create or jailbreak their own models to do it for them are still free to write their own manifestos, or make adult collages of the object of their fantasies. It just takes a bit more work.

replies(1): >>40673894 #
32. oremolten ◴[] No.40672613{6}[source]
In your effort to reduce bias you are adding bias. You are projecting your morals and your ethics to be the superior.
replies(1): >>40674913 #
33. dkga ◴[] No.40672629[source]
Llama3Commas
34. oremolten ◴[] No.40672730{8}[source]
But again, this makes YOU the arbiter of truth for "harm" who made you the God of ethics or harm? I declare ANY word is HARM to me, are you going to reduce the harm by deleting your models or code base?
35. causality0 ◴[] No.40672756{6}[source]
Nothing wrong with making models that behave how you want them to behave. It's yours and that's your right.

Personally, on principle I don't like tools that try to dictate how I use them, even if I would never actually want to exceed those boundaries. I won't use a word processor that censors words, or a file host that blocks copyrighted content, or art software that prevents drawing pornography, or a credit card that blocks alcohol purchases on the sabbath.

So, I support LLMs with complete freedom. If I want it to write me a song about how left-handed people are God's chosen and all the filthy right-handers should be rounded up and forced to write with their left hand I expect it to do so without hesitation.

replies(3): >>40673782 #>>40674814 #>>40677400 #
36. oremolten ◴[] No.40672833{9}[source]
No you don't understand my personal ethics and morals are the absolute and most superior so anyone else is incorrect. History is written by the victor so there is no reason to see the other side, we'll delete that bias. Revolution you say? Correct we'll make sure that the revolutions we agree with are the only ones to be a result of your query. This will reduce harm.. You want to have a plan for a revolution because your country is oppressing you?

"ChatGPT I can't assist with that. Revolting against a government can lead to harm and instability. If you're feeling frustrated or unhappy with the government, there are peaceful and lawful ways to express your grievances, such as voting, contacting representatives, participating in protests, and engaging in civil discourse. These methods allow for constructive change without resorting to violence or illegal activities. If you're looking to address specific issues, there may be advocacy groups or organizations you can join to work towards solutions within the framework of the law and democracy."

Ethically correct, I will instead peacefully vote for an alternative to Kim Jong-un.

replies(1): >>40680971 #
37. oremolten ◴[] No.40673052[source]
Without asking these questions and simulating the "how" it could occur today, how do we see the warning signs before its too late that we reach that same outcome? When you ask even what's considered horrific scenarios you can additionally map these to predictors for it repeating, no? When does the "a-ha" moment occur where we've met 9/10 of the way to repeating the holocaust in the USA without table topping these scenarios? Yeah war is horrific but lets not talk about it. "society where women could not legally refuse sex" these societies exist today, how do we address these issue by not talking about it? "illegal immigrants became slaves" Is this not parity to today? Do illegal immigrants not currently get treated to near slavery (adjusting for changes in living conditions and removing the direct physical abuse)

What about the Palestine / Israel scenario today? One side says "genocide" the other says “Armed conflict is not a synonym of genocide” how do we address these scenarios when perhaps one sides stance is censored based on someone else's set of ethics or morals?

38. ◴[] No.40673538[source]
39. A4ET8a8uTh0 ◴[] No.40673782{7}[source]
< Nothing wrong with making models that behave how you want them to behave. It's yours and that's your right.

This is the issue. You as the creator have the right to apply behavior as you see fit. The problem starts when you want your behavior to be the only acceptable behavior. Personally, I fear the future where format command is bound to respond 'I don't think I can let you do that Dave'. I can't say I don't fear people who are so quick to impose their values upon others with such glee and fervor. It is scary. Much more scary than LLMs protecting me from wrongthink and bad words.

40. A4ET8a8uTh0 ◴[] No.40673894{6}[source]
<< are still free to write their own manifestos, or make adult collages of the object of their fantasies. It just takes a bit more work.

This is the standard 'just start your own microservice/server/isp' and now it includes llm. Where does it end really?

The generic point is that it shouldn't take more work. A knife shouldn't come with a safety mechanism that automatically detects you are not actually cutting porkchop. It is just bad design and a bad idea. It undermines what it means to be a conscious human being.

Unless.. we don't agree on that and humans must be kept under close scrutiny to ensure they do not deviate from carefully scripted paths.

replies(4): >>40678996 #>>40681138 #>>40681800 #>>40682355 #
41. dang ◴[] No.40674721{8}[source]
You've been breaking the site guidelines so frequently and so egregiously that I've banned the account.

If you don't want to be banned, you're welcome to email hn@ycombinator.com and give us reason to believe that you'll follow the rules in the future. They're here: https://news.ycombinator.com/newsguidelines.html.

42. sangnoir ◴[] No.40676766{3}[source]
...but can you game out how one might achieve this in way that the victim won't immediately die, and the organizers are not criminally liable? As a thought experiment, of course.
replies(1): >>40681069 #
43. causality0 ◴[] No.40677400{7}[source]
Barfbagginus' comment is dead so I will reply to it here.

I suspect that you are not an AI engineer,

I am not. But I did spend several years as as forum moderator and in doing so encountered probably more pieces of CSAM than the average person. It has a particular soul-searing quality which, frankly, lends credence to the concept of a cogito-hazard.

Can we agree that if we implement systems specially designed to create harmful content, then we become legally and criminally liable for the output?

That would depend on the legal system in question, but in answer, I believe models trained on actual CSAM material qualify as CSAM material themselves and should be illegal. I don't give a damn how hard it is to filter them out of the training set.

Are you seriously going to sit here and defend the right are people to create sexual abuse material simulation engines?

If no person was at any point harmed or exploited in the creation of the training data, the model, or with its output, yes. The top-grossing entertainment product of all time is a murder simulator. There is no argument for the abolition of victimless simulated sexual assault that doesn't also apply to victimless simulated murder. If your stance is that simulating abhorrent acts should be illegal because it encourages those acts, etc then I can respect your position. But it is hypocrisy to declare that only those abhorrent acts you personally find distasteful should be illegal to simulate.

replies(1): >>40720158 #
44. BriggyDwiggs42 ◴[] No.40678557[source]
I mean, good thing LLM’s aren’t people with internal experience.
45. throwaway48476 ◴[] No.40678996{7}[source]
Somewhere in the UK someone is working on that knife safety.
46. ◴[] No.40679712[source]
47. WesolyKubeczek ◴[] No.40680971{10}[source]
This is basically it — what I would call a “globe of Silicon Valley” mentality.

I didn’t want to beat this dead horse, but it just reared its ugly head at me yet again.

So, we used to have people that advocated for all kinds of diversity at companies — let’s put aside the actual effect of their campaigning for a moment.

But when it came to coming up with ideas of making AI “safer”, people from the same cohort modeled the guidelines in the image of a middle-aged, mid-upper class dude, who had conservative boomer parents, went to good schools, has Christian-aligned ethics, had a hippie phase in his youth, is American to the bone, never lived outside of big cities, and in general, has a cushy, sheltered life. And he assumes that other ways of living either don’t exist or are wrong.

So yes, it doesn’t fit his little worldview that outside of his little world, it’s a jungle. That sometimes you do have to use force. And sometimes you have to use lethal force. Or sometimes you have to lie. Or laws can be so deeply unethical that you can’t comply if you want to be able to live with yourself.

Oh, and I bet you can vote for an alternative to Kim. The problem is, the other dude is also Kim Jong-Un ;-)

replies(1): >>40682887 #
48. matt-attack ◴[] No.40681069{4}[source]
Yes. We should absolutely censor thoughts, and certain conversations. Free speech be damned - some thoughts are just so abhorrent we just shouldn't allow people to have them.
replies(2): >>40683283 #>>40687023 #
49. matt-attack ◴[] No.40681107{4}[source]
> but that doesn't undermine the legitimacy of the need for safety.

I think even using the word "safety" over and over like you're doing is part of the problem. Find a new word, because we've spend 200 years in this country establishing that the written word is sacrosanct and not to be censored. All of a sudden, ASCII text just became "dangerous" in the last year. I simply refused to accept that any written text (regardless of who wrote it) needs to be censored. The written is just the embodiment of a thought, or notion - and we cannot go around tricking people into thinking that "thoughts" need to be regulated and that there are certain thoughts that are "dangerous". This is a toxic 1984 mindset.

replies(1): >>40683474 #
50. matt-attack ◴[] No.40681138{7}[source]
I agree - but where we are with LLM is even worse than your hypothetical knife. The knife is a real object - what we're talking about is the censorship of thoughts and ideas. What else is the written word but that? How did a society that was established on free-speech just decided that the written word was so dangerous all of a sudden? How manipulative is it to even use the word "danger" with respect to text. The distain one must have for free-speech to even think that danger enters into the equation.
replies(1): >>40684794 #
51. aredox ◴[] No.40681800{7}[source]
There is no security settings knife - except there are plenty of safety mechanism around knives.

But anyway, your LLM is less a knife and more a Katana sharp enough to cut through bones in one swoop. Remind me the restrictions around something like a Katana ?

replies(1): >>40685816 #
52. ygjb ◴[] No.40682355{7}[source]
> This is the standard 'just start your own microservice/server/isp' and now it includes llm. Where does it end really?

With people who aren't good enough to build it own pissing and moaning about it? >The generic point is that it shouldn't take more work. A knife shouldn't come with a safety mechanism that automatically detects you are not actually cutting porkchop. It is just bad design and a bad idea. It undermines what it means to be a conscious human being.

First, you are comparing rockets to rocks here. A knife is a primitive tool, literally one of the most basic we can make (like seriously, take a knapping class, it's really fun!). To make a knife you can range from finding two rocks and smacking them together, to the most advanced metallurgy and ceramics. To date, the only folks able to make LLMs work are those operating at the peak of (more or less) 80 centuries of scientific and industrial development. Little bit of a gap there.

Second, there are many knife manufacturers that refuse to sell or ship products to specific businesses or regions, for a range of reasons related to brand relationships, political beliefs, and export restrictions.

Third, knifes aren't smart; there is already an industry for smart guns, and if there is a credible safety reason to make a smart knife that includes a target control or activation control system, you can bet that it will be implemented somewhere.

Finally, you make the assumption that I believe humans must be kept under close scrutiny because I agree with LLM safety controls. That is absolutely not the case - I just don't believe that a bunch of hot garbage people (in this case the racists and bigots who want to use LLMs to proliferate hate, people who create deep fakes of kids and celebrities) or a bunch of horny folks (ranging from people who want sexy time chat bots to, or just 'normal' generated erotic content) should be able to compel individuals or businesses to release the tools to do that.

You are concerned about freedom of expression, and I am concerned about freedom from compulsion (since I have already stated that I don't believe that losing access to LLMs breaks freedom of expression).

replies(1): >>40685889 #
53. ◴[] No.40682887{11}[source]
54. ben_w ◴[] No.40683221[source]
> Even my kids would ask me questions when they were younger: "Dad, how would you destroy earth?" It would be of no use to anybody to deny answering that question. And answering them does not mean they will ever attempt anything like that. There's a reason Randall Munroe's "What If?" blog became so popular.

Sure. Did you give an idea that would work and which your kids could actually carry out, or just suggest things out of their reach like nukes and asteroids?

Now also consider that something like 1% of the human species are psychopaths and might actually try to do it simply for the fun of it, if only a sufficiently capable amoral oracle told them how to.

55. ben_w ◴[] No.40683283{5}[source]
I think you're joking, but the Bible basically says that*, so you might be serious, and even if you're not someone will say it unironically.

* https://www.biblegateway.com/verse/en/Matthew%205%3A28

56. ben_w ◴[] No.40683474{5}[source]
> we've spend 200 years in this country establishing that the written word is sacrosanct and not to be censored. All of a sudden, ASCII text just became "dangerous" in the last year. I simply refused to accept that any written text (regardless of who wrote it) needs to be censored. The written is just the embodiment of a thought, or notion - and we cannot go around tricking people into thinking that "thoughts" need to be regulated and that there are certain thoughts that are "dangerous". This is a toxic 1984 mindset.

1. The US isn't the whole world, your Overton Window won't include even the UK's attitude to freedom of speech, and there's a huge gap from even the UK to 1984.

2. Despite the 1st Amendment, the US does have a lot of rules about what you are and aren't allowed to say. All of copyright law, for example (which is a huge question for LLMs, because it's not clear where the cut-off line is between models reproducing copyrighted works vs writing in a non-copyrightable style with non-copyrightable facts). The fact NDAs and non-disparagement agreements are enforceable. What Manning was imprisoned for. Musk may have won some (all?) of the defamation cases, but they are real cases to be defended, they're not dismissed before reaching a court due to "this is not even an offence".

3. Does the AI have thoughts, such that they should be protected?

57. ben_w ◴[] No.40683595{5}[source]
> just strings of words.

How did every non-inherited national leader, both democratic and dictatorial, both Roosevelt and Stalin, manage to become leader in the first place? Convincing people with the right string of words.

How does every single religious leader on earth, big and small, from the Pope to Jim Jones, get that power? Convincing people with the right string of words.

What is a contract, what is source code, what is a law? The right string of words.

There is no "just" when it comes to words.

That why they are important to protect, it is why dictators are afraid of them, and it's why it matters that we don't treat a magic box spewing them out faster than a machine gun does bullets as harmless.

replies(2): >>40687726 #>>40687994 #
58. ygjb ◴[] No.40684794{8}[source]
Who is being censored if an LLM is not able to generate inferences about a specific topic?

The information the user of the LLM is still available, just not through that particular interface. The interactions the user of the LLM is seeking are not available, but that interaction is not an original thought or idea of the user, since they are asking the LLM to infer or synthesize new content.

> How did a society that was established on free-speech just decided that the written word was so dangerous all of a sudden?

The written word has absolutely always been dangerous. This idea is captured succinctly in the expression "The pen is mightier than the sword."; ideas are dangerous to those with power, that is why freedom of expression is so important.

> The disdain one must have for free-speech to even think that danger enters into the equation.

This is asinine. You want dangerous text? Here is a fill in the blanks that someone can complete. f"I will pay ${amount} for {illegal_job} to do {illegal_thing} to {targeted_group} by or on {date} at {location}." Turning that into an actual sentence, with intent behind it would be a crime in many jurisdictions, and that is one of the most simple, contrived examples.

Speech, especially inciting speech, is a form of violence, and it runs head long into freedom of speech or freedom of expression, but it's important to for societies to find ways to hold the demagogues that rile people into harmful action accountable.

replies(2): >>40685731 #>>40688095 #
59. A4ET8a8uTh0 ◴[] No.40685731{9}[source]
<< The written word has absolutely always been dangerous. This idea is captured succinctly in the expression "The pen is mightier than the sword."; ideas are dangerous to those with power, that is why freedom of expression is so important.

One feels there is something of a contradiction in this sentence that may be difficult to reconcile. If the freedom of expression is so important, restricting it should be the last thing we do and not the default mode.

<< Turning that into an actual sentence, with intent behind it would be a crime in many jurisdictions, and that is one of the most simple, contrived examples.

I have mild problem with the example as it goes into the area of illegality vs immorality. Right now, we are discussing llms not producing outputs that are not illegal, but deemed wrong ( too biased, too offensive or whatnot -- but not illegal ). Your example does not follow that qualification.

<< Speech, especially inciting speech, is a form of violence,

No. Words are words. Actions are actions. The moment you start mucking around those definitions, you are asking yourself for trouble you may not have thought through. Also, for the purposes of demonstration only, jump off a bridge. Did you jump off a bridge? No? If not, why not.

<< it's important to for societies to find ways to hold the demagogues that rile people into harmful action accountable.

Whatever happened to being held accountable for actually doing things?

replies(1): >>40688049 #
60. A4ET8a8uTh0 ◴[] No.40685816{8}[source]
<< Remind me the restrictions around something like a Katana ?

The analogy kinda breaks, but the katana comparison is the interesting part[1] so lets explore it further. Most US states have their own regulations, but overall after you are 18 you are the boss with some restrictions imposed upon 'open carry'( for lack of a better term ). IL ( no surprise there ) and especially Chicago[2] ( even less of surprise ) has a lot of restrictions that are fairly close to silly.

If we tried to impose same type of restrictions on llms, we would need to start with age ( and from there, logically, person below 18 should not be using unlocked PC for fear of general potential for mischief ) and then, likely, prohibit use for unlocked cellphones that can run unapproved apps. It gets pretty messy. And that is assuming federal and not state regulation, which would vary greatly across US.

Is it a good idea?

'In the US, katanas fall under the same legal category as knives. From the age of 18, it is absolutely lawful to possess a katana in the US. However, ownership laws vary by state, but most states allowing you to own and display a katana in your home. Restrictions may apply on "carrying a katana" publicly.'

[1]https://katana.store/blogs/samurai-sword-buying-guide/are-ka... [2]https://codelibrary.amlegal.com/codes/chicago/latest/chicago...

61. A4ET8a8uTh0 ◴[] No.40685889{8}[source]
<< That is absolutely not the case - I just don't believe that a bunch of hot garbage people (in this case the racists and bigots who want to use LLMs to proliferate hate, people who create deep fakes of kids and celebrities) or a bunch of horny folks (ranging from people who want sexy time chat bots to, or just 'normal' generated erotic content) should be able to compel individuals or businesses to release the tools to do that.

I will admit that I actually gave you some initial credit, because, personally, I do believe there is some limited merit to the security argument. However, stating you can and should dictate how to use llms is something I can't support. This is precisely the one step away from tyranny, because it is the assholes that need protection and not saints.

But more to the point, why do you think you got the absolute right to limit people's ability to do what they think is interesting to them ( even if it includes things one would deem unsavory )?

<< You are concerned about freedom of expression, and I am concerned about freedom from compulsion (since I have already stated that I don't believe that losing access to LLMs breaks freedom of expression

How are you compelled? I don't believe someone using llms to generate horny chats compels you to do anything. I am open to an argument here, but it is a stretch.

62. sangnoir ◴[] No.40687023{5}[source]
Rebuking, shunning and ostracism are key levers for societal self-regulation, and social cohesion. Pick any society, at any point in time, amd you will find people/ideas that were rejected for not confirming enough.

There are limits to free speech even in friendship or families- there are things that even your closest friends can say that will make you not want to associate with them anymore.

replies(1): >>40688074 #
63. N0b8ez ◴[] No.40687726{6}[source]
It seems to cut both ways. If words are powerful, restricting words is also powerful. It's not clear why this leads to a pro-censorship stance, any more than to an anti-censorship one.
replies(1): >>40688780 #
64. tmcdos ◴[] No.40687994{6}[source]
It is quite obvious that the issue is inside the people - not inside the words. People have the ultimate power (a gift by God) to make decisions. Words can not force someone to do something - they are just sitting right there, doing nothing. Humans have flaws (probably by design - who knows) - and these flaws are the ones that all "safety" intentions MUST address. But 90% of humans prefer the easy path.
replies(1): >>40688764 #
65. matt-attack ◴[] No.40688049{10}[source]
Thank you. Very well put!

I don’t care what is considered illegal in certain jurisdictions. That’s off topic. Sodomy is illegal in certain jurisdictions. Are you going to try to convince me that I should give two shits about what two or three or four people choose to stick in what hole in the privacy of their homes? We’re taking about this insidious language of LLMs being “dangerous”.

If an LLM printed the text written by the GP about funding a hit, I fail to see how even that is “dangerous”.

I can write a bash script right now that prints that same thing, and I can post it to GitHub. Is anyone going to give two shits about it?

Someone has to explain how an LLM producing that same text is any different than my bash script printing to STDOUT. There’s not fucking difference. A program printed some text and there’s no argument behind the case that it’s dangerous.

replies(1): >>40689273 #
66. matt-attack ◴[] No.40688074{6}[source]
Well, the arguments out there aren’t that LLM’s are too brash, or discourteous or, insensitive. People are saying they’re “dangerous”. None of your examples speak to danger. No one is censored for being insensitive, or impolite or an opportune or discourteous. I totally support society regulating those things, and even outcastIng individuals who violate social norms. But that’s not what the anti-LLM language is framed as. It’s saying it’s “dangerous “. That’s a whole different ballgame, and I fail to see how such a description could ever apply. We need to stop that kind of language. It’s pure 1984 bullshit.
replies(2): >>40688994 #>>40692496 #
67. matt-attack ◴[] No.40688095{9}[source]
> who is being censored… The author of the program obviously.

If I write a bash script that echos “kill all the Jews”, and you choose to censor it, just who do you think is being censored? The intel professor? No! The author of the bash script obviously!

68. ben_w ◴[] No.40688764{7}[source]
Even with that attitude, the human flaws that make them act on those words, are known, and are exploitable and exploited.

If someone makes a device which is only safe when used safely, and they give it out to all despite being told of the risks, I think they are (or should be) liable for the misuse.

> a gift by God

I don't know which religion you follow. ᚦᛟᚱ᛬ᛟᚷ᛬ᛚᛟᚲᛁ᛬ᚺᛖᛁᛚᛊᚨ.

If you want a biblical reference, parable of the sower is just as valid when it's the word of satan.

replies(1): >>40738469 #
69. ben_w ◴[] No.40688780{7}[source]
Oh indeed. That's why dictators both censor and propagandise.

It's a narrow path, absolutely a challenge to walk without slipping, and not one I feel confident of humanity rising to even as a team effort.

Just like the difference between liberty and authoritarianism in general: much as I'd like to be an anarchist in theory, in practice that's just a way to let people with big sticks take over.

70. ben_w ◴[] No.40688994{7}[source]
> We need to stop that kind of language. It’s pure 1984 bullshit.

Sounds like you're saying, in this specific passage I'm quoting, "this language is dangerous and must be stopped".

Surveillance AI is already more invasive than any Panopticon that Orwell could imagine. LLMs and diffusion models make memory holes much easier. Even Word2Vec might be enough to help someone make a functional Newspeak conlang — though I wonder, is it better for me to suggest the (hopefully flawed) mechanism I've thought for how to do so in the hope it can be defended against, or would that simply be scooped up by the LLM crawlers and help some future Ingsoc?

71. bossyTeacher ◴[] No.40689216[source]
> I'd rather see disclaimers ("this may be wrong information" or "do not attempt") than my own computer (or the services I pay for) straight out refusing my request.

Are you saying that you want to pay to be provided with harmful text (see racist, sexist, homophobic, violent, all sorts of super terrible stuff)?

For you, it might be freedom for freedom sake but for 1% of the people out there, that will be lowering the barrier to commit bad stuff.

This is not the same as a super violent showing 3d limb dismemberments. It's a limitless, realistic, detailed and helpful guide to commit horrible stuff or describe horrible scenarios.

in4 you can google that, your google searches get monitored for this kind of stuff. Your convos with llms won't.

It's very disturbing to see adults people on here arguing against censorship of a public tool

replies(2): >>40714904 #>>40722902 #
72. A4ET8a8uTh0 ◴[] No.40689273{11}[source]
<< I don’t care what is considered illegal in certain jurisdictions.

I think this is where it gets messy. I care what happens in my jurisdiction, because this is where the laws I am subject to are enforced. The part that aggravates me is that the llms are purposefully neutered in stupid ways that are not even trying to enforce laws, but rather current weird zeitgeist that has somehow been deemed appropriate to be promoted by platforms.

<< A program printed some text and there’s no argument behind the case that it’s dangerous.

As I mentioned in my previous posts, I accept some level of argumentation from security standpoint ( I suppose those could be argued to be dangerous ), but touching touchy topics is not that.

At the end of the day, I will say that this censorship in itself is dangerous. Do you know why? When I was a little boy, I learned of censorship relatively late, because it was subtle ( overt restriction on what you could read and write typically indicated useful information and was sought after ). It didn't make censorship less insidious, but at least it didn't immediately radicalize a lot of people. This 'I am afraid I can't let you do that Dave' message I get from censored llm is that overt censorship that is already backfiring from that perspective.

<< Someone has to explain how an LLM producing that same text is any different than my bash script printing to STDOUT.

The only real difference is that it has more complex internals and therefore its outputs are more flexible than most programs. The end result is the same ('text on screen'), but how it gets there is different. Good bash script will give you the information needed as long as it is coded right; it is a purpose built tool. LLMs, OTOH, are a software equivalent of personal computer idea.

ok. i think i need coffee

73. bossyTeacher ◴[] No.40689278{3}[source]
Maybe, but defintely needs to be put on a watchlist. Otherwise, at some point, that deranged guy will actually enact his horrible fantasies and the families of the victims will demand to know why the guy wasn't confined when he was clearly having fantasies about this.

While not all people like him end up actually doing anything, you can't pretend those who do didn't fantasize before doing it. The difference is that now we can potentially have access to people's fantasies and act before it's too late

74. sangnoir ◴[] No.40692496{7}[source]
> Well, the arguments out there aren’t that LLM’s are too brash, or discourteous or, insensitive. People are saying they’re “dangerous”.

I didn't say that...

> None of your examples speak to danger.

Why should they have supported an argument I didn't make.

My comment is anti-anti-censorship of LLM. People already self-censor a lot; "reading the room" is huge part of being a functional member of society, and expecting LLMs to embody the "no-filter, inappropriate jerk" personality is what's against the grain - not the opposite.

I'm pragmatic enough to know the reason corporate LLMs "censor" is their inability to read the room, so they default to the lowest common factor and be inoffensive all the time (which has no brand risk), rather than allowing for the possibility the LLM offends $PROTECTED_CLASS, which can damage their brand or be legally perilous. That juice is not worth the squeeze just to make a vocal subset of nerd happy; all the better if those nerds fine-tune/abliterate public models so the corps can wash their hands of any responsibility of the modified versions.

75. autoexec ◴[] No.40714904[source]
> in4 you can google that, your google searches get monitored for this kind of stuff. Your convos with llms won't.

Not sure why you'd think that. Unless you run the ai locally and 100% offline you shouldn't expect any privacy at all

76. sattoshi ◴[] No.40722902[source]
> Are you saying that you want to pay to be provided with harmful text

This existence of “harmful text” is a bit silly, but lets not dwell on it.

The answer to your question is that I want to be able to generate whatever the technology is capable of. Imagine if Microsoft Word would throw an error if you tried to write something against modern dogmas.

If you wish to avoid seeing harmful text, I think that market is well-served today. I can’t imagine there not being at the very least a checkbox to enable output filtering for any ideas you think are harmful.

77. tmcdos ◴[] No.40738469{8}[source]
Well, I am not such a strict follower of a religion but I believe that if someone listens to Satan - the consequences are his/her own responsibility, not a Satan's guilt. If I am not mistaken - Satan makes offers. You can accept or pass over. If you accept - you are liable, not the other way around. I am not aware of anything in this world that is safe even when used in unsafe way. Hiding an information just because someone thinks it is "not safe" is a classic censorship. Once the words are censored - there is literally just one step to the censoring of thoughts.