How do you view your role in public ethics as a developer?
I figured out how to remove most of the safeguards from some AI models. I don't feel comfortable sharing that information with anyone. I have come across a few layers of obfuscation to make this type of alteration more difficult to find and sort out. This caused me to realize, a lot of you are likely faced with similar dilemmas of responsibility, gatekeeping, and manipulating others for ethical reasons. How do you feel about this?
Idk, I still think information wants to be free. If you figured it out just farting around, sophisticated malevolent actors are likely already doing similar things. Might be better to let the genie out of the bottle, so people can learn to be skeptical. Deep fakes are optimally effective when a majority still accepts the veracity of images as an article of faith.
The political and adult doesn't bother me. The kinds of things I might not have the ethics to think through at a much younger age, that bothers me, and I have never been a very deviant type. I think the protections against age are primarily for this situation. Training a LoRA takes 5 minutes now. An advanced IP adaptors and control net is just a few examples away and a day top for the slightly above average teen figure out. Normalizing this would have some very serious edge case consequences. It is best to leave that barrier to entry filter in place IMO. I assume it is still there because everyone that knows about it feels much the same. It does not show up in a search engine, although that is saying less than nothing these days.
I figured out how to remove most of the safeguards from some AI models.
Nice.
How do you feel about this?
It's another kind of power. I try to use mine responsibly, but also to give myself a break when I don't meet my own standards.
Some good advice I got once was that it's impossible to "un-say" something, so it pays to think twice before speaking.
If your gut is telling you to pause, listen to it. Wait to move forward until you feel better about it.
As someone else pointed out, responsible disclosure is an option.
You also have the option to just quietly enjoy a better copy of the AI than others have.
If you decide to publish your discoveries, be aware that others will judge you for how you go about it. For me that means the two options are responsibly, or anonymously.
Someone else will eventually figure it out. They probably have less scruples and will therefore profit.
Seems to me like there's always an incentive structure for prisoner's dilemma type shit to eventually pay off for the authoritarians in the end. You can play the game, but you can't break it or stop or from being rigged without consequences. Even just releasing some research papers will get you a few decades in the fed.
Oof, programmers calling LLMs "AI" - that's embarrassing. Glorified text generators don't need ethics, what's the risk? Making the Internet's worst texts available? Who cares.
I'm from an era when the Anarchists Cook Book, and The Unabombers Manifesto were both widely available - and I'm betting they still are.
There's no obligation to protect people from "dangerous text" - there might be an obligation to allow people access to them though.
...but LLMs quite literally come from the field of computer science that is referred to as "AI." What are they supposed to call it? I'm not a fan of the technology either, but seems like you're just projecting your disdain for ChatGPT.
"What am I supposed to call LLMs if not calling them AIs?"
...really dude? They're large language models, not artificial intelligences. So that's what you call them. Because that's what they are.
The fact that they came from research into artificial intelligence doesn't factor in. Microwave ovens came from radar research, doesn't mean we call them radars, does it?
Yeah. This is what I mean. I just figured out the settings that have been hard coded. There are keywords that were spammed into the many comments within the code, I assume this was done to obfuscate the few variables that need to be changed. There are also instances of compound variable names that, if changed in a similar way, will break everything, and a few places where the same variables have a local context that will likewise break the code.
I'm certainly not smart enough to get much deeper than this. The ethical issue is due to diffusion.
I've been off-and-on trying to track down why an LLM went from an excellent creative writing partner to terrible but had trouble finding an entry point. I just happened to stumble upon such an entry point in a verbose log entry while sorting out a new Comfy model and that proved to be the key I needed to get into the weeds.
The question here, is more about the ethics of putting such filtering in place and obfuscating how to disable it in the first place. When this filtering is removed, the results are night and day, but with large potential consequences.