Ask HN: Security in the Age of AI
This is an interesting subject that almost feels like a Pandora's box to me.
How does one actually go about securing AI models be it unimodels or multi-modal systems?
Is there a specific framework of thinking you'd recommend? Never giving them full autonomy over critical infrastructure, whether it is mono or multi. These models are great at speeding things up but without human checks over their actions, over time errors will be exacerbated. Add into the fact that there is little verification to the interior logic, allowing anything to be completely controlled by an AI is asking for trouble. While this might work in the short term over time the models will get better and humans lazier. Really we need a cultural shift in business and automation tasks. One that would make the people using the models to assess their actions using it and one that breeds complete distrust of the AIs. If no one gives the keys of power/destruction to the models there's little to no problems, but currently businesses and the workers within them view human action as both more incorrect and unreliable than ones controlled by computers. So inevitably they will be given more power to offload the tasks to an autonomous agent with no checks in place. Even if we accomplish the cultural and practical changes very generally outlined before, people will point to the fear of an AI "waking up" and then taking the keys to power itself. While there are many steps needed to get to that point, if we take the outlined precautions there is still room for that to happen. To tackle that problem, I personally believe that we should focus on narrow, specialized AIs that are completely disconnected from one another. In this way we don't centralize enough computing power to run a "Super Intelligence". The more obfuscation between a centralized AI that could "trick" people through psychological attack vectors will make it significantly more safe. So in conclusion, don't give them nukes, change the way people view work and automation, and keep them small and far apart from one another. sound points! do you have any oss projects that I can look at to build guardrails for my ai models? Unless you are a large scale company, luckily there isn't much responsibility on your part. Really the danger vectors I identified are more geared to entities with massive resources and also have the power to dictate public discourse. But as we get more efficient models and hardware, there might be dangers the small scale researcher could let loose. Cant wait! In all honesty, the first step is to educate yourself on the actual mechanisms behind today's models, and realize they are not super intelligent Gods and selling them like it will only hurt the public and industry. Aim for specially tailored models that have an off switch and do not directly "talk" to critical hardware. Don't copy ai generated code wholesale, since downstream these could be exploited by a hacker. Don't make hacking AIs like thats just asking for pain. The biggest one is think about the downstream effects your final product could have on both the positive and the negative. Today everyone focuses on the positive while ignoring the possible dooms. Dont be blinded by the possible future, see what could go wrong and plan for it so it doesnt become a problem. Also thank you for the kind words. :D