This report identifies vulnerabilities in GPT-4, o1, and o3 models that allow disallowed content generation, revealing weaknesses in current alignment mechanisms.
Am I the only one that feels it’s a bit strange to have such safeguards in an AI model? I know most models aren’t available online but some models are available to download and run locally right? So what prevents me from just doing that if I wanted to get around the safeguards? I guess maybe they’re just doing it so that they can’t be somehow held legally responsible for anything the AI model might say?
Am I the only one that feels it’s a bit strange to have such safeguards in an AI model? I know most models aren’t available online but some models are available to download and run locally right? So what prevents me from just doing that if I wanted to get around the safeguards? I guess maybe they’re just doing it so that they can’t be somehow held legally responsible for anything the AI model might say?
The idea is they’re marketable worker replacements
If you have a call center you want to switch to ai, it’s easy though to make them pull up relevant info. It’s harder to stop them from being misused
If your call center gets slammed for using racial slurs, that’s an issue
Remember, they’re trying to sell AI as drop in worker replacement