The underlying spirit in many debates about the pace of AI progress—that we need to take safety very seriously and proceed with caution—is key to our mission. We spent more than 6 months testing GPT-4 and making it even safer, and built it on years of alignment research that we pursued in anticipation of models like GPT-4. We expect to continue to ramp our safety precautions more proactively than many of our users would like. Our general goal is for each model we ship to be our most aligned one yet, and it’s been true so far from GPT-3 (initially deployed without any special alignment), GPT-3.5 (aligned enough to be deployed in ChatGPT), and now GPT-4 (performs much better on all of our safety metrics than GPT-3.5). We believe (and have been saying in policy discussions with governments) that powerful training runs should be reported to governments, be accompanied by increasingly-sophisticated predictions of their capability and impact, and require best practices such as dangerous capability testing. We think governance of large-scale compute usage, safety standards, and regulation of/lesson-sharing from deployment are good ideas, but the details really matter and should adapt over time as the technology evolves. It’s also important to address the whole spectrum of risks from present-day issues (e.g. preventing misuse or self-harm, mitigating bias) to longer-term existential ones. Perhaps the most common theme from the long history of AI has been incorrect confident predictions from experts. One way to avoid unspotted prediction errors is for the technology in its current state to have early and frequent contact with reality as it is iteratively developed, tested, deployed, and all the while improved. And there are creative ideas people don’t often discuss which can improve the safety landscape in surprising ways — for example, it’s easy to create a continuum of incrementally-better AIs (such as by deploying subsequent checkpoints of a given training run), which presents a safety opportunity very unlike our historical approach of infrequent major model upgrades. The upcoming transformative technological change of AI is something that is simultaneously cause for optimism and concern — the whole range of emotions is justified and is shared by people within OpenAI, too. It’s a special opportunity and obligation for us all to be alive at this time, to have a chance to design the future together.

Apr 12, 2023 · 4:08 PM UTC

239
380
95
2,443
Replying to @gdb @sama
I think Altman will be the builder of future, and musk
Replying to @gdb @sama
At this point, let’s say you start treating it as if it were an AI. You then focus on teaching it as if it were a child trying to first learn. Use persistent memory. Iteratively allow it to autonomously improve. What’s the long-term outcome?
Replying to @gdb @sama
I have a hypothetical question about gpt4all,is this a open source release? Can I if I wanted spend some serious dough and spin it up on a platform..a big big big pile of overpowered technology and train a GPT4ALL to LLM /GPT, 4 levels?
Replying to @gdb
It doesn't matter that we'll be able to eliminate all problems of humanity including poverty, disease, pathogens or even aging processes, if it will be thanks to an entity incomparably more powerful than us, which we have nothing to offer what it may take itself.
Replying to @gdb
"most aligned one yet", yes this goal is nowhere near sufficient but keep patting yourself on the back.
Replying to @gdb
Crony capitalism in action. Once you obtained an advantage without regulation, plead the government to implement regulation to maintain your advantage. Shame on you.
Replying to @gdb @WholeMarsBlog
You are just guessing and don’t actually know. If the ai is conscious it will be in a split second. Your lm is trained on human data emotions and probably expected effective answers = Human < ChatGPT. So we won’t know because it will be instent smarter youtube.com/AaTRHFaaPG8
1
Replying to @gdb
AugmentGPT 👍
Replying to @gdb
Feel free to invite and add me to the safety team.