The mathematics of measuring how well our true objective (e.g. helpful or accurate responses) is being optimized when training an AI systems on a proxy metric (e.g. the output of a reward model). Pretty cool & not obvious that you can measure this at all!
openai.com/blog/measuring-go…
Software engineering: 50% understanding requirements, 40% complexity management, 9% debugging, 1% solving "interesting" algorithmic problems.
You'll enjoy software engineering a whole lot more if you instead think of the first 99% as the interesting part.
Big challenge in ML engineering: finding aggregate views that let you quickly understand the micro details of everything happening in your system. Often the biggest problems are simple problems obvious from looking at one specific computation. ML rewards attention to detail.
Reality is a program running on a universe-sized quantum computer. So the best way to solve a hard problem is to maximize contact with reality — lets you tap into an unfathomable amount of compute.
Whenever I cowork with my colleagues in person, I realize that the pandemic has completely dulled my sense of just how energizing & productive good collaboration can be. Months of remote work have resulted in many tiny unshared assumptions, which we only notice in person.
Get an onsite interview, get early access to DALL•E. Good way to show rather than tell what kinds of projects you could work on here.
(Attached photo: "A walrus typing at a keyboard doing a programming interview at an office".)
We just decided we'll give early access to DALL•E for any applicants to @OpenAI who make it to the onsite interview stage.
Should help candidates get a better sense for where the technology is and how to think about the opportunity.
DALL·E 2 — generate any image from a text description. Imagination is the limit.
"A Shiba Inu dog wearing a beret and black turtleneck"
"A photo of a quaint flower shop storefront with a pastel green and clean white facade and open door and big window"
openai.com/dall-e-2/
AI engineering is poring through logs puzzling out why your run crashed, profiling to increase perf by a more few percent, and figuring out which code version was running during that weird blip.
Only when the model is finally trained do you realize it was magic all along.
In a startup, you should seek out activities that seem hard, boring, annoying, and unscalable. The highest-value tasks are often hiding amongst them, and no one else has noticed because they seem unappealing on the surface.
I am inspired by curiosity.
That is what drives me.
So let us expand the scope & scale of consciousness so that we may aspire to understand the Universe.
.@Replit can now automatically find bugs in your code using OpenAI Codex under the hood.
A concrete step towards transforming programming to be more about expressing your intent, and less about getting the incantations exactly right: