Emailing users when a new feature is enabled for their account isn't even the kind of thing that would distract an existing very busy developer.
You could literally hire an entirely new guy, give him instructions to build such an email system, and let him put the right triggers on the user account permissions database to send out the right emails at the right time.
And then, when it's built, you can start adding more features like sending the emails only when demand is low and/or at times of day when you get the best click through rate. And then next you can start measuring the increase in revenue from sending those emails.
Before long, you have a whole marketing and comms team. Which you probably want as a big company anyway.
The peanut gallery could (and does) say this about 1000 little features that 1000 different people claim to be so necessary that OpenAI is incompetent for not having, yet none of those people agree that those other features are priorities.
But this issue far predates their current success. GPT2 was held back for a while. GPT3 launched as "waitlist only" with an application process, and so did GPT3.5.
This is a large part of what held them back: GPT3.5 had most of the capabilities of the initial ChatGPT release, just with a different interface. Yet GPT3.5 failed to get any hype because the rollout was glacial. They made some claims that it was great, but to verify this for yourself you had to wait months. Only when they finally made a product that everyone could try out at the same time, with minimal hassle, did OpenAI turn from a "niche research company" to the fastest growing start-up. And this seems to have been a one-time thing, now they are back to staggered releases.
> GPT3.5 had most of the capabilities of the initial ChatGPT release, just with a different interface
I believe two other factors were the cost (especially of fine tuned models, IIRC fine tuned davinci cost $0.20 per thousand tokens) and also that OpenAI hadn't very clearly shown just how much higher the quality could get once RLHF'd. I remember asking davinci-instruct to parse some data, and the reliability really seemed much lower than ChatGPT at launch, to the point that, at the time, I thought GPT-4 was secretly powering ChatGPT.
At what point to you go from startup to not when you have 10 billion invested and countless employees and is practically a sub branch of microsoft. Sounds cooler though I guess
I'm not sure when a company begins to not be a start up, but by the time they have a wave of news claiming their product kills teenagers, or they're engaging in antitrust, or they're effectively using space labor, that's when they are definitely no longer a start up.
That hasn't happened yet for OpenAI, but I'm sure it will happen eventually, and then we'll know.
I think you stop being a startup when there are engineers who do not know the CEO. I would guess OpenAI is still a startup by that definition (they don't have that many engineers IIRC) but I don't actually know.
That's really a function of what kind of CEO the company has, and what do you mean by "know". I worked remotely for a company employing hundreds of people, around for couple decades and with offices in different regions of the world, and I still got to talk to the CEO a couple times, and he knows me by name, all by virtue of bumping into him a couple times on corridor while on one of my infrequent visits to the office.
> I suspect they do care about communicating with customers, but it's total chaos and carnage internally.
This is my best guess as well, they are rocketing down the interstate at 200mph and just trying to keep the wheels on the car. When you're absolutely killing it I guess making X% more by being better at messaging just isn't worth it since to do that you'd have to take someone off something potentially more critical. Still makes me a little sad though.
Whether or not you’re profitable has very little to do with how valuable others think you are. And usually having competitors is something that validates your market.
Clearly, you can be a company like Microsoft where nobody is challenging your dominance in PC operating systems, and you can make huge sums of money. So competitors certainly aren't vital.
Or if you've cleverly sewn up a market with patents or trade secrets or a giant first-mover advantage or network effects, and nobody's got any chance of challenging your dominance in your specific market niche - again that could be highly profitable.
On the other hand, if you're selling ten-dollar bills for five dollars, you might have millions of satisfied paying customers, but no competitors because nobody else thinks your unit economics make sense. Or if you run a DVD rental store, you might be profitable and have many return customers, but you might not attract competitors because they don't think DVD rental is a growth business.
So some people consider a lack of competition an ominous sign.
> And usually having competitors is something that validates your market
a whole bunch of AI startups were founded around the same time. surely each can't validate the market for the others and be validated by the others in turn
The surviving ones can. The same way that multiple species of trees, growing on the same patch of the ground, desperately competing for sunlight, together validate that the soil underneath is fertile.
When dealing with a tech where people have credible reasons to believe it can be enormously harmful on every possible time scale, maybe it would behoove them to not rocket down the interstate at 200mph?
There is always going to be people who against any new technology and who makes up reasons to be against it.
The best defence is to move so quickly that you are an established part of the business framework by the time these forces can gather, or to go so slowly that nobody takes you as a threat.
No, successfully navigate past this version of Covid vacine deniers, 5g conspiracists etc.
In ten years we will enjoy a higher productivity due to AI and a richer society as a result. We have already seen it with protein folding which AI is amazing at[0].
The only reasonable fear of AI is for some jobs and that China gets their first.
Right, it is perfectly valid to only accept the potential good points while neglecting all the potential bad points.
This is no different then saying "Look, nuclear weapons aren't actually dangerous, if they were we'd all be dead because the apocalypse would have already happened", which is probably the dumbest take on the close calls and real risks that exist.
I think their main goal is to be perceived as the most advanced AI company. Why? Because that's how you get the best people working for you. The main determinant of success for companies like OpenAI is people.
Considering the field and progress that is being made I find this idea terrifying. All the big problems like "How will we actually control what we're building?" being answered "that's too hard; let's punt and solve that after we figure out how to consume voice data". One way or another this is likely the last technological advance that humans will make.
> How will we actually control what we're building?
This question is poorly formed because it’s not clear who the “we” is. If it’s you and me, that train left the station a while ago. If it’s any humans, well Sam Altman is probably a human and all of these are impressive products, but still just tools.
I don't know if @skeeter2020's assertion is correct, but it is certainly the goal.
To use a fictional but entirely apt quote:
> I say your civilization because as soon as we started thinking for you, it really became our civilization, which is, of course, what this is all about: Evolution, Morpheus, evolution. Like the dinosaur. Look out that window. You had your time. The future is our world, Morpheus. The future is our time.
As long as it's better at some of the advances that would make it even better at those advances, or better at more advances, then it'll quickly become better than us in approximately everything, and at that point humans become NPCs of their own story.
Are either of those things indicative of "fastest growth ever"? Maybe 100 million users, but we live in a world where it's hard to know how meaningful that is (ie Tucker's X videos getting 300 million views).
This happens when there are thousands other simple things and a lot of complicated things. When your devs are stretched, you sort by priority and I can tell you this is not that important.
I suspect they do care about communicating with customers, but it's total chaos and carnage internally.