(no title)
cjrd | 3 years ago
> Given both the competitive landscape and the safety implications of large-scale models like GPT-4, this report contains no further details about the architecture (including model size), hardware, training compute, dataset construction, training method, or similar.
- OpenAI
shpx|3 years ago
awesomeMilou|3 years ago
xvector|3 years ago
SXX|3 years ago
jryan49|3 years ago
dx034|3 years ago
Not defending their actions, but it's not that common that new very valuable products are directly available for retail users to use.
toriningen|3 years ago
They cannot disclose anything, since it would make it apparent that GPT-4 cannot have a number of parameters that low, or the gradients would have faded out on the network that deep, and so on.
They don't want any competition, obviously, but with their recent write-up on "mitigating disinformation risks", where they propose to ban non-governmental consumers from having GPUs at all (as if regular Joe could just run 100'000 A100s in his garage), so perhaps this means the lowest border for inference and training is a lot lower than we have thought and assumed?
Just a wild guess...