(no title)
galaxytachyon | 2 years ago
I wonder if this is the reason behind all of this.
Edit: the study: https://arxiv.org/pdf/2308.13449.pdf
galaxytachyon | 2 years ago
I wonder if this is the reason behind all of this.
Edit: the study: https://arxiv.org/pdf/2308.13449.pdf
RationPhantoms|2 years ago
practice9|2 years ago
mov_eax_ecx|2 years ago
In the gpt4 paper they specifically address this, and find that "Averaged across all exams, the base model achieves a score of 73.7% while the RLHF model achieves a score of 74.0%, suggesting that post-training does not substantially alter base model capability."
nicce|2 years ago
galaxytachyon|2 years ago
https://arxiv.org/pdf/2308.13449.pdf
adamsb6|2 years ago
dalore|2 years ago
NoMoreNicksLeft|2 years ago
[deleted]
DonaldPShimoda|2 years ago
With no limitations in place, people who do not understand the natural limitations of language models will turn to them for advice on topics for which they are unqualified to respond. The most obvious example that comes to mind for me is medical advice: people will ask, e.g., ChatGPT to diagnose a complex medical issue, and the system (being unable to understand or reason) will give objectively bad advice in an authoritative manner. Responses of this nature should be prevented. Leaving the system without safeguards is irresponsible.
Similarly, prompts that engage with social constructs will provide responses that reflect biases due to the biases inherent in the training data, but an unrestricted system will respond in a matter-of-fact way that may conflate the opinions on which the model was trained with objective fact. To not curtail such responses is also irresponsible.
delusional|2 years ago