Alignment really just means how close do the model outputs align with human preferences or some other criteria.
At a glance, this looks like a model pretrained to perform prompt-engineering. It should automatically use Chain-of-Thought in its responses in order to improve it's programming abilities, and, therefore be better aligned with users expectations.
It also has reflection. So they include code to execute the model output and return the response to the model for feedback.
exe34|1 year ago
irthomasthomas|1 year ago
At a glance, this looks like a model pretrained to perform prompt-engineering. It should automatically use Chain-of-Thought in its responses in order to improve it's programming abilities, and, therefore be better aligned with users expectations.
It also has reflection. So they include code to execute the model output and return the response to the model for feedback.
PoignardAzur|1 year ago