(no title)
extheat | 1 year ago
What would be more interesting is training a large model on pure (code, assembly) pairs like a normal translation task. Presumably a very generalized model would be good at even doing the inverse: given some assembly, write code that will produce the given assembly. Unlike human language there is a finite set of possible correct answers here and you have the convenience of being able to generate synthetic data for cheap. I think optimizations would arise as a natural side effect this way: if there's multiple trees of possible generations (like choosing between logits in an LLM) you could try different branches to see what's smaller in terms of byte code or faster in terms of execution.
hughleat|1 year ago
> What would be more interesting is training a large model on pure (code, assembly) pairs like a normal translation task.
It is that.
> Presumably a very generalized model would be good at even doing the inverse: given some assembly, write code that will produce the given assembly.
Is has been trained to disassemble. It is much, much better than other models at that.
quonn|1 year ago
ChatGPT does this, unreliably.