top | item 46235033

(no title)

hackernewds | 2 months ago

> Writing code is the default behavior from pre-training

what does this even mean? could you expand on it

discuss

order

bongodongobob|2 months ago

He means that it is heavily biased to write code, not remove, condense, refactor, etc. It wants to generate more stuff, not less.

elzbardico|2 months ago

Because there are not a lot of high quality examples of code edition on the training corpora other than maybe version control diffs.

Because editing/removing code requires that the model output tokens for tools calls to be intercepted by the coding agent.

Responses like the example below are not emergent behavior, they REQUIRE fine-tuning. Period.

  I need to fix this null pointer issue in the auth module.
  <|tool_call|>
  {"id": "call_abc123", "type": "function", "function": {"name": "edit_file",     "arguments": "{"path": "src/auth.py", "start_line": 12, "end_line": 14, "replacement": "def authenticate(user):\n    if user is None:\n        return   False\n    return verify(user.token)"}"}}
  <|end_tool_call|>

snet0|2 months ago

I don't see why this would be the case.

joaogui1|2 months ago

During pre-training the model is learning next-token prediction, which is naturally additive. Even if you added DEL as a token it would still be quite hard to change the data so that it can be used in a mext-token prediction task Hope that helps