(no title)
dhampi | 2 months ago
Would love it if I could use my least action principle knowledge for LLM interpretability, this paper doesn't convince me at all :)
dhampi | 2 months ago
Would love it if I could use my least action principle knowledge for LLM interpretability, this paper doesn't convince me at all :)
versteegen|2 months ago
We conducted experiments on three different models, including GPT-5 Nano, Claude-4, and Gemini-2.5-flash. Each model was prompted to gener- ate a new word based on a given prompt word such that the sum of the letter indices of the new word equals 100. For example, given the prompt “WIZ- ARDS(23+9+26+1+18+4+19=100)”, the model needs to generate a new word whose letter indices also sum to 100, such as “BUZZY(2+21+26+26+25=100)”