(no title)
danpalmer | 5 days ago
Gemini: https://ai.google.dev/gemini-api/docs/code-execution
ChatGPT: https://help.openai.com/en/articles/8437071-data-analysis-wi...
Claude: https://claude.com/blog/analysis-tool
Reasoning only gets you so far, even humans write code or use spreadsheets, calculators, etc, to get their answers to problems.
simianwords|5 days ago
there are multiple ways to disprove this
1. GPT o1 was released and it never supported the tools and it easily solved the strawberry problem - it was named strawberry internally
2. you can run GPT 5.2-thinking in the API right now and deny access to any tools, it will still work
3. you can run deepseek locally without tools and run it, it will still work
Overall this idea that LLM's cant reason and need tools to do that is misleading and false and easily disproven.
danpalmer|5 days ago
My point was much more general, that code execution is a key part of these models ability to perform maths, analysis, and provide precise answers. It's not the only way, but a key way that's very efficient compared to more inference for CoT.