top | item 37670980

(no title)

jacknobody | 2 years ago

I think it's a matter of convention, and that all the maths I can think of survives the following definitions:

If infinity were to be defined as immeasurably large, and zero as immeasurably small, we would be saying infinity=(some finite number)/0 and 0=(some finite number/infinity

This still leaves us unable to find an definite answer to the multiplication of 0 and infinity, because we might have chosen any 2 finite numbers for the definitions above.

discuss

order

anankaie|2 years ago

The problem is that the "infinitely small" is already defined, and not equal to zero: The infinitesimal. It is not so simple as defining what the result should be - one needs to make sure this is consistent with the rest of mathematics, or rather, with the mathematics generated by the rest of your axioms and definitions, without resulting in paradox.

If anything, division by zero generates the set of all numbers in your field. That is consistent with division being the operation of finding the multiplicative inverse, but inconsistent with the idea that arithmetic operations have only a single result, rather than a set - but this already exists in finding k-roots of numbers, so it is not completely unreasonable to treat some arithmetic operations as potentially resulting in sets.

Is it a useful definition? Unclear.