top | item 3643045

(no title)

bwarp | 14 years ago

I was the proud owner of an ARM copro [1] many years ago (I still have the Master it was plugged into) and the first Acorn RISC machine (an A310 with 512k RAM if I remember correctly).

They were and still are extremely powerful and productive machines.

[1] http://en.wikipedia.org/wiki/BBC_Micro_expansion_unit#ARM_Ev...

discuss

order

gaius|14 years ago

Returning to your original post,

We really are building things with too deep abstraction hierarchies causing knowledge to be divided

Abstraction is necessary, true, but it's not clear to me what the abstraction-level we have now really gets us. In other words, say we had a BBC micro with a 2Ghz 6502 in it. What productive computing tasks that we do now could it not do? Or let's imagine an Atari ST with a 2Ghz 68000, to get us a little more memory. What could it not do, that we need to do now? I'm struggling to think of anything.

bwarp|14 years ago

It doesn't get us anything at all other than a fucking huge rabbit hole to stare down every time you do anything. Lets look at a pretty naff case for .Net CLR on x86 for windows workflow:

application -> xaml -> framework -> server -> container -> c# -> cil -> bytecode -> x86 instructions -> microcode.

Now forth on a 68000:

Forth screen -> 68k instructions

To be honest, for what I consider to be life and death stuff, a 10MHz 68000 is good enough (I have one in my TI92 calculator).

william42|14 years ago

The main thing is that you'd need a new set of abstractions for security, and then you'd need to implement HTML5 on it anyways to do all the things we can do on a computer now.