OP added "ATI" by themselves rather than keeping original one. What a shame.
Anyway:
> t provides an additional six displays up to 4K with the Intel HD graphics that has three, giving a total of nine outputs. The Radeon Graphics supports DisplayPort 1.4 with HDR and HDMI 2.0b with HDR10 support, along with FreeSync/FreeSync2. As a result, when the graphics output changes from Intel HD Graphics to Radeon graphics, users will have access to FreeSync, as well as enough displays to shake a stick at (if the device has all the outputs).
Yes, if those NOCs/HTPCs provide all of those capabilities, otherwise it's just marketing words. In reality, only top notch has more than 1 DisplayPort, I guess.
The already-announced Hades Canyon NUC has 2x DisplayPort, 2x HDMI 2.0a and 2x Thunderbolt 3. That amounts to six ports capable of driving a 4K display. Thunderbolt 3 ports can drive two 4K displays at 60Hz if the controller supports it.
I suspect that other products using this processor will be less generous in terms of connectivity, simply because it's total overkill for the overwhelming majority of users.
I could imagine an OEM like Apple or Dell looking at these to free up space in their board layout on their notebooks or small form factor computers
In such a scenario ECC memory is not really a high priority, no?
Looking at the PCI-Express lanes available those other 8 CPU lanes look ripe for a thunderbolt 3 controller, with the rest of the peripherals being powered from the PCH (or does that make no sense at all?)
Not sure why Intel just doesn't release a full system in package for a basic laptop system, with 16GB or 32GB main memory and 256GB or 512GB of NVMe flash?
I think Intel still makes flash chips and they originally started out as a DRAM manufacturer as well..
Because that would mean discarding an entire package if just one component failed or is damaged during production.
It would also mean that manufacturers would have to order different parts for two systems that are identical, expect from memory or storage. In your example, with 16 or 32GB of RAM and 256 or 512GB of storage, you'd end up having to order four different SKUs from Intel, and Intel would have to manufacture four different SKUs, plus one without memory and storage.
Logistically I don't think it makes sense to add storage and memory to the package, it just adds inflexibility and more SKUs.
They make something similar. An Intel NUC is basically a full system laptop, with the laptop-specific bits missing (no screen, keyboard, battery, etc).
You can buy it barebones, but you can also buy some of them with everything pre-installed (including RAM and flash storage and licensed Windows OS)
Why are AMD helping to sell a competitor's product? The article says it's strictly business, but I would have thought the quick buck made today by selling graphics chips to Intel would be outweighed by the long term benefit (e.g. growth in the combined CPU/GPU market) this provides to Intel.
This is a joint venture against Nvidia basically. ML is the next big thing, and Nvidia has smartly positioned GPUs as the best way to do ML. AMD needs to claw back marketshare from Nvidia, and partnering with Intel is a quick way to do it. Intel also needs to keep Nvidia in check.
Now they should co-develop and promote a well designed open source ML framework, something that can compete with CUDA. AMD isn't up to the task, but Intel is.
In the UK that sort of merger would almost certainly be blocked by the Competition Commission: I'd be interested to hear what the situation is in the US, where Intel and AMD are based.
Yeah, because all they have to do is just tweak some code and re-release... /s It takes 2 - 4 years to design a chip. If Intel doesn't release the new chip they started designing 2 - 4 years ago, they waste that development cost, fall behind AMD, and you complain that they are behind the times. If they do release the chip you complain they haven't fixed the bug you just learned about a week ago. It took a year for everyone to fix their software, and you want Intel to magically come out with new hardware? Sheesh.
I think all the "Intel sucks" comments should have a disclaimer "Full disclosure: I've never designed hardware in my life, I have no idea how to run a business, I really just hate Intel and this is an excuse to vent my hatred."
[+] [-] a012|8 years ago|reply
Anyway:
> t provides an additional six displays up to 4K with the Intel HD graphics that has three, giving a total of nine outputs. The Radeon Graphics supports DisplayPort 1.4 with HDR and HDMI 2.0b with HDR10 support, along with FreeSync/FreeSync2. As a result, when the graphics output changes from Intel HD Graphics to Radeon graphics, users will have access to FreeSync, as well as enough displays to shake a stick at (if the device has all the outputs).
Yes, if those NOCs/HTPCs provide all of those capabilities, otherwise it's just marketing words. In reality, only top notch has more than 1 DisplayPort, I guess.
[+] [-] jdietrich|8 years ago|reply
I suspect that other products using this processor will be less generous in terms of connectivity, simply because it's total overkill for the overwhelming majority of users.
https://www.anandtech.com/show/12226/intels-hades-canyon-nuc...
[+] [-] prewett|8 years ago|reply
However, I read the article hoping to figure out why Intel is doing this, and no luck there. Are they giving up on the Intel graphics?
[+] [-] mrmondo|8 years ago|reply
[+] [-] rwx------|8 years ago|reply
[+] [-] dna_polymerase|8 years ago|reply
[+] [-] Tsiklon|8 years ago|reply
In such a scenario ECC memory is not really a high priority, no?
Looking at the PCI-Express lanes available those other 8 CPU lanes look ripe for a thunderbolt 3 controller, with the rest of the peripherals being powered from the PCH (or does that make no sense at all?)
[+] [-] mozumder|8 years ago|reply
I think Intel still makes flash chips and they originally started out as a DRAM manufacturer as well..
[+] [-] stonogo|8 years ago|reply
You can ask them, I guess.
[+] [-] mrweasel|8 years ago|reply
It would also mean that manufacturers would have to order different parts for two systems that are identical, expect from memory or storage. In your example, with 16 or 32GB of RAM and 256 or 512GB of storage, you'd end up having to order four different SKUs from Intel, and Intel would have to manufacture four different SKUs, plus one without memory and storage.
Logistically I don't think it makes sense to add storage and memory to the package, it just adds inflexibility and more SKUs.
[+] [-] maxsilver|8 years ago|reply
You can buy it barebones, but you can also buy some of them with everything pre-installed (including RAM and flash storage and licensed Windows OS)
[+] [-] chx|8 years ago|reply
[+] [-] pjc50|8 years ago|reply
[+] [-] raverbashing|8 years ago|reply
[+] [-] sleavey|8 years ago|reply
[+] [-] frou_dh|8 years ago|reply
So Intel goes cap-in-hand to AMD for their tech.
[+] [-] fgonzag|8 years ago|reply
Now they should co-develop and promote a well designed open source ML framework, something that can compete with CUDA. AMD isn't up to the task, but Intel is.
[+] [-] Shivetya|8 years ago|reply
[+] [-] PascLeRasc|8 years ago|reply
[+] [-] msh|8 years ago|reply
[+] [-] justinclift|8 years ago|reply
[+] [-] jacksonsabey|8 years ago|reply
[+] [-] sp332|8 years ago|reply
[+] [-] make3|8 years ago|reply
[+] [-] bartread|8 years ago|reply
[+] [-] Shivetya|8 years ago|reply
[+] [-] leandrod|8 years ago|reply
[+] [-] gsich|8 years ago|reply
[+] [-] capisce|8 years ago|reply
[+] [-] mrmondo|8 years ago|reply
[+] [-] pricetag|8 years ago|reply
[+] [-] itsnotvalid|8 years ago|reply
[+] [-] aceoflulu|8 years ago|reply
[deleted]
[+] [-] smallnamespace|8 years ago|reply
Also who negotiated the deal to let Intel slap only its logo on the packaging, even though the AMD die is clearly larger?
[+] [-] TazeTSchnitzel|8 years ago|reply
[+] [-] adrianratnapala|8 years ago|reply
Well, take a wild guess at which way the money is going to flow.
[+] [-] dna_polymerase|8 years ago|reply
I don't know how they have the audacity to release new CPUs while all their products just went to shit...
[+] [-] prewett|8 years ago|reply
I think all the "Intel sucks" comments should have a disclaimer "Full disclosure: I've never designed hardware in my life, I have no idea how to run a business, I really just hate Intel and this is an excuse to vent my hatred."
[+] [-] t3rmi|8 years ago|reply
[+] [-] pjmlp|8 years ago|reply