top | item 46660705

(no title)

gregbot | 1 month ago

Great. So 538TWh per year is 61 GW so roughly 61 GW * $9.38 = $576 billion staggered over the 80 year life of nuclear plants is $7.2 billion per year of capital expenditure.

For comparison, wind is about $5/W. Assuming a 35% capacity factor and a 30 year expected lifetime for the latest turbines that comes to $10.0 billion per year of capital expenditure with no storage or fossil backup systems or extra capacity given weather variability.

https://en.wikipedia.org/wiki/Cost_of_electricity_by_source https://en.wikipedia.org/wiki/Electricity_sector_in_France

discuss

order

matthewdgreen|1 month ago

PV solar is between $.97 and $1.16 per watt, so that's going to be the front line. With storage you can get from $1.60 to $2. This is already the bulk of the power generation in Europe and is only going to increase. The idea that you're going to run nuclear plants at 95% capacity factor economically is also very suspect in a continent saturated with cheap PV solar.

gregbot|1 month ago

US NREL Puts it at $2/W with no storage and ~20% capacity factor. Lifetime of latest panels is unknown but optimistically is 25 years. Assuming perfect and free storage that comes to $24.4 billion per year of capital expenditure for a country the size of France to be 100% solar. So no, it would not be more economical to use solar over nuclear. Wind would be better but when you add the full system costs of storage and backup intermittent heavy systems are vastly more expensive and emit more carbon than nuclear ones. https://discussion.fool.com/t/levelized-full-system-costs-of...

Intermittents are only gaining market share because their unreliable and intermittent power which is less valuable is being purchased by governments at prices that far exceed what it is worth. In other words, massive hidden subsidies. Without those, there would be next to no intermittents on the grid anywhere.

See “Market matching costs” here: https://en.wikipedia.org/wiki/Cost_of_electricity_by_source