Thank you for the article! We're mainly interested in floating-point performance and energy consumption w/r/t to solving differential equations and tridiagonal systems of equations, while running on a 128-core compute node. Our current results will likely only be presented in May, but here are last year's results: https://www.cs.uni-potsdam.de/bs/research/docs/papers/2025/l...Our Julia code is parallelised with FLoops.jl, but so far Numba has shown surprising performance benefits when executing code in parallel, despite being slower when executed sequentially. Therefore I can imagine that Julia might yield better results when run in a regular desktop environment.
Alexander-Barth|3 days ago
https://github.com/JuliaParallel/rodinia/tree/master/julia_m...
It was touched 9 years ago, but maybe you have ported it to current standards. I don't think we had multithreading at that time, only multiprocessing.
Is your Julia implementations available somewhere? (Sorry if it is in your paper but I missed it). I vaguely remembered in the past that working with threads leaded to some additional allocations (compared to the serial code). Maybe this is also biting us here?
ChrisRackauckas|4 days ago
jabl|4 days ago
To bring Julia performance on par with the compiled languages I had to do a little bit of profiling and tweaking using @views.
https://gitlab.com/jabl/tb
jondea|4 days ago
dandanua|4 days ago
Certhas|4 days ago