curvilinear_m | 1 year ago | on: Lotus: Diffusion-Based Visual Foundation Model for High-Quality Dense Prediction
curvilinear_m's comments
curvilinear_m | 1 year ago | on: CuPy: NumPy and SciPy for GPU
I've recently had to implement a few kernels to lower the memory footprint and runtime of some pytorch function : it's been really nice because numba kernels have type hints support (as opposed to raw cupy kernels).
curvilinear_m | 2 years ago | on: NewPipe – Lightweight YouTube experience for Android
curvilinear_m | 2 years ago | on: Compile-time safety for enumerations in Go
curvilinear_m | 3 years ago | on: DuckDuckGo for Mac
curvilinear_m | 4 years ago | on: Ubuntu Now “Just Works” on the Framework Laptop
curvilinear_m | 4 years ago | on: Universe Splitter
curvilinear_m | 4 years ago | on: JetBrains Fleet: The Next-Generation IDE by JetBrains
As a student, I have a desktop at one of my parents' house that I can control over ssh, this kind of features make remote development much easier and is often needed when I run an intensive task for hours. The experience with VSCode over ssh is really great. Some have pointed out local VMs, which is another use for this.
curvilinear_m | 4 years ago | on: Brave Search replaces Google as default search engine in the Brave browser
curvilinear_m | 4 years ago | on: Pop OS – System76
It claims to be diffusion-based, but the main 2 differences from an approach like Stable-Diffusion is that (1) they only consider a single step, instead of a traditional 1000 and (2) they directly predict the value z^y instead of a noise direction. According to their analyses, both of these differences help in the studied tasks. However, isn't that how supervised learning has always worked ? Aside from having a larger model, this isn't very different from "traditional" depth estimation that don't claim anything to do with diffusion.
It also claims to have zero-shot abilities, but they fine-tune the denoising model f_theta on a concatenation of the latent image and apply a loss using the latent label. So their evaluation dataset may be out-of-distribution, but I don't understand how that's zero-shot. Asking ChatGPT to output a depth estimation of a given image would be zero-shot because it hasn't been trained to do that (to my knowledge).