briggers | 2 years ago | on: Ask HN: Could you share your personal blog here?
briggers's comments
briggers | 3 years ago | on: Ask HN: Is having a personal blog/brand worth it for you?
I write fairly deep ML performance tuning articles at https://paulbridger.com and the (many) hours I've spent on each article have been hugely worth it.
Many people reach out to me via this work, and when we talk they already see me as an expert or already want to work with me.
I need to blog more, thanks for the reminder.
briggers | 3 years ago | on: Ask HN: What bits of fundamental knowledge are productivity multipliers?
briggers | 4 years ago | on: Podman can transfer container images without a registry
docker run -d -p 5000:5000 --name registry registry:2
https://docs.docker.com/registry/#:~:text=The%20Registry%20i....
briggers | 5 years ago | on: Reloadr – Hot code reloading tool for Python
It’s less about saving the time to re-run something, and more about removing conceptual overhead (I think).
briggers | 5 years ago | on: A terminal-based workflow for research, writing, and programming
briggers | 5 years ago | on: Object Detection at 1840 FPS with TorchScript, TensorRT and DeepStream
briggers | 5 years ago | on: Object Detection at 1840 FPS with TorchScript, TensorRT and DeepStream
briggers | 5 years ago | on: Object Detection at 1840 FPS with TorchScript, TensorRT and DeepStream
briggers | 5 years ago | on: Object Detection at 1840 FPS with TorchScript, TensorRT and DeepStream
briggers | 5 years ago | on: Object Detection at 1840 FPS with TorchScript, TensorRT and DeepStream
briggers | 5 years ago | on: Object Detection at 1840 FPS with TorchScript, TensorRT and DeepStream
briggers | 5 years ago | on: Object Detection from 9 FPS to 650 FPS
To me, seeing the GIL held for 40% of time and significant time spent waiting on GIL by other threads was a fairly strong indicator. Keen to hear your thoughts/experience on it.
briggers | 5 years ago | on: Object Detection from 9 FPS to 650 FPS
Everything lostdog says. I've had experience speeding up tracking immensely using the same big hammer I talk about in the article - moving the larger parts of tracking compute to GPU.
Also, in a tracking pipeline you'll generally have the big compute on pixels done up front. Object detection and ReID take the bulk of the compute and can be easily batched and run in parallel. The results (metadata) can then be fed into a more serial process (but still doing the N<->N ReID comparisons on GPU).
briggers | 5 years ago | on: Object Detection from 9 FPS to 650 FPS
Alternatively, there are some quite fast OSS libraries for object detection. Nvidia's retinanet will export to a TensorRT engine which can be used with DeepStream.
briggers | 5 years ago | on: Object Detection from 9 FPS to 650 FPS
Completely agree that almost none of the SoTA github repos are really ready for production and making this stuff work can be pretty hard.
Getting this done on C++ and moving up to the next level of performance is the focus of my next article :)
briggers | 5 years ago | on: Show HN: bbox-visualizer – Make drawing and labeling bounding boxes easy as cake
More useful to me would be something similar that operates on tensors on the GPU.
Doing image annotations on host/CPU often becomes a bottleneck.
briggers | 5 years ago | on: AI slays top F-16 pilot in DARPA dogfight simulation
briggers | 5 years ago | on: What would you do if you lost your Google account?
briggers | 6 years ago | on: Standardizing OpenAI’s deep learning framework on PyTorch
Are you aware of the Sequential module? It allows you to chain together layers into a single variable, making this repetition disappear into a single forward/__call__ on the Sequential.