top | item 45367085

Docker Hub Is Down

199 points| cipherself | 5 months ago |dockerstatus.com

100 comments

order
[+] c0balt|5 months ago|reply
Exceeded their quota, probably, based on my recent experience with dockerhub
[+] __turbobrew__|5 months ago|reply
Anyone have recommendations for an image cache? Native kubernetes a plus.

What would be really nice is a system with mutating admission webhooks for pods which kicks off a job to mirror the image to a local registry and then replaces the image reference with the mirrored location.

[+] edoceo|5 months ago|reply
We do a local (well, internal) mirror for "all" these things. So, we're basically never stuck. It mirrors our CPAN, NPM, Composer, Docker and other of these web-repos. Helps on the CI tooling as well.
[+] da768|5 months ago|reply
Not Google Artifact Registry... Our Docker Hub pull-through mirror went down with the Docker Hub outage. Images were still there but all image tags were gone
[+] issei|5 months ago|reply
I've been using https://github.com/enix/kube-image-keeper on some of my clusters - it is a local docker registry running on cluster, with a proxy and mutation webhooks. I also evaluated spegel, but currently it isn't possible to setup on GKE
[+] andrewstuart2|5 months ago|reply
CNCF has harbor [0], which I use at home and have deployed in a few clusters at work, and it works well as a pull through cache. In /etc/containers/registries.conf it's just another line below any registry you want mirrored.

    [[registry]]
    location = "docker.io"
    [[registry.mirror]]
    location = "core.yourharbor.example.com/hub"
Where hub is the name of the proxy you configured for, in this case, docker.io. It's not quite what you're asking for but it can definitely be transparent to users. I think the bonus is that if you look at a podspec it's obvious where the image originates and you can pull it yourself on your machine, versus if you've mutated the podspec, you have to rely on convention.

[0] https://goharbor.io/

[+] alias_neo|5 months ago|reply
Depending on what other (additional) features you're willing to accept, the GoHarbor[0] registry supports pull-through as well as mirroring and other features, it's a nice registry that also supports other OCI stuff like Helm charts, and does vulnerability scanning with "Interrogation Services" like Trivy.

I've been using it at home and work for a few years now, might be a bit overkill if you just want a simple registry, but is a really nice tool for anyone who can benefit from the other features.

[0] https://goharbor.io/

[+] philipallstar|5 months ago|reply
You can use Artifactory as a "front" for a variety of registries, including Docker, so it'll pull once and then use its cached image.
[+] tfolbrecht|5 months ago|reply
I usually do upstream image mirroring as part of CI. Registries are built into GitLab, AWS (ECR), GitHub, etc
[+] gvkhna|5 months ago|reply
Github actions buildx also going down is a really unintended consequence. It would be great if we could mirror away from docker entirely at this point but I digress.
[+] tfolbrecht|5 months ago|reply
There's a registry image for OCI containers that is pretty painless to set up and low maintenance, can use s3 as a storage backend.

https://hub.docker.com/_/registry

Your git provider probably also has a container registry service built in.

[+] switz|5 months ago|reply
I didn't even really realize it was a SPOF in my deploy chain. I figured at least most of it would be cached locally. Nope, can't deploy.

I don't work on mission-critical software (nor do I have anyone to answer to) so it's not the end of the world, but has me wondering what my alternate deployment routes are. Is there a mirror registry with all the same basic images? (node/alpine)

I suppose the fact that I didn't notice before says wonderful things about its reliability.

[+] tom1337|5 months ago|reply
I guess the best way would be to have a self-hosted pull-through registry with a cache. This way you'd have all required images ready even when dockerhub is offline.

Unfortunately that does not help in an outage because you cannot fill the cache now.

[+] matt_kantor|5 months ago|reply
> I don't work on mission-critical software

> wondering what my alternate deployment routes are

If the stakes are low and you don't have any specific need for a persistent registry then you could skip it entirely and push images to production from wherever they are built.

This could be as simple as `docker save`/`scp`/`docker load`, or as fancy as running an ephemeral registry to get layer caching like you have with `docker push`/`docker pull`[1].

[1]: https://stackoverflow.com/a/79758446/3625

[+] XCSme|5 months ago|reply
It's a bit stupid that I can't restart (on Coolify) my container, because pulling the image fails, even though I am already running it, so I do have the image, I just need to restart the Node.js process...
[+] sublinear|5 months ago|reply
Somewhat unrelated, but GitLab put out a blog post earlier this year warning users about Docker Hub's rate limiting: https://about.gitlab.com/blog/prepare-now-docker-hub-rate-li...

We chose to move to GitLab's container registry for all the images we use. It's pretty easy to do and I'm glad we did. We used to only use it for our own builds.

The package registry is also nice. I only wish they would get out of the "experimental" status for apt mirror support.

[+] s_ting765|5 months ago|reply
Status report says issue with authentication fixed but it's far worse than that. This incident also took down docker pull for public images with it.
[+] rickette|5 months ago|reply
Well to be fair: this doesn't happen very often. It's quite a stable service in my experience.
[+] miller_joe|5 months ago|reply
I was hoping google cloud artifact registry pull-thru caching would help. Alas, it does not.

I can see an image tag available in the cache in my project on cloud.google.com, but after attempting to pull from the cache (and failing) the image is deleted from GAR :(

[+] qianli_cs|5 months ago|reply
I think it was likely caused by the cache trying to compare the tag with Docker Hub: https://docs.docker.com/docker-hub/image-library/mirror/#wha...

> "When a pull is attempted with a tag, the Registry checks the remote to ensure if it has the latest version of the requested content. Otherwise, it fetches and caches the latest content."

So if the authentication service is down, it might also affect the caching service.

[+] rshep|5 months ago|reply
I’m able to pull by the digest, even images that are now missing a tag.
[+] breatheoften|5 months ago|reply
In our ci setting up the docker buildx driver to use the artifact registry pull through cache involves (apparently) an auth transaction to dockerhub which fails out
[+] esafak|5 months ago|reply
What's the easiest way to cache registries like docker, pypi, and npm these days?
[+] lambda|5 months ago|reply
The images I use the most, we pull and push to our own internal registry, so we have full control.

There are still some we pull from Docker Hub, especially in the build process of our own images.

To work around that, on AWS, you can prefix the image with public.ecr.aws/docker/library/ for example public.ecr.aws/docker/library/python:3.12 and it will pull from AWS's mirror of Docker Hub.

[+] pm90|5 months ago|reply
Someone mentioned Artifactory; but its honestly not needed. I would very highly recommend an architecture where you build everything into a docker image and push it to an internal container registry (like ecr; all public clouds have one) for all production deployments. This way, outages only affect your build/deploy pipeline.
[+] viraptor|5 months ago|reply
You pull the images you want to use, preferably with some automated process, then push them to your own repo. And anyways use your own repo when pulling for dev/production. It saves you from images disappearing as well.
[+] manasdas|5 months ago|reply
Therefore keep a local registry mirror. You will get it from local cache all the time.
[+] taberiand|5 months ago|reply
So that's why. This gave me the kick I needed to finally switch over the remaining builds to the pull-through cache.
[+] XCSme|5 months ago|reply
Yup, my Coolify deployments were failing and I didn't know why : https://softuts.com/docker-hub-is-down/

Also, isn't it weird that it takes so long to fix given the magnitude of the issue? Already down for 3 hours.

[+] philip1209|5 months ago|reply
Development environment won't boot. Guess I'll go home early.
[+] MASNeo|5 months ago|reply
For what it's worth, my debugging made me install the latest docker version. So the outage is good for something ;-)
[+] zenmac|5 months ago|reply
This is one of the reasons I don't want to use docker on production machines and have started to use systemd again!!
[+] Too|5 months ago|reply
Hard to see if this is /s or not. Nobody is forcing you to run images straight from dockerhub lol. Every host keeps the images already on it. Running a in-house registry is also a good idea.