Ask HN: What do you self-host?
538 points| aeleos | 6 years ago
I am currently running an Unraid server with some docker containers, here are a few of them: Plex, Radarr, Sonarr, Ombi, NZBGet, Bitwarden, Storj, Hyrda, Nextcloud, NginxProxyManager, Unifi, Pihole, OpenVPN, InfluxDB, Grafana.
Some comments were deferred for faster rendering.
mavidser|6 years ago
All web-services are reverse-proxied through traefik
At home:
On a remote server:tnsittpsif|6 years ago
nerdponx|6 years ago
teddyh|6 years ago
I think the problem is entirely caused by the US having absolutely abysmal private internet speeds and capacity. Since you can’t then have your own server at home, you are forced to have it elsewhere with sensible internet connections.
It’s as if, in an alternate reality, no private residences had parking space for cars; no garages, no street parking. Everyone would be forced to either use public transport, taxis and chauffeur services to get anywhere. Having a private vehicle would be an expensive hobby for the rich and/or enthusiasts, just like having a personal server is in our world.
IanCal|6 years ago
I do everything for little to nothing in my life, and there's no reasonable default as to where the line is other than a cost/benefit comparison.
For my for many years owning a car was far more expensive than renting or getting taxis when needed. Owning a car absolutely would have been an expensive hobby, and the same is true for many in cities.
Having a personal server is exceptionally cheap. I had a VPS unnoticed recently I'd forgotten to cancel which cost about 10 dollars per year. That's about one minimum wage hour where I live. If you mean literally a personal server a raspberry pi can easily run a bunch of things and can cost about the same as a one off.
It's time, and upfront costs of software. If I want updates, and I do (security at least) I need some ongoing payments for those, and then I need to manage a machine. That management is better done by people other than me (as even if they earned the same as me they'd be faster and better) and they can manage more machines without a linear increase in their time.
So why self host? Sometimes it'll make sense, but the idea it should be the default to me doesn't hold. Little needs to be 100% in house, and sharing things can often be far more efficient. Software just happens to be incredibly easy to share.
bhauer|6 years ago
The trend definitely traces to the advent and eventual domination of asymmetric Internet connectivity. My first DSL connection was symmetric, so peer-to-peer networking and running servers ("self-hosting") were just natural. Since then, asymmetric bandwidth has ruled the US.
It's not so much that connectivity technology in the US is strictly poor—many cities have options providing hundreds of megabits or a gigabit or more of aggregate bandwidth. It's that the capacity allocation of some shared delivery platforms (e.g., cable) is dramatically biased toward download/consumption, and against upload/share/host. And there's no way for consumers to opt for a different balance. I'd gladly take 500/500 versus 1000/50. Even business accounts, which for their greatly increased costs are a refuge of symmetric connectivity and static IPs, are more commonly asymmetric today.
I think that this capacity imbalance and bias toward consumption snowballs and reinforces the broader assumptions of consumption at the edge (why make a product you self-host when most people don't have the proper connectivity?). This in turn means more centralization of services, applications, and data.
Nevertheless, even with mediocre upload speeds (measured in mere tens of megabits), I insist on self-hosting data and applications as much as I can muster. All of my devices are on my VPN (using the original notion of "VPN," meaning quite literally a virtual private network; not the more modern use of VPN to mean "encrypted tunnel to an Internet browsing egress node located in a data center"). For example, why would I use Dropbox when I can just access my network file system from anywhere? To me, it's a matter of simplicity. Everything I use understands a simple file system.
cyphar|6 years ago
It also acts as an NFS server for my media center (Kodi -- though I really am not a huge fan of LibreELEC) to pull videos, music, and audiobooks from. Backups are done using restic (and ZFS snapshots to ensure they're atomic) and are pushed to BackBlaze B2.
I used to run an IRC bouncer but Matrix fills that need these days. I might end up running my own Gitea (or gitweb) server one day though -- I don't really like that I host everything on GitHub. I have considered hosting my own email server, but since this is all done from a home ISP connection that probably isn't such a brilliant idea. I just use Mailbox.org.
[1]: https://github.com/cyphar/cyphar.com/tree/master/srv
douglascoding|6 years ago
I plan to use Wireguard too, so I shouldn't run on containers? Can you elaborate on that?
mwcampbell|6 years ago
This is a bit tangential, but to clarify, do you mean that you listen to audiobooks on your TV using Kodi? Do you also have a way of syncing them to a more portable device, like your phone?
sdan|6 years ago
Overleaf: https://sdan.xyz/latex
A URL Shortener: https://sdan.xyz
All my websites (https://sdan.xyz/drf, https://sdan.xyz/surya, etc.)
My blog(s) (https://sdan.xyz/blog, https://sdan.xyz/essays)
Commento commenting server (I don't like disqus)
Monitoring (https://sdan.xyz/monitoring, etc.)
Analytics (using Fathom Analytics) and some more stuff!
djsumdog|6 years ago
I wrote this to setup my web server, mail server and VPN server, and auto-generate all my VPN keys.
https://github.com/sumdog/bee2
RulerOf|6 years ago
I see a lot of people putting their home stuff behind CloudFlare, but when I reviewed their free tier, I didn’t actually see any security benefit to outweigh the privacy loss, and I didn’t see that covered on your blog post.
whalesalad|6 years ago
Here’s my home lab: https://imgur.com/a/aOAmGq8
I don’t self host anything of value. It’s not cost effective and network performance isn’t the best. Google handles my mail. GitHub can’t be beat. I use Trello and Notion for tracking knowledge and work, whether personal or professional. Anything else is on AWS. I do have a VPN though so I can access all of this when I’m not home.
The NAS is for backing up critical data. R720 was bought to experiment with Amazon Firecracker. It’s usually off at this point. Was running ESXI, now running Windows Server evaluation.
The desktop on the left is the new toy. I’m learning AD and immersing myself 100% in the Microsoft stack. Currently getting an idiomatic hybrid local/azure/o365 setup going. The worst part about planning a MS deployment is having to account for software licensing that is done on a per-cpu-core basis.
Marsymars|6 years ago
The status quo is radically anti-consumer, IMO, as radical as abolition of all copyright would be.
LeoPanthera|6 years ago
tbyehl|6 years ago
Have been meaning to move more to colo, especially my Wordpress install and some Wordpress.com-hosted sites, but inertia.
[0] https://support.cloudflare.com/hc/en-us/articles/204899617-A...
[1] https://www.cloudflare.com/ips/
NewDimension|6 years ago
Pmop|6 years ago
zelly|6 years ago
Docker running random stuff
Used to run Pihole until I got an Android and rooted it. Used to mess with WebDAV and CalDAV. Nextcloud is a mess; plain SFTP fuse mounts work better for me. My approach has gone from trying to replicate cloud services to straight up remoting over SSH (VNC or terminal/mosh depending connectivity) to my home computer when I want to do something. It's simple and near unexploitable.
This is the way it should always have been done from the start of the internet. When you want to edit your calendar, for example, you should be able to do it on your phone/laptop/whatever as a proxy to your home computer, actually locking the file on your home computer. Instead we got the prolifetation of cloud SaaSes to compensate for this. For every program on your computer, you now need >1 analogous but incompatible program for every other device you use. Your watch needs a different calendar program than your gaming PC than your smart fridge, but you want a calendar on all of them. M×N programs where you could have just N, those on your home computer, if you could remote easily. (Really it's one dimension more than M×N when you consider all the backend services behind every SaaS app. What a waste of human effort and compute.)
dmos62|6 years ago
Why computer at home though? For someone who moves around a lot and doesn't invest into "a home", this would be bothersome. Not to mention it's more expensive, in terms of energy and money. I think third-party data centers are fine for self-hosting.
oarsinsync|6 years ago
ricardbejarano|6 years ago
* MinIO: for access to my storage over the S3 API, I use it with restic for device backups and to share files with friends and family
* CoreDNS: DNS cache with blacklisted domains (like Pihole), gives DNS-over-TLS to the home network and to my phone when I'm outside
* A backup of my S3-hosted sites, just in case (bejarano.io, blog.bejarano.io, mta-sts.bejarano.io and prefers-color-scheme.bejarano.io)
* https://ideas.bejarano.io, a simple "pick-one-at-random" site for 20,000 startup ideas (https://news.ycombinator.com/item?id=21112345)
* MediaWiki instance for systems administration stuff
* An internal (only accessible from my home network) picture gallery for family pictures
* TeamSpeak server
* Cron jobs: dynamic DNS, updating the domain blacklist nightly, recursively checking my websites for broken links, keeping an eye on any new release of a bunch of software packages I use
* Prometheus stack + a bunch of exporters for all the stuff above
* IPsec/L2TP VPN for remote access to internal services (picture gallery and Prometheus)
* And a bunch of internal Kubernetes stuff for monitoring and such
I still have to figure out log aggregation (probably going to use fluentd), I want to add some web-based automation framework like NodeRED or n8n.io for random stuff. I'd also like to host some password manager but I still have to study that.
I also plan on rewriting wormhol.org into supporting any S3 backend, so that I can bind it's storage with MinIO.
And finally, I'd like to move off single-disk storage and get a decent RAID solution to provide NFS for my cluster, as well as a couple more nodes to add redundancy and more compute.
Edit: formatting.
whycombagator|6 years ago
I would be _very_ interested in a write up/explanation of this set up
devthane|6 years ago
itm|6 years ago
https://github.com/epoupon/lms for music
https://github.com/epoupon/fileshelter to share files
Eveything is packaged on debian buster (amd64 and armhf) and run behind a reverse proxy.
djsumdog|6 years ago
One UI question? Is there a reason you left off volume controls? That's something that always annoys me still about Bandcamp and I had submitted a patch to Mastodon to create a volume control for their video component.
EvanAnderson|6 years ago
kstenerud|6 years ago
I have around 10 desktops that run in containers in various places for various common tasks I do. Each one has a backed up homedir, and then I have a ZFS-backed fileserver for centralized data. I connect to them using chrome remote desktop or x2go. I've had my work machine die one time too many, so with these scripts I can go from a blank work machine to exactly where I left off before the old one died, in a little over an hour. None of my files are stuck to a particular machine, so I can run on a home server, and then when I need to travel, transfer the desktop to a laptop, then transfer it back again when I get home. Takes about 10 minutes to transfer it.
https://github.com/kstenerud/virtual-builders
I also run most of my server apps this way:
https://github.com/kstenerud/virtual-builders/tree/master/ma...
letstrynvm|6 years ago
Incoming mail points directly to an RPi at home on dsl... Postfix + Dovecot IMAP. It's externally accessible, my dedicated server does the dynamic dns to point to the RPi; the domain MX points to that. Outgoing mail forwards through the dedicated server, which has an IP with good reputation and DKIM.
This gets me a nice result that my current and historical email is delivered directly to, and stays at, home, and my outgoing mail is still universally accepted. There's no dependency on google or github. There's no virtualization, no docker, no containers, just Linux on the server and on the rpi to keep up to date. It uses OS packages for everything so it stays up to date with security updates.
rolleiflex|6 years ago
I also host Aether P2P (https://getaether.net) on a Raspberry Pi-like device, so it helps the P2P network. But I’m biased on that last one, it’s my own software.
thegeekbin|6 years ago
I blog about this stuff if anyone’s interested: https://thegeekbin.com/
Youden|6 years ago
h1d|6 years ago
You don't want less tested web app to expose some security hole for someone to start snooping on your traffic toward BitWarden after SSL termination.
If you don't want an extra box at home, you can always get a $5/mo cloud instance for public stuff, where you don't have to worry about increased electricity bill from DDoS having CPU spiked or choking your home network.
DominoTree|6 years ago
On the front end I have two 1Gbit circuits (AT&T and Google) going into an OPNSense instance doing load-balancing and IPS running on a Dell R320 with a 12-thread Xeon and 24GB of RAM
Services are hosted on a Dell R520 with 48GB RAM and two 12-thread Xeons running Ubuntu and an up-to-date ZFS on Linux build.
Media storage handled by two Dell PowerVault 1200 SAS arrays.
Back-end is handled by a Cisco 5548UP and my whole apartment is plumbed for 10Gbit.
Havoc|6 years ago
Holy hell. How did that come about?
menssen|6 years ago
I live in a stable first-world democracy. Or, since it seems to be getting less stable recently, maybe a better way to put it is: I participate in a stable global economy. If "the cloud" catastrophically fails to the point where I lose all of the above without warning, I will likely have bigger problems than never being able to watch a favorite tv show again.
I wonder if this exposes two kinds of people: those who value mobility, and are more comfortable limiting the things that are important to them to a laptop and a bug-out bag, and those who value stability, and are inclined to build self-sufficient infrastructure in their castles.
elagost|6 years ago
I don't self host a lot of services (and the ones that do could go away tomorrow without hurting me much) but I only have one cloud resource - email. It kind of has to be that way for various reasons; I'd self host if I could reasonably do so. I also think I value my $75/mo more than I value an endless stream of entertainment.
(edit: just wanted to say, thanks for posting this. It is a valuable discussion point.)
olalonde|6 years ago
[0] https://github.com/huan/docker-simple-mail-forwarder
folkhack|6 years ago
* It's a target for my rsync backups for all my client systems (most critical use); Docker TIG stack (Telegraf, InfluxDB, Grafana) which monitors my rackmount APC UPS, my Ubiquiti network hardware, Docker, and just general system stats; Docker Plex; Docker Transmission w/VPN; Docker Unifi; A custom network monitor I built that just pings/netcats certain internal and external hosts (not used too seriously but it comes in handy); and finally a neglected Minecraft server.
I went for low power consumption since it's an always-on device and power comes at a premium here + fanless. I highly suggest the NUC as it's a highly capable device and with plenty of power if upgraded a bit!
ekianjo|6 years ago
kissgyorgy|6 years ago
mackrevinack|6 years ago
yankcrime|6 years ago
https://dischord.org/2019/07/23/inside-the-sausage-factory/
At home I have:
The DS412+ is my main network storage device, with various things backed up to the Microserver. Aside from the OEM services it also runs Minio (I use this for local backups from Arq), nzbget, and Syncthing in Docker containers.Mister_Snuggles|6 years ago
FreeBSD server running various things:
* Home Assistant, Node-RED, and some other home automation utilities running in a FreeBSD Jail.
* UniFi controller in a Debian VM.
* Pi-Hole in a CentOS VM.
* StrongSwan in a FreeBSD VM.
* ElasticSearch, Kibana, Logstash, and Grafana running in a Debian VM.
* PostgreSQL on bare metal.
* Nginx on bare metal, this acts as a front-end to all of my applications.
I also have:
* Blue Iris on a dedicated Windows box. This was a refurbished business desktop and works well, but my needs are starting to outgrow it.
* A QNAP NAS for general storage needs.
Future plans are always interesting, so in that vein here are my future plans:
Short term:
* Move my home automation stuff out of the FreeBSD Jail into a Linux VM. The entire Home Assistant ecosystem is fairly Linux-centric and even though it works on FreeBSD, it's more pain than I'd really like. Managing VMs is also somewhat easier than managing Jails, though I'm sure part of this is that I'm using ezjail instead of something more modern like iocage.
* Get Mayan-EDMS up and running. I hate paper files, this will be a good way to wrangle all of them. I've used it before, but didn't get too deep into it. This time I'm going all-in.
Medium term:
* Replace my older cameras with newer models.
* Possibly upgrade my Blue Iris machine to a more powerful refurbished one.
* Create a 'container VM', which will basically be a Linux VM used for me to learn about containers.
Long term:
* Replace my FreeBSD server with new hardware running a proper hypervisor (e.g., Proxmox, VMware ESXi). This plan is nebulous as what I have meets my needs, this is more about learning new tools and ways of doing things.
boredpenguin|6 years ago
• Apache: hosting a few websites and a personal (private) wiki.
• Transmission: well, as an always-on torrent client. Usually I add a torrent here, wait for it to download and then transfer it via SFTP to my laptop.
• Gitea: mostly to mirror third party repos I need or find useful.
• Wireguard: as a VPN server for all my devices and VPS, mostly so I don't need to expose SSH to the internet. Was really easy to setup and it's been painless so far.
0x0aff374668|6 years ago
stiray|6 years ago
- httpd
- nextcloud (mostly for android syncing, for normal file operations I prefer sftp). Nextcloud is great but the whole js/html/browser is clumsy.
- roundcube (again mostly imap but just to have alternative when phone isnt available - I havent used it for ages)
- postfix
- dovecot
- squid on separate fib with paid vpn (mitming all the traffic, removing all internet "junk" from my connections, all my devices, including android are using it over ssh tunnel).
- transmission, donating my bandwidth to some OSS projects
- gitolite, all my code goes there
I think this is it.
Everything is running on mitx board, with 16gb of ram, 3x 3tb toshiba hdds in zraid and additional 10tb hitachi disk. FreeBSD. 33 watts.
jasonkester|6 years ago
it costs about $800/month for the half cage and all the hardware in it, when you amortise it out. And there's plenty of performance overhead for when one project gets a lot of attention or I want to ad something new.
Pretty much the only thing I use cloud computing for is the nightly job for S3stat, because it fits the workload pattern that EC2 was designed for. Namely, it needs to run 70 odd hours of computing every day, and gets 3 hours to do it in.
For SaaS sized web stuff, self hosting still makes the most sense.
kemenaran|6 years ago
So I set up Yunohost [0] on a small box, and now I install self hosted services whenever I need them. Installing a new service is a breeze–but more importantly, upgrading them is a breeze to.
For now I self host Mattermost, Nextcloud, Transmission.
[0] https://yunohost.org
hendry|6 years ago
Tbh I run hot and cold about self hosting since after work, I really really want to be able relax at home.
Not wonder why the hell my nuc hasn't come up after a reboot. Or why is it so hard to increase the disk space on my FreeNAS https://www.ixsystems.com/community/threads/upgrading-storag...
dcchambers|6 years ago
I wasn't happy with any of the free wiki hosting solutions available so I ended up self-hosting a mediawiki site. It's been...challenging...to convince my wife and family to adapt and use wiki markup.
I've been considering switching to something that uses standard markdown instead since it's easier to write with.
Jaruzel|6 years ago
I used to self-host a lot more, but have been paring back recently.
canada_dry|6 years ago
Home automation/security system + 'Alexa': completely home grown using python + android + arduino + rpi + esp32
dnate|6 years ago
I have hosted media folders/streaming applications for friends and family, but this has been by far my most used and most useful hack.
Macha|6 years ago
* Unbound for dns-over-tls and single point of config hostnames for my home network
* Syncthing for file sync
* offlineimap to backup my email accounts
* Samba for a home media library
* cron jobs to backup my shares
* Unifi controller
On my todo list:
* Scheduled offsite backup (borg + rsync.net being the top contender currently)
* Something a bit more dedicated to media streaming than smb. some clients like vlc handle it fine, others do not.
* Pull logs for my various websites locally
vermilingua|6 years ago
What do you all spend on this sort of thing? Whether hosting remotely or on local hardware, what would you say is the rough monthly/annual cost to move your Netflix/Spotify/etc equiv to a self-hosted setup (excluding own labor)?
chrissnell|6 years ago
- A weather station that lives on a pole on the yard. Powered by GopherWX https://github.com/chrissnell/gopherwx
- InfluxDB for weather station
- Heatermeter Barbecue controller
- oauth2_proxy, fronted by Okta, to securely access the BBQ controller while I'm away. This proxy is something that everyone with applications hosted on their home network should look into. Combined with Okta, it's much easier than running VPN.
In the public cloud, I host nginx, which runs a gRPC proxy to the gopherwx at home. I wrote an app to stream live weather from my home station to my desktops and laptops and show it in a toolbar.
nginx in the cloud also hosts a public website displaying my live weather, pulled as JSON over HTTPS from gopherwx at home.
ohiovr|6 years ago
dmclamb|6 years ago
I have a second raspberry pi running a version of Kali Linux. I only hack my own stuff for learning.
Once upon a time I ran a public facing website and quake server, and published player stats. No time these days for much play.
zzo38computer|6 years ago
geek_at|6 years ago
https://github.com/HaschekSolutions/opentrashmail
hanklazard|6 years ago
(I guess these may not really be “self-hosted” since I don’t make them publically accessible through ports ... just vpn in to my home network)
yogsototh|6 years ago
- my websites with nginx
- IRC (ngircd)
- ZNC
- espial for bookmarks and notes
- node-red to automate RSS -> twitter and espial -> pinboard
- transmission
- some reddit bots manager I’ve written in Haskell+Purescript.
- some private file upload system mostly to share images in IRC in our team
- goaccess to self host privacy respecting analytics
At home, Plex.
moutansos|6 years ago
Basically all the stuff I don't want to pay a cloud provider to host.
Overall the R720 with 48GB of ram has been one of my best buys hands down. down the road I plan on grabbing a second server and a proper NAS or unraid setup.
nilsandrey|6 years ago
- docker (just dev env with a lot of images, almost everything I can is tested in there, and maybe used there too. Just on VM if is a desktop gadget or app)
- Calibre- Windows Media share feature for remote videos on devices and TV (, don't like it really, mess with subtitles and really will look for a docker oss alternative)
Wish list:
- wallabag
- firefox-sync (stuck on Chrome yet, no alternative on this found)
- email sync
It's not so great for now. Looking on this thread for contacts and calendar (currently used from the cloud classic providers)
ehnto|6 years ago
Everything. I keep infrastructure simple as I found as a developer, infrastructure configuration, dependency issues and updates took an extraordinary amount of time while providing zero benefit for products of a small to medium size. I do have a plan in place should I need to scale, but it is not worth maintaining an entirely different stack full of dependencies for the off chance I get a burst in traffic I can't handle.
notinventedhear|6 years ago
k_sze|6 years ago
- mail server in Docker container
- ZNC in Docker container
- Shadowsocks server
- Wekan as a Snap
- My blog, statically generated using Pelican, served from nginx
At home, I only have a Synology NAS that is exposed to the internet.
munmaek|6 years ago
I am unhappy with the complexity of Mayan EDMS. I'm debating moving to Paperless. All I want is a digital file system that 1) looks at directories and automatically handles files 2) has user permissions/personal files so I can let my family use it 3) has a web form for uploads.
I am planning to change gitea to sourcehut- the git service as well as builds.
Any ideas for things a raspberry pi 3 & 4 could be useful for?
Fiahil|6 years ago
I use NFS on the NAS for the storage unit. It's the only thing I need to backup.
bob1029|6 years ago
Relying on streaming providers, cloud email services, etc., has left me in a very foul mood lately and I feel like I need to take back control. My biggest trigger was when I purchased an actual physical audio CD (this year; because NONE of the popular streaming providers offer the album), ripped it to FLAC, and then realized I had no reliable/convenient way to expose this to my personal devices. I used to have a very elaborate setup with subsonic doing music hosting duty, and all of my personal devices were looped in on it. This was vastly superior to Spotify, et. al., but the time it takes to maintain the collection and services was perceived to be not worth it. From where I am sitting now, its looking like its worth it again.
How long until media we used to enjoy is squeezed completely out of existence because a handful of incumbent providers feel its no longer "appropriate" for whatever money-grabbing reasons?
kixiQu|6 years ago
* Pleroma/Mastodon - I had been using Pleroma, but I'm not happy about a few things, so I bit the bullet to upgrade to a t3.small and am now running Mastodon. I love all the concepts of the fediverse, though the social norms are still being ironed out.
* Write Freely (https://writefreely.org/) at https://lesser.occult.institute for my blog (right now mostly holds hidden drafts)
* Matrix (Synapse) and the Riot.im frontend for a group chat. I'm a little conflicted, because right now the experience around enabling E2EE is very alarming for low-tech users and a pain for anyone who signs in from many places, and if it isn't enabled I have better security just messaging my friends with LINE. That said, I really want to write some bots for it. Group chats are the future of social networking, they all say...
greenyouse|6 years ago
Surprisingly (at least to me), there are some really big companies like Microsoft, IBM/RedHat, and others pushing this workflow. The editor is supposed to basically be VSCode in browser and compatible with most extensions.
I'm using my RPi as a jump box and have some commands to turn on my home desktop + mount the file system and that kind of stuff when connecting. I've used it in the past and it's worked nicely.
I got k8s running but got blocked by some bugs when installing Che. Looks neat though. It would be cool to have a 2007 macbook with the computing power of a 2990WX workstation :).
winrid|6 years ago
The orchestrator can now deploy itself! All declarative service configuration with autoscaling etc. It manages the infra and service deployment for me. Thinking about open sourcing.
Nginx/nchan, NodeJS, static sites (vanilla/angular/react deployments), nfs, MongoDB, Redis
pjc50|6 years ago
I still have the email domain, because it's easier to run it forever than migrate all the things you signed up for. But actually running my own email is too much of an obligation and need to keep up on all the anti spam measures.
john61|6 years ago
unknown|6 years ago
[deleted]
bluedino|6 years ago
VMware ESXi, with VM's for Squid, DNS, MySQL, Nginx, Apache, basic file server, Gitlab, and one that's basically for IRSSI
Strongly considering just moving everything to Debian with containers for everything, easier to manage than VM's.
minimaul|6 years ago
On colo’d hardware:
- off-site backup server (Borg backup on top of zfs) - this is a dedicated box
- a mix of VMs and docker containers - mostly custom web apps
- email (it’s easier than you think)
At home:
- file server using zfs
- Nextcloud
- more custom web apps
- tvheadend
- VPN for remote access (IKEv2)
- gitlab
- gitlab ci
Also run an IPSec mesh between sites for secure remote access to servers etc
While my workplace uses AWS a massive amount, I still prefer to run my own hardware and software. Cloud services are not for me.
fractalf|6 years ago
harlanji|6 years ago
mmcnl|6 years ago
* Nextcloud - your own Dropbox! Amazing stuff.
* VPN - simple Docker service that is super reliable and easy to set up (docker-ipsec-vpn-server)
* Ghost - a very nice lean and mean blogging CMS
* MQTT broker for temperature sensors
* Samba server
* Deluge - Torrent client for local use
* Sabnzbd - NZB client
* Gitea - my own Git server
* Mail forwarder - very handy if you just want to be able to receive email on certain addresses without setting up a mailbox
* Pihole - DNS ad-blocking
* Jellyfin - self-hosted Netflix
It's become sort of my hobby to self-host these kind of things. I use all of these services almost daily and it's very rewarding to be able to fully self-host it. I also really love Docker, self-hosting truly entered a new era thanks to readily avaibable Docker images that make it very easy to experiment and run things in production without having to worry about breaking stuff.
conradfr|6 years ago
Of course you can't even tell Macos to not suspend wifi or whatever if you close the lid while on battery so now I'm trying to move it to a Raspberry Pi 4 but I've got an obscure ssl error with OTP22 on it while querying an api, so I'm trying to debug that instead ... oh the joy.
All my side projects and some clients are hosted old school style in a dedicated servers. I do overpay because that's the same price and machine since 2013 and yet it's still way cheaper than any cloud offering, especially because of the hosted databases pricings.
CarelessExpert|6 years ago
TT-RSS + mercury-parser + rss-bridge + Wallabag to replace Feedly and Pocket.
Syncthing + restic + rclone and some home grown scripting for backups.
Motion + MotionEye for home security.
Deluge + flexget + OpenVPN + Transdroid.
Huginn + Gotify for automation and push notifications.
Apache for hosting content and reverse proxying.
Running on a NUC using a mix of qemu/kvm and docker containers.
ekianjo|6 years ago
- Nginx
- Nextcloud (with Calendar/Contacts on it)
- IRC client (thelounge)
- IRC server
- DLNA server
- Ampache server
- video and photo library thru NFS (locally only)
- OpenVPN
- Shiori for bookmarks
- Gitea for private projects
- Syncthing (to keep a folder synchronized across my devices)
- Jenkins
Spivak|6 years ago
Zash|6 years ago
gargron|6 years ago
platz|6 years ago
theshrike79|6 years ago
The only things I host are either just hobbies or non-essentials:
At home: - Node-red for home automation - PiHole for ad filtering on the local network - Plex on my NAS for videos - A Raspi for reading my Ruuvitags and pushing the info to MQTT On Upcloud and DigitalOcean and a third place: - Unifi NVR (remote storage for security cameras) - Flexget + Deluge for torrents - InfluxDB + Grafana for visualizing all kinds of stuff I measure - Mosquitto for MQTT
_b8r0|6 years ago
- Nextcloud
- Mailu.io
- Huginn
- Gotify
- Airsonic
- Gitea
All on a dedicated box. Planning to add password sync, wallabag, syncthing a VPN and a few other features. Other boxes I have run various things from DNS to backup MXes and a WriteFreely instance on OpenBSD.
Internally I host a ton of stuff, mostly linked to a Plex instance.
algaeontoast|6 years ago
apple4ever|6 years ago
4 Ubuntu 16.04 servers:
- Nginx/PHP for Wordpress - MySQL - Redis - Mail
Planning to expand the the Nginx/PHP servers to at least two, and add load balancers. All certs are provided by an Ansible script using Lets Encrypt (yuck).
At home:
Proxmox running on two homebuilt AMD FX 8320 servers with 32GB each, with drives provided by FreeNAS on a homebuilt Supermicro server with about 10TB of usable space (on both HDDs and SSDs)
Ubuntu 16.04 Servers:
- 2x DNS - 2x DHCP - GitLab - Nagios - Grafana - InfluxDB - Redmine - Reposado - MySQL
Other:
- Sipecs
All set up via Ansible.
Next will set up a Kubernetes cluster (probably as far as I’ll get with containers).
DrAwdeOccarim|6 years ago
> Resilio Sync for iPhone pictures backups and "drop box" file access
> Transmission server
> SMB share of NAS to supply OSMC boxes on every TV
> Nighthawk N7000 running dd-wrt with a 500gb flash drive attached as storage for my Amcrest wifi cameras
> Edgerouter Lite running VPN server
> Hassbian for my zwave home automation stuff
> A pi with cheap speakers that I can log into and play a phone ringing sound so my wife will look at her phone!
HellfireHD|6 years ago
Also, kudos to those brave souls who are running Tor exit nodes!
Edit: Forgot a bunch
preid24|6 years ago
lostmsu|6 years ago
Now I only host my own project: http://billion.dev.losttech.software:2095/
Also regular Windows file sharing which I use for media server and backups.
Though I'd like to expand that. Maybe a hosted GitLab.