I for one use and self-host Meshcentral. The GUI is ugly, but it works well.
I for one use and self-host Meshcentral. The GUI is ugly, but it works well.
In term of software compatibility, on Linux, you have the option of making chroots. Since the kernel devs makes a lot of effort to preserve compatibility, old software can still work fine. If I remember correctly, some kernel devs tested a while ago some really really old versions of bash, gcc, etc, and they still work fine with modern kernels.
Red Hat. Probably Canonical too.
I know it for a fact since I worked for a bank that chose Red Hat and since I also know someone working for Red Hat.
For those wondering, it also works with a Linux VM:
It’s not easy to set up, but it works. I’m able to run some games like Borderlands 3 running at ~50FPS with a resolution of 1920x1080 with visual effects set to the max (important: disable vsync in the games !).
Only problem is disk access. It tends to add some latency. So with games coded with their ass (ex: Raft), the framerate drops a lot (Raft goes down to 20FPS sometimes).
Yes I would count this game as self-hosted (as long as you don’t need a third-party service to start it). And yes I agree it is a pretty wide definition. But at the same time, I really think there are a lot of good reasons to not dismiss it:
To be honest, when it comes to self-hosting, I can’t shake this feeling that a lot of people are dismissing desktop apps immediately just because they are not cool nor hype anymore.
Regarding Syncthing, if I’m not mistaken, the Web UI can be opened to the network (most likely for headless servers) but by default it is only reachable through the loopback.
Regarding OP, for me, it wasn’t entirely clear at first whether they wanted network access or not. They clarified it later in comments.
It is “hosted” on your workstation. There is no need for a server-client relationship for self-hosting.
By requiring a server-client relationship, you’re making self-hosting uselessly hard to deploy and enforce a very specific design when others (P2P, file sync, etc) can solve the same problems more efficiently. For example, in my specific case, with Paperwork + Nextcloud file sync, my documents are distributed on all my workstations and always available even if offline. Another example is Syncthing which IMO fits the bill for self-hosting, but doesn’t fit your definition of self-hosted.
No it does not.
Self-hosted implies self-hosted. AFAIK, the end goal is being as autonomous as possible technologically-speaking. Why exclude desktop applications ?
AFAIK, unfortunately Dia hasn’t been maintained and hasn’t got a new release for a really long time. It’s still using GTK2.
You don’t even need Docker for draw.io: https://github.com/jgraph/drawio-desktop/releases
You can use du -sh
to figure out what’s using most of the space. Something along the line of:
sudo -i
du -sh /home /usr /var
du -sh /var/*
du -sh /var/log/*
# etc
If it’s one of your log files (likely), you can run something like tail -n 100 /var/log/[culprit]
or tail -F /var/log/[culprit]
to see what is being flooded in this log file exactly. Then you can try to fix it.
As suggested by others, your processes may be using too much memory. However I would also suggest you keep an eye on the output of dmesg
. Maybe one of your disks is failing.
I think there is some confusion here.
Paperwork is a desktop application, not a web application. (eh, self-hosting doesn’t necessarily always imply web applications ! :). I for one use Nextcloud and nextcloud-desktop to keep my Paperwork work directories in sync on all my computers.
Paperless is a web application. Paperwork is a desktop application.
Paperwork seems to fit most of the bill except for one thing: it won’t scroll to where the search hit is (but it will highlight the matching keywords).
Just beware Paperwork won’t just create an index. It’ll organize the PDF its own way in its own work directory.
(full disclosure: I’m its main dev)
I use OPNSense virtualized on top of Proxmox. Each physical interface of the host system (ethX
and friends) is in its own bridge (vmbrX
), and for each bridge, the OpenSense VM also has a virtual interface that is part of the bridge. It has worked flawlessly for months now.
I worked for a bank. When they decided to deploy Linux on their infrastructure, they chose RHEL and they have signed a big contract with RedHat for tech support.
Overall, they chose RedHat for the same reason they chose Microsoft before: tech support. They have >10000 engineers, and yet somehow they think they absolutely need tech support… They pay a lot for it. In my building, they even got a Microsoft engineer once a week on-site until Covid. I don’t know for the other people working for this bank, but I asked for Microsoft support only once in 2 years. In the end, their guy sent me back an email telling me “I’ve transmitted your question to the corresponding engineering team” and … diddlysquat.
Now to be fair, for paying customers, RHEL and Microsoft both ensure security updates for a really a long time. Red Hat pays a lot of people to backport security patches from upstream to previous versions. It allows companies like the bank I worked for to keep running completely crappy and obsolete software for an insane amount of time without having to worry too much about security vulnerabilities.
Anyway regarding RedHat contributions, a lot of them are subtle.
This list is far from exhaustive. I’m sure they have paid for a lot of other things you’re using daily without knowing it.
My wife and I use a Nextcloud application called Cospend.