We are still actively working on the spam issue.
Home servers are about learning and expanding your horizons. De-botnet your life. Learn something new. Serving applications to yourself, your family, and your frens feels good. Put your /g/ skills to good use for yourself and those close to you. Store their data with proper availability redundancy and backups and serve it back to them with a /comfy/ easy to use interface.
Most people get started with NAS. It’s nice to have a /comfy/ home for all your data. Streaming your movies/shows around the house and to friends. Know all about NAS? Learn virtualization. Spin up some VMs. Learn networking by setting up a pfSense box and configuring some VLANs. There's always more to learn and chances to grow. Think you’re god tier already? Setup openstack and report back to /hsg/.
Things that are online today might not be online forever. It's good to have a copy of something because you never know when it might get taken down due to copyright strikes or Big Tech censorship.
- 1 Hardware
- 2 Operating systems
- 3 File Systems and RAID
- 4 Containers
- 5 Server software
- 6 System administration software
- 7 External Links
- 8 See also
What hardware you get depends mostly on your use case. A simple file server can be run on an SBC with a couple hard drives attached. If you want to do more fancy things like virtualization, streaming 4K movies, etc you are going to want better hardware. If you plan on using ZFS or Btrfs, server grade hardware and ECC RAM are recommended but not required.
There are many roads to the Home Server. Each one has upsides and downsides. It's up to you to decide what works best for your requirements.
SBC and NUC
For simple home server use, such as a file server or single user direct play PLEX server, these options might be an appealing and inexpensive and energy efficient option for your home server. Expect performance issues if you try to scale though, and don't expect to be able to run multiple virtual machines or do heavy transcoding, you don't have much options for expansion and little to no options for upgrading either. You likely won't be able to add much more ram, and definitely won't be able to increase processing power unless you go the clustering route and purchase multiple units. If your use case is compute intensive or would require expansion cards (like a GPU for example) SBCs are likely not a good option.
Connecting hard drives via USB docks may have some performance impact. Use USB 3 where possible and don't attach too many drives to a single port.
- Intel NUC
- Search your NUC here for more information on it
Repurpose Old Hardware
If you have an old gaming PC, workstation, laptop, or spare parts lying around, you might be able to get away with using them as your server (provided they are not too old). Performance and capability will vary wildly from machine to machine. When in doubt post specs in /hsg/ and ask.
Laptops are not really designed for 24/7 use, but their battery does act as a built in UPS to a certain extent. Preferably use one with USB3 or better, older laptops might only have USB2 ports which will bottleneck any attached HDDs.
Build Your Own
If you have the money, buying new hardware is a viable (but expensive) option. Knowing exactly what you are getting and peace of mind that you can RMA any DOA items, as well as reasonable shipping prices are good reasons to buy new. A combination of new hardware and repurposed older hardware is also an option if you are on a bit of a budget.
If you are looking to build a ZFS/FreeNas server be sure to get a motherboard and CPU that support ECC RAM. Server motherboards are recommended as they have many features such as IPMI, Intel NICs, NIC teaming support, and more.
Supermicro/Asrock Rack are good options. "Prosumer" boards are usually incredibly expensive and not worth the money.
Modern AMD Ryzen CPUs all "unofficially" support ECC RAM, just make sure your motherboard supports it. Pretty much all AMD CPUs also support virtualization.
If using Intel CPUs check ark.intel.com for information on your CPU. Some features to look for:
- Intel Quick Sync Video allows for hardware accelerated transcoding.
- VT-x and VT-d are must haves if you plan on using virtualization of any kind.
- ECC RAM Support
Buying Used enterprise hardware can be a cheap, but somewhat unreliable option. Waiting for a good deal might not be for everyone but the rewards are great. Tremendous amounts of storage potential at a relatively cheap price. Some listings even come with drives preinstalled. Be aware though, rack mount hardware is usually pretty loud and older Xeons can be extremely energy inefficient.
Things to look out for when buying used:
- 32bit systems have a hard limit on Ram. Avoid at all costs.
- Some older legacy systems do not support UEFI and thus cannot boot UEFI OS's.
- Older hardware specs may become performance bottlenecks (Earlier SATA/PCIe/SAS/USB revisions).
- Some Very old (8-10 years) high end CPUs actually perform worse than modern low-to-mid end CPUs.
- Lack of support for potentially desired features such as QSV, 1GB/10GB Ethernet, m.2, etc.
- Some disk shelf/server backplanes are SAS only and won't accept standard Sata drives.
- Good places to find old server hardware:
Only buy a prebuilt NAS if you want to spend more and get less.
They are typically woefully underpowered for the price and you’re better served with a $65 Odroid than a $300 QNAP/Synology with a shitty Celeron and 1gb of ram. That said, they are the most noob friendly option with a GUI interface for setup.
It's massively cheaper to buy WD Easystores or WD Elements (when they go on sale) than it is to buy an equivalent size NAS hard drive like WD Red/Iron wolf. Just remember: YOU VOID YOUR WARRANTY. If your drive fails you are most likely fucked. When you buy regular NAS drives you are basically paying more for the warranty.
Some other things to consider if you decide to shuck:
- Shucked drives under 8TB might be SMR drives.
- Shucked drives lack the middle mounting hole that most other drives have. You may need an adaptor for your HDD trays if your case doesn’t support them. Some cases might not have adaptors at all, research before buying your case!
- Some 8 and 10TB drives are air-filled rather than filled with helium. These air-filled drives can run significantly hotter than the helium ones. Check the model number with Crystal Disk Info before shucking, if it has an H it is most likely a helium drive. If you have airflow constraints in your case, it might be better to try and get Helium drives, otherwise it shouldn't be much of an issue.
3.3v pin issue
Shucked drives WILL NOT BOOT with most consumer power supplies. This is because of a feature on enterprise drives that lets administrators reboot hard drives by powering the 3.3 volt pin which isn’t used on consumer hard drives. Consumer PSUs, of course, always power this pin, so the hard drive will be stuck in an infinite boot loop and never power on. This can be solved by covering the first three pins on the hard drive with insulating Kapton tape.
- DO NOT use liquid electrical tape. This can damage the drive.
- DO NOT cut the SATA power cable this can damage the drive and your PSU.
- Molex to SATA adapters DO work but be careful, as some of the poorly made ones can catch fire. I wouldn’t risk it.
SMR v CMR
SMR stands for "shingled magnetic recording" It's an alternative method to conventional magnetic recording (CMR) that traditional hard drives use. While SMR technology allows for greater data density, they are also slow compared to CMR. These drives are bad for NAS use cases and especially bad for ZFS due to compatibility issues. Just avoid them all together. All Seagate NAS drives are CMR. Easystores/Elements 8tb and above should be safe.
SSDs are recommended for the OS and programs only, or for use in cache, or ARC cache in ZFS in place of expensive ram.
Don't buy SSDs for main storage unless you want to spend tens of thousands. Check out the SSD buying guide for more on SSDs.
Expanding Your Storage
If you find you have run out of SATA ports on your motherboard but require more storage, there are a number of options for increasing the number of drives your server can support. The best and recommended approach is to use a SAS HBA with SAS to Sata breakout cables. Each SAS port can support up to 4 Sata drives (or even more if you use an expander). You can find used LSI SAS HBAs on eBay for relatively cheap which have ~2 internal ports, or 8 total SATA drives. Avoid SAS1 cards as they are far too old by now and have some limitations. If your case can no longer support more drives, you may want to look into buying an External SAS HBA, which will allow you to connect drives in an external enclosure directly to your server.
- Some videos on SAS controllers and cables which I found very helpful. If you are new to using SAS you should watch these:
Sata HBAs and port multipliers/Expanders are not recommended. They are garbage and not worth buying. SATA port multipliers specifically can cause issues when you try to us any kind of RAID with them.
There are some counterfeit LSI cards on the market, avoid Chinese sellers, sellers with no return policy, etc.
Racks and Cases
- Home server case guide
- Lack rack - Meme-y but practical and cheap solution for rack-mount equipment. Be wary of putting too much weight on them though
There are many options for which OS to use for your server, Ultimately it depends on your needs and budget. Some options like Proxmox require specific hardware which will increase overall cost of your server.
OMV is good enough if all you are storing is rarely accessed media. Standard Linux like Ubuntu supports software RAID and file systems like OpenZFS, and runs on pretty much anything. TrueNAS CORE (formerly FreeNAS) is BSD based and fairly simple to install and use but server grade hardware and ECC RAM are recommended.
- Debian based distros like Ubuntu are easy to work with and have plenty of documentation.
- Ubuntu usually does retarded things with their packages and versions (lib*-ubuntu1.l2), and pulls unstable software from Debian Sid.
- CentOS is a free alternative to RedHat and is also fairly popular.
Gentoo is usually too much trouble to be worth it, but it's ok.
Use mdadm, ZFS on Linux (ZoL) or Snapraid for data redundancy.
TrueNAS CORE is the free version of the premium TrueNAS and successor/replacement for FreeNAS. TNC is a FreeBSD based OS that utilizes ZFS for storage and has many available plugins for things like PLEX, bitorrent, and more. Has simple, easy to use GUIs to set up your services such as samba shares, etc.
Keep in mind it will install to the ENTIRE DRIVE and you won't be able to use the install drive for anything else. A small, cheap, M.2 SSD is a good option for the OS drive. Server grade hardware and ECC RAM is recommended.
Good for storing infrequently changed files like media files.
Website doesn't have much to show for. Check the forums for guides.
Supports Snapraid as a plugin. Can be used with mergerFS to pool drives together.
A Linux based Virtualization Environment that has built in ZFS support, Linux container support, and more.
Your CPU must support Virtualization in order to run Proxmox. (VT-x on Intel, AMD-V for AMD).
ECC RAM is recommended as per usual with ZFS.
A free version of the OS Synology uses for their NAS products. Obviously you won't get any support from Synology if you use this.
Supports differently sized physical disks and adding hard drives to expand as needed.
Comes with it's own RAID solution that technically isn't real RAID because all parity is stored on one or two disks. Not free, you need to fork over some money to buy it.
Supports differently sized physical disks and adding hard drives to expand as needed.
File Systems and RAID
You may want to consider a RAID array for long-term file storage. A proper RAID array can protect you against sudden drive failures, and some software RAID have more features to combat data degradation.
When deciding on what RAID level to use, try to aim for at least two disk redundancy for arrays larger than 4 disks. Rebuilding a RAID array is an intensive process and it's not uncommon for a second disk to fail during the process. RAID 5 and equivalents only offer 1 disk redundancy, so if another disk fails during your rebuild you're fucked and any data not backed up is lost.
Software RAID vs Hardware RAID
Software RAID typically has a number of features that are more beneficial than just standard RAID. Best in class at this moment is ZFS, which has automatic file self healing and file system level checksums to combat bitrot.
Btrfs is also good if you want an alternative to ZFS, but is still in development so be careful.
UnRAID is more noob friendly (but you pay for it).
Snapraid + merger FS is a viable (free) alternative to UnRAID if you don't feel like spending money on your OS.
A long standing, reliable file system and software RAID solution that works on BSD and Linux.
Supports up to 3 disk redundancy (RaidZ3) and has checksums and self healing to prevent data corruption.
Has some limitations, one major one being expansion is cumbersome.
If you plan on using RaidZ make sure you know what your array is going to be beforehand, you won't be able to add to it later.
Server hardware and at least 8GB ECC RAM recommended. Using old hardware or consumer hardware is not recommended.
A tool for creating and managing Linux software RAID arrays.
You can create file systems directly on the RAID arrays, and then use mergerFS to pool the file systems together.
Technically supports disks of different sizes, but it requires multiple partitions on the drives and is not recommended.
No built in checksums, Can use dm-integrity to detect errors, but has no way of dealing with them.
Required learning for management of drives on Proxmox.
A bit more confusing than partitioning drives normally but is very flexible. Allows for thin provisioning of storage, and pooling of multiple drives or mdadm arrays into logical volumes
Another reliable file system. Unlike ZFS it doesn't have built in software RAID features.
Default file system for most Linux distros. Does everything a good file system should do and more.
It's "B-Tree", not "Butter".
Has many of the same features as ZFS, including checksums and self healing.
STILL UNDER DEVELOPMENT USE AT OWN RISK. Potential for data loss.
RAID 1 features are stable on the most recent Linux kernels.
A Union file system that pools multiple file systems together under one mount point, allowing them to appear as one.
Has some advantages over LVM, you can use multiple disks with data already on them instead of having to create LVM volumes/groups. If a disk fails, since the data is not striped across multiple disks (like with LVM), data loss can be less drastic.
Works with multiple different file systems at the same time, including Windows's NTFS. Use with Snapraid or mdadm for disk redundancy.
Available as a plugin for OMV.
Has an impressive list of features including up to 6 disk redundancy and the ability to add hard drives to expand as needed.
Supports differently sized disks, allowing for more flexibility with expansion (your data disks must be equal to or smaller than your parity disks! If you try to add a new data disk that is larger than your parity drives you will run into issues).
Technically not "real" RAID and has some limitations. Read the manual.
Can be used with mergerFS to pool drives together while retaining a level of redundancy.
Available as a plugin for OMV.
Containers are a method of isolating running software from both the host OS and other software. You may also hear them called Jails or Chroot Jails if you are running some variant of BSD (such as TrueNAS CORE/FreeNAS).
There are a number of reasons why you would want this:
- Less overhead than standard virtual machines because you aren't virtualizing the kernel.
- Isolated software cannot interfere with each other or the host. If a container crashes it won't effect anything else.
- Like VMs, containers are portable. You can create a container, configure it however you want, and deploy it anywhere.
- Like VMs, removing containers and starting from scratch or a backup in the event of a fatal crash is easy.
- Docker and Podman containers are incredibly easy to deploy and you can find pre-configured container images online.
Best practice is to keep the base OS as clean as possible and install each individual application (such as PLEX, Samba, etc) in their own container. This makes your server much more stable since there is virtually no chance of a containerized application crashing your server, or an installation gone wrong from ruining your host OS.
LXC is the standard Linux containers. Available on most distros. You will likely be using these if you are running a Proxmox server. Since Linux containers are essentially just separate instances of Linux, you can't run Windows programs in them without using WINE.
LXD is a newer, more user friendly version of LXC. Has better management options for containers.
Instead of running as though it was an entire OS like LXC, Docker only virtualizes a single application. Can run on Windows as well as Linux. You will still need WINE to run Windows apps on Linux. Freemium software. Base software is free for individuals (you).
An alternative to Docker. Those using Docker can easily switch without issues. Unlike Docker, it does not use a single large server daemon. Uses "pods" which can contain more than one container.
Jails are BSD's version of containers. Since TrueNAS CORE/FreeNAS is FreeBSD based you will be using these instead of LXC/LXD.
- TrueNas Jail documentation
- Give Jails access to host storage - Jail version of Bind mounting
- FreeBSD Jail documentation
For a greater range of self hosting solutions and services see awesome self hosted software.
System administration software
For a greater range of sysadmin solutions and services see awesome sysadmin software.
- Home server hardware - Hayden James' home lab setup
- STH Forums - Good general resource for server questions
- Learn Command line
- HP T620 plus - Decent cheap computer. You can use it as a VPN, pfSense firewall, and more: