.. _sysreqs_sample_target_node_config_storage: Storage nodes ------------- We recommend that you separate Ceph nodes for scalability and robustness. The hardware estimate provided below is based on the requirement of 0.5 cores per Ceph-OSD CPU and 1 GB of RAM per 1 TB of Ceph-OSD space. You can configure all Ceph storage and journal hard disks in JBOD (Just a Bunch of Disks) mode on the RAID controller or plug them directly to the available SATA or SAS ports on the mainboard. Each storage node must include: +------------------------+---------------------------------+ | CPU | Single-socket CPU with at least | | | 4 physical cores | +------------------------+---------------------------------+ | RAM | 24 GB | +------------------------+---------------------------------+ | Storage | RAID 1 controller with at least | | | 500 GB capacity for the host | | | operating system disk | | | | | | For production installations, | | | set the Ceph object replication | | | factor to 3 or greater. | +------------------------+---------------------------------+ | Network | * For testing: 2 x 1 Gbit/s NICs| | | * For production: 2 x 10 Gbit/s | | | NICs | +------------------------+---------------------------------+ | Storage | * 18 TB for Ceph storage | | | (6 x 3 TB) | | | * 1-2 x 64 GB SSDs or more, for | | | the Ceph journal | +------------------------+---------------------------------+