I'm setting up a ceph cluster (first time for me) which in the end will be made of ~100 disks spread over 10 hosts. I'm going with a single erasure coded data pool to maximize disk space; my constraints are ~80% efficiency and a fault tolerance of 2 disks. This can be achieved most simply with a k=8 m=2 erasure code, but also with k=16 m=4 with the bonus of tolerating up to 4 disk faults.
I'm thus wondering which are the downsides of growing the number of stripes; a few come to my mind (e.g. increased CPU and network overhead due to increased file fragmentation) but given my very poor knowledge of the subject I'm not sure. I'd really appreciate any insight on this topic.
bluestore_min_alloc_size_hdd. If you have many small files your chunks will eat up space if you don't change the allocation size. So in conclusion, I would advise against using 48 chunks but rather stay somewhere between 8 and 18 chunks. We have a few customers with erasure-coded pools, both 18 chunks and 9 chunks work quite well.