It would but you should not: http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-December/014846.html Kevin Am Di., 8. Jan. 2019 um 15:35 Uhr schrieb Rodrigo Embeita <rodrigo@xxxxxxxxxxxxxxx>: > > Thanks again Kevin. > If I reduce the size flag to a value of 2, that should fix the problem? > > Regards > > On Tue, Jan 8, 2019 at 11:28 AM Kevin Olbrich <ko@xxxxxxx> wrote: >> >> You use replication 3 failure-domain host. >> OSD 2 and 4 are full, thats why your pool is also full. >> You need to add two disks to pf-us1-dfs3 or swap one from the larger >> nodes to this one. >> >> Kevin >> >> Am Di., 8. Jan. 2019 um 15:20 Uhr schrieb Rodrigo Embeita >> <rodrigo@xxxxxxxxxxxxxxx>: >> > >> > Hi Yoann, thanks for your response. >> > Here are the results of the commands. >> > >> > root@pf-us1-dfs2:/var/log/ceph# ceph osd df >> > ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS >> > 0 hdd 7.27739 1.00000 7.3 TiB 6.7 TiB 571 GiB 92.33 1.74 310 >> > 5 hdd 7.27739 1.00000 7.3 TiB 5.6 TiB 1.7 TiB 77.18 1.45 271 >> > 6 hdd 7.27739 1.00000 7.3 TiB 609 GiB 6.7 TiB 8.17 0.15 49 >> > 8 hdd 7.27739 1.00000 7.3 TiB 2.5 GiB 7.3 TiB 0.03 0 42 >> > 1 hdd 7.27739 1.00000 7.3 TiB 5.6 TiB 1.7 TiB 77.28 1.45 285 >> > 3 hdd 7.27739 1.00000 7.3 TiB 6.9 TiB 371 GiB 95.02 1.79 296 >> > 7 hdd 7.27739 1.00000 7.3 TiB 360 GiB 6.9 TiB 4.84 0.09 53 >> > 9 hdd 7.27739 1.00000 7.3 TiB 4.1 GiB 7.3 TiB 0.06 0.00 38 >> > 2 hdd 7.27739 1.00000 7.3 TiB 6.7 TiB 576 GiB 92.27 1.74 321 >> > 4 hdd 7.27739 1.00000 7.3 TiB 6.1 TiB 1.2 TiB 84.10 1.58 351 >> > TOTAL 73 TiB 39 TiB 34 TiB 53.13 >> > MIN/MAX VAR: 0/1.79 STDDEV: 41.15 >> > >> > >> > root@pf-us1-dfs2:/var/log/ceph# ceph osd pool ls detail >> > pool 1 'poolcephfs' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 471 fla >> > gs hashpspool,full stripe_width 0 >> > pool 2 'cephfs_data' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 256 pgp_num 256 last_change 471 lf >> > or 0/439 flags hashpspool,full stripe_width 0 application cephfs >> > pool 3 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 256 pgp_num 256 last_change 47 >> > 1 lfor 0/448 flags hashpspool,full stripe_width 0 application cephfs >> > pool 4 '.rgw.root' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 471 flags ha >> > shpspool,full stripe_width 0 application rgw >> > pool 5 'default.rgw.control' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 47 >> > 1 flags hashpspool,full stripe_width 0 application rgw >> > pool 6 'default.rgw.meta' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 471 f >> > lags hashpspool,full stripe_width 0 application rgw >> > pool 7 'default.rgw.log' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 471 fl >> > ags hashpspool,full stripe_width 0 application rgw >> > >> > >> > root@pf-us1-dfs2:/var/log/ceph# ceph osd tree >> > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF >> > -1 72.77390 root default >> > -3 29.10956 host pf-us1-dfs1 >> > 0 hdd 7.27739 osd.0 up 1.00000 1.00000 >> > 5 hdd 7.27739 osd.5 up 1.00000 1.00000 >> > 6 hdd 7.27739 osd.6 up 1.00000 1.00000 >> > 8 hdd 7.27739 osd.8 up 1.00000 1.00000 >> > -5 29.10956 host pf-us1-dfs2 >> > 1 hdd 7.27739 osd.1 up 1.00000 1.00000 >> > 3 hdd 7.27739 osd.3 up 1.00000 1.00000 >> > 7 hdd 7.27739 osd.7 up 1.00000 1.00000 >> > 9 hdd 7.27739 osd.9 up 1.00000 1.00000 >> > -7 14.55478 host pf-us1-dfs3 >> > 2 hdd 7.27739 osd.2 up 1.00000 1.00000 >> > 4 hdd 7.27739 osd.4 up 1.00000 1.00000 >> > >> > >> > Thanks for your help guys. >> > >> > >> > On Tue, Jan 8, 2019 at 10:36 AM Yoann Moulin <yoann.moulin@xxxxxxx> wrote: >> >> >> >> Hello, >> >> >> >> > Hi guys, I need your help. >> >> > I'm new with Cephfs and we started using it as file storage. >> >> > Today we are getting no space left on device but I'm seeing that we have plenty space on the filesystem. >> >> > Filesystem Size Used Avail Use% Mounted on >> >> > 192.168.51.8,192.168.51.6,192.168.51.118:6789:/pagefreezer/smhosts 73T 39T 35T 54% /mnt/cephfs >> >> > >> >> > We have 35TB of disk space. I've added 2 additional OSD disks with 7TB each but I'm getting the error "No space left on device" every time that >> >> > I want to add a new file. >> >> > After adding the 2 additional OSD disks I'm seeing that the load is beign distributed among the cluster. >> >> > Please I need your help. >> >> >> >> Could you give us the output of >> >> >> >> ceph osd df >> >> ceph osd pool ls detail >> >> ceph osd tree >> >> >> >> Best regards, >> >> >> >> -- >> >> Yoann Moulin >> >> EPFL IC-IT >> >> _______________________________________________ >> >> ceph-users mailing list >> >> ceph-users@xxxxxxxxxxxxxx >> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@xxxxxxxxxxxxxx >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com