[ERR] OSD_FULL: 1 full osd(s) - with 73% used

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi everyone!
I'm trying to understand why Ceph changed the state of one osd as full if
it is 73% used and full_ratio is 0.97.

Follow some information:

# ceph health detail
(...)
[ERR] OSD_FULL: 1 full osd(s)
    osd.11 is full
(...)

# ceph osd metadata
(...)

> {
>         "id": 11,
>         "arch": "x86_64",
>         "back_iface": "",
>         "bluefs": "1",
>         "bluefs_dedicated_db": "0",
>         "bluefs_dedicated_wal": "0",
>         "bluefs_single_shared_device": "1",
>         "bluestore_bdev_access_mode": "blk",
>         "bluestore_bdev_block_size": "4096",
>         "bluestore_bdev_dev_node": "/dev/dm-41",
>         "bluestore_bdev_devices": "nvme0n1",
>         "bluestore_bdev_driver": "KernelDevice",
>         "bluestore_bdev_partition_path": "/dev/dm-41",
>         "bluestore_bdev_rotational": "0",
>         "bluestore_bdev_size": "1000200994816",
>         "bluestore_bdev_support_discard": "1",
>         "bluestore_bdev_type": "ssd",
>         "ceph_release": "octopus",
>         "ceph_version": "ceph version 15.2.16
> (d46a73d6d0a67a79558054a3a5a72cb561724974) octopus (stable)",
>         "ceph_version_short": "15.2.16",
>         "cpu": "Intel(R) Xeon(R) CPU E5-2630 v4 @ 2.20GHz",
>         "default_device_class": "ssd",
>         "device_ids":
> "nvme0n1=Samsung_SSD_970_EVO_Plus_1TB_S59ANJ0N308050F",
>         "device_paths":
> "nvme0n1=/dev/disk/by-path/pci-0000:01:00.0-nvme-1",
>         "devices": "nvme0n1",
>         "distro": "ubuntu",
>         "distro_description": "Ubuntu 18.04.6 LTS",
>         "distro_version": "18.04",
>         "front_iface": "",
>         "hostname": "pf-us1-dfs6",
>         "journal_rotational": "0",
>         "kernel_description": "#182-Ubuntu SMP Fri Mar 18 15:53:46 UTC
> 2022",
>         "kernel_version": "4.15.0-173-generic",
>         "mem_swap_kb": "117440508",
>         "mem_total_kb": "181456152",
>         "network_numa_unknown_ifaces": "back_iface,front_iface",
>         "objectstore_numa_node": "0",
>         "objectstore_numa_nodes": "0",
>         "os": "Linux",
>         "osd_data": "/var/lib/ceph/osd/ceph-11",
>         "osd_objectstore": "bluestore",
>         "osdspec_affinity": "",
>         "rotational": "0"
>     } (...)


# cat /var/lib/ceph/osd/ceph-11/type

> bluestore


# ls -l /var/lib/ceph/osd/ceph-11 | grep block

> lrwxrwxrwx 1 ceph ceph  50 Mar 22 21:44 block ->
> /dev/mapper/YpB2cx-HlyU-VPqT-Abaz-Dutx-iMrz-Tty6o1


# lsblk

> nvme0n1                                                      259:0    0
> 931.5G  0 disk
> └─ceph--7ef8f83a--d055--4a59--8d6b--c564544c5a55-osd--block--c7c03dfd--a1b1--4182--9e61--bf264f293f2b
>                                                            253:0    0
> 931.5G  0 lvm
>   └─YpB2cx-HlyU-VPqT-Abaz-Dutx-iMrz-Tty6o1                   253:41   0
> 931.5G  0 crypt


# ceph osd dump | grep full_ratio

> full_ratio 0.97
> backfillfull_ratio 0.95
> nearfull_ratio 0.9


# ceph osd df | head
>
> ID   CLASS  WEIGHT    REWEIGHT  SIZE     RAW USE  DATA     OMAP      META     AVAIL     %USE   VAR   PGS  STATUS
>  11   nvme   1.29999   1.00000  932 GiB   694 GiB  319 GiB   301 GiB   74 GiB  238 GiB  74.49  1.09   96      up


Any ideas?
Thanks!
-- 

Rodrigo Werle
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux