Re: Libcephfs : ceph_readdirplus_r() with ceph_ll_lookup_vino() : ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



CC ceph-user
(Apologies, Forgot last time I replied)

On Tue, Sep 26, 2023 at 1:11 PM Joseph Fernandes <josephaug26@xxxxxxxxx>
wrote:

> Hello Venky,
>
> Did you get a chance to look into the updated program? Am I
> missing something?
> I suppose it's something trivial I am missing, As I see these API used in
> NFS Ganesha also and would have been tested there.
>
> https://github.com/nfs-ganesha/nfs-ganesha/blob/2a57b6d53295426247b200cd100ba0741b12aff9/src/FSAL/FSAL_CEPH/export.c#L322
>
> Thanks,
> ~Joe
>
> On Fri, Sep 22, 2023 at 10:52 PM Joseph Fernandes <josephaug26@xxxxxxxxx>
> wrote:
>
>> Hello Venky,
>>
>> Nice to hear from you :) Hope you are doing well.
>>
>> I tried as you suggested,
>>
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root# mkdir
>> dir1 dir2
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root# echo
>> "Hello Worldls!" > file2
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root# echo
>> "Hello Worldls!" > file1
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root# ls
>> dir1  dir2  file1  file2
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root#
>>
>> Create a new snapshot called  "sofs-4-6"
>>
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root#  ceph fs
>> subvolume snapshot ls cephfs 4
>> [
>>     {
>>         "name": "sofs-4-1"
>>     },
>>     {
>>         "name": "sofs-4-2"
>>     },
>>     {
>>         "name": "sofs-4-3"
>>     },
>>     {
>>         "name": "sofs-4-4"
>>     },
>>     {
>>         "name": "sofs-4-5"
>>     },
>>     {
>>         "name": "sofs-4-6"
>>     }
>> ]
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root#
>>
>> And Delete the file and directory from the active filessystem
>>
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root# rm -rf *
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root# ls
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root# cd .snap
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root/.snap# ls
>> _sofs-4-6_1099511627778
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root/.snap# cd
>> _sofs-4-6_1099511627778/
>> root@ss-joe-01(bash):/mnt/cephfs/volumes/_nogroup/4/user_root/.snap/_sofs-4-6_1099511627778#
>> ls
>> dir1  dir2  file1  file2
>> root@ss-joe-01
>> (bash):/mnt/cephfs/volumes/_nogroup/4/user_root/.snap/_sofs-4-6_1099511627778#
>>
>> Now I modified the program and executed it, I am getting the same result.
>>
>>  root@ss-joe-01(bash):/home/hydrauser# ./snapshot_inode_lookup
>>
>> =================================/volumes/_nogroup/4/user_root/=====================================
>>
>> Path/Name        :"/volumes/_nogroup/4/user_root/"
>> Inode Address    : 0x7f7c10008f00
>> Inode Number     : 1099511628293
>> Snapshot Number  : 18446744073709551614
>> Inode Number     : 1099511628293
>> Snapshot Number  : 18446744073709551614
>> . Ino: 1099511628293 SnapId: 18446744073709551614 Address: 0x7f7c10008f00
>> .. Ino: 1099511627778 SnapId: 18446744073709551614 Address: 0x7f7c100086f0
>>
>>
>> =================================1099511628293:7=====================================
>>
>> Path/Name        :"1099511628293:7"
>> Inode Address    : 0x7f7c10009710
>> Inode Number     : 1099511628293
>> Snapshot Number  : 7
>> Inode Number     : 1099511628293
>> Snapshot Number  : 7
>> . Ino: 1099511628293 SnapId: 7 Address: 0x7f7c10009710
>> .. Ino: 1099511628293 SnapId: 7 Address: 0x7f7c10009710
>>
>>
>> =================================/volumes/_nogroup/4/user_root/.snap/_sofs-4-6_1099511627778=====================================
>>
>> Path/Name
>>  :"/volumes/_nogroup/4/user_root/.snap/_sofs-4-6_1099511627778"
>> Inode Address    : 0x7f7c10009710
>> Inode Number     : 1099511628293
>> Snapshot Number  : 7
>> Inode Number     : 1099511628293
>> Snapshot Number  : 7
>> . Ino: 1099511628293 SnapId: 7 Address: 0x7f7c10009710
>> .. Ino: 1099511628293 SnapId: 18446744073709551615 Address: 0x55efc15b4640
>> file1 Ino: 1099511628297 SnapId: 7 Address: 0x7f7c1000a030
>> dir1 Ino: 1099511628294 SnapId: 7 Address: 0x7f7c1000a720
>> dir2 Ino: 1099511628295 SnapId: 7 Address: 0x7f7c1000ada0
>> file2 Ino: 1099511628296 SnapId: 7 Address: 0x7f7c1000b420
>>
>>
>> =================================1099511628293:7=====================================
>>
>> Path/Name        :"1099511628293:7"
>> Inode Address    : 0x7f7c10009710
>> Inode Number     : 1099511628293
>> Snapshot Number  : 7
>> Inode Number     : 1099511628293
>> Snapshot Number  : 7
>> . Ino: 1099511628293 SnapId: 7 Address: 0x7f7c10009710
>> .. Ino: 1099511628293 SnapId: 18446744073709551615 Address: 0x55efc15b4640
>> file1 Ino: 1099511628297 SnapId: 7 Address: 0x7f7c1000a030
>> dir1 Ino: 1099511628294 SnapId: 7 Address: 0x7f7c1000a720
>> dir2 Ino: 1099511628295 SnapId: 7 Address: 0x7f7c1000ada0
>> file2 Ino: 1099511628296 SnapId: 7 Address: 0x7f7c1000b420
>>
>> root@ss-joe-01(bash):/home/hydrauser#
>>
>> I have attached the modified program and ceph client logs from this run.
>>
>> Cheers,
>> Joe
>>
>> On Fri, Sep 22, 2023 at 8:54 PM Venky Shankar <vshankar@xxxxxxxxxx>
>> wrote:
>>
>>> Hi Joseph,
>>>
>>> On Fri, Sep 22, 2023 at 5:27 PM Joseph Fernandes <josephaug26@xxxxxxxxx>
>>> wrote:
>>> >
>>> > Hello All,
>>> >
>>> > I found a weird issue with ceph_readdirplus_r() when used along
>>> > with ceph_ll_lookup_vino().
>>> > On ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757)
>>> quincy
>>> > (stable)
>>> >
>>> > Any help is really appreciated.
>>> >
>>> > Thanks in advance,
>>> > -Joe
>>> >
>>> > Test Scenario :
>>> >
>>> > A. Create a Ceph Fs Subvolume "4" and created a directory in root of
>>> > subvolume "user_root"
>>> >
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23#
>>> > ceph fs subvolume ls cephfs
>>> > [
>>> >     {
>>> >         "name": "4"
>>> >     }
>>> > ]
>>> > root@ss-joe-01
>>> >
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23#
>>> >
>>> >
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23#
>>> > ls -l
>>> > total 0
>>> > drwxrwxrwx 2 root root 0 Sep 22 09:16 user_root
>>> > root@ss-joe-01
>>> >
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23#
>>> >
>>> >
>>> > B. In the "user_root" directory create some files and directories
>>> >
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root#
>>> > mkdir dir1 dir2
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root#
>>> > ls
>>> > dir1  dir2
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root#
>>> > echo
>>> > "Hello Worldls!" > file1
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root#
>>> > echo "Hello Worldls!" > file2
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root#
>>> > ls
>>> > dir1  dir2  file1  file2
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root#
>>> > cat file*
>>> > Hello Worldls!
>>> > Hello Worldls!
>>> >
>>> >
>>> > C. Create a subvolume snapshot   "sofs-4-5". Please ignore the older
>>> > snapshots.
>>> >
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23#
>>> > ceph fs subvolume snapshot ls cephfs 4
>>> > [
>>> >     {
>>> >         "name": "sofs-4-1"
>>> >     },
>>> >     {
>>> >         "name": "sofs-4-2"
>>> >     },
>>> >     {
>>> >         "name": "sofs-4-3"
>>> >     },
>>> >     {
>>> >         "name": "sofs-4-4"
>>> >     },
>>> >     {
>>> >         "name": "sofs-4-5"
>>> >     }
>>> > ]
>>> > root@ss-joe-01
>>> >
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23#
>>> >
>>> > Here "sofs-4-5" has snapshot id 6.
>>> > Got this from libcephfs and have verified at Line
>>> > snapshot_inode_lookup.cpp#L212. (Attached to the email)
>>> >
>>> >
>>> >            #Content within the snapshot
>>> >
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23#
>>> > cd .snap/
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/.snap#
>>> > ls
>>> > _sofs-4-1_1099511627778  _sofs-4-2_1099511627778
>>> _sofs-4-3_1099511627778
>>> >  _sofs-4-4_1099511627778  _sofs-4-5_1099511627778
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/.snap#
>>> > cd _sofs-4-5_1099511627778/
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/.snap/_sofs-4-5_1099511627778#
>>> > ls
>>> > user_root
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/.snap/_sofs-4-5_1099511627778#
>>> > cd user_root/
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/.snap/_sofs-4-5_1099511627778/user_root#
>>> > ls
>>> > dir1  dir2  file1  file2
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/.snap/_sofs-4-5_1099511627778/user_root#
>>> > cat file*
>>> > Hello Worldls!
>>> > Hello Worldls!
>>> > root@ss-joe-01
>>> >
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/.snap/_sofs-4-5_1099511627778/user_root#
>>> >
>>> >
>>> > D. Delete all the files and directories in "user_root"
>>> >
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root#
>>> > rm -rf *
>>> > root@ss-joe-01
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root#
>>> > ls
>>> > root@ss-joe-01
>>> >
>>> (bash):/mnt/cephfs/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root#
>>> >
>>> >
>>> > E. Using Libcephfs in a C++ program do the following,(Attached to this
>>> > email)
>>> >
>>> >    1. Get the Inode of "user_root" using ceph_ll_walk().
>>> >    2. Open the directory using Inode received from  ceph_ll_walk()
>>> and do
>>> >    ceph_readdirplus_r()
>>> >    We don't see any dentries(except "." and "..")  as we have deleted
>>> all
>>> >    files and directories in the active filesystem. This is expected and
>>> >    correct!
>>> >
>>> >
>>> >
>>> >
>>> =================================/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root/=====================================
>>> >
>>> >    Path/Name
>>> >
>>>  :"/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root/"
>>> >    Inode Address    : 0x7f5ce0009900
>>> >    Inode Number     : 1099511629282
>>> >    Snapshot Number  : 18446744073709551614
>>> >    Inode Number     : 1099511629282
>>> >    Snapshot Number  : 18446744073709551614
>>> >    . Ino: 1099511629282 SnapId: 18446744073709551614 Address:
>>> 0x7f5ce0009900
>>> >    .. Ino: 1099511627779 SnapId: 18446744073709551614 Address:
>>> >    0x7f5ce00090f0
>>> >
>>> >
>>> >    3. Using  ceph_ll_lookup_vino() get the Inode * of "user_root" for
>>> >    snapshot 6, Here "sofs-4-5" has snapshot id 6.
>>>
>>> Try this for path "/volumes/_nogroup/4/" (snapshots live a level-up in
>>> the subvolume path) for the ceph version you use.
>>>
>>> >    Got this from libcephfs and have verified at Line
>>> >    snapshot_inode_lookup.cpp#L212. (Attached to the email
>>> >    4. Open the directory using Inode * received from
>>> ceph_ll_lookup_vino()
>>> >    and do  ceph_readdirplus_r()
>>> >    We don't see any dentries (except "." and "..") This is NOT
>>> expected and
>>> >    NOT correct, as there are files and directories in the snaphot 6.
>>> >
>>> >
>>> >
>>> =================================1099511629282:6=====================================
>>> >
>>> >    Path/Name        :"1099511629282:6"
>>> >    Inode Address    : 0x7f5ce000a110
>>> >    Inode Number     : 1099511629282
>>> >    Snapshot Number  : 6
>>> >    Inode Number     : 1099511629282
>>> >    Snapshot Number  : 6
>>> >    . Ino: 1099511629282 SnapId: 6 Address: 0x7f5ce000a110
>>> >    .. Ino: 1099511629282 SnapId: 6 Address: 0x7f5ce000a110
>>> >
>>> >
>>> >    5. Get the Inode of "user_root/ .snap/_sofs-4-5_1099511627778 / "
>>> using
>>> >    ceph_ll_walk().
>>> >    6. Open the directory using Inode received from  ceph_ll_walk()
>>> and do
>>> >    ceph_readdirplus_r()
>>> >    We see ALL dentries of all files and directories in the snapshot.
>>> This
>>> >    is expected and correct!
>>> >
>>> >
>>> >
>>> =================================/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root/.snap/_sofs-4-5_1099511627778/=====================================
>>> >
>>> >    Path/Name
>>> >
>>>  :"/volumes/_nogroup/4/f0fae76f-196d-4ebd-b8d0-528985505b23/user_root/.snap/_sofs-4-5_1099511627778/"
>>> >    Inode Address    : 0x7f5ce000a110
>>> >    Inode Number     : 1099511629282
>>> >    Snapshot Number  : 6
>>> >    Inode Number     : 1099511629282
>>> >    Snapshot Number  : 6
>>> >    . Ino: 1099511629282 SnapId: 6 Address: 0x7f5ce000a110
>>> >    .. Ino: 1099511629282 SnapId: 18446744073709551615 Address:
>>> >    0x5630ab946340
>>> >    file1 Ino: 1099511628291 SnapId: 6 Address: 0x7f5ce000aa90
>>> >    dir1 Ino: 1099511628289 SnapId: 6 Address: 0x7f5ce000b180
>>> >    dir2 Ino: 1099511628290 SnapId: 6 Address: 0x7f5ce000b800
>>> >    file2 Ino: 1099511628292 SnapId: 6 Address: 0x7f5ce000be80
>>> >
>>> >    7. Now Again using  ceph_ll_lookup_vino() get the Inode * of
>>> "user_root"
>>> >    for snapshot 6, Here "sofs-4-5" has snapshot id 6.
>>> >    8.  Open the directory using Inode * received from
>>> >    ceph_ll_lookup_vino() and do  ceph_readdirplus_r()
>>> >    Now! we see all the files and Directories in the snapshot!
>>> >
>>> >
>>> >
>>> =================================1099511629282:6=====================================
>>> >
>>> >    Path/Name        :"1099511629282:6"
>>> >    Inode Address    : 0x7f5ce000a110
>>> >    Inode Number     : 1099511629282
>>> >    Snapshot Number  : 6
>>> >    Inode Number     : 1099511629282
>>> >    Snapshot Number  : 6
>>> >    . Ino: 1099511629282 SnapId: 6 Address: 0x7f5ce000a110
>>> >    .. Ino: 1099511629282 SnapId: 18446744073709551615 Address:
>>> >    0x5630ab946340
>>> >    file1 Ino: 1099511628291 SnapId: 6 Address: 0x7f5ce000aa90
>>> >    dir1 Ino: 1099511628289 SnapId: 6 Address: 0x7f5ce000b180
>>> >    dir2 Ino: 1099511628290 SnapId: 6 Address: 0x7f5ce000b800
>>> >    file2 Ino: 1099511628292 SnapId: 6 Address: 0x7f5ce000be80
>>> >
>>> >
>>> > Am I missing something using these APIs?
>>> >
>>> > File attached to this email
>>> > Full out of the program attached to the email.
>>> > - snapshot_inode_lookup.cpp_output.txt <Attached>
>>> > C++ Program - snapshot_inode_lookup.cpp  <Attached>
>>> > /etc/ceph/ceph.conf - <attached>
>>> > Ceph Client Log during the run of this C++ program -
>>> client.log<attached>
>>> >
>>> > Compile Command:
>>> > g++ -o snapshot_inode_lookup  ./snapshot_inode_lookup.cpp -g -ldl -ldw
>>> > -lcephfs -lboost_filesystem --std=c++17
>>> >
>>> > Linux Details,
>>> > root@ss-joe-01(bash):/home/hydrauser# uname -a
>>> > Linux ss-joe-01 5.10.0-23-amd64 #1 SMP Debian 5.10.179-1 (2023-05-12)
>>> > x86_64 GNU/Linux
>>> > root@ss-joe-01(bash):/home/hydrauser#
>>> >
>>> > Ceph Details,
>>> >
>>> > root@ss-joe-01(bash):/home/hydrauser# ceph -v
>>> > ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy
>>> > (stable)
>>> > root@ss-joe-01(bash):/home/hydrauser#
>>> > root@ss-joe-01(bash):/home/hydrauser# ceph -s
>>> >   cluster:
>>> >     id:     fb43d857-d165-4189-87fc-cf1debce9170
>>> >     health: HEALTH_OK
>>> >
>>> >   services:
>>> >     mon: 3 daemons, quorum ss-joe-01,ss-joe-02,ss-joe-03 (age 4d)
>>> >     mgr: ss-joe-01(active, since 4d), standbys: ss-joe-03, ss-joe-02
>>> >     mds: 1/1 daemons up
>>> >     osd: 3 osds: 3 up (since 4d), 3 in (since 4d)
>>> >
>>> >   data:
>>> >     volumes: 1/1 healthy
>>> >     pools:   3 pools, 49 pgs
>>> >     objects: 39 objects, 1.0 MiB
>>> >     usage:   96 MiB used, 30 GiB / 30 GiB avail
>>> >     pgs:     49 active+clean
>>> >
>>> > root@ss-joe-01(bash):/home/hydrauser#
>>> > root@ss-joe-01(bash):/home/hydrauser# dpkg -l | grep ceph
>>> > ii  ceph                                               17.2.5-1~bpo11+1
>>> >           amd64        distributed storage and file system
>>> > ii  ceph-base                                          17.2.5-1~bpo11+1
>>> >           amd64        common ceph daemon libraries and management
>>> tools
>>> > ii  ceph-base-dbg                                      17.2.5-1~bpo11+1
>>> >           amd64        debugging symbols for ceph-base
>>> > ii  ceph-common                                        17.2.5-1~bpo11+1
>>> >           amd64        common utilities to mount and interact with a
>>> ceph
>>> > storage cluster
>>> > ii  ceph-common-dbg                                    17.2.5-1~bpo11+1
>>> >           amd64        debugging symbols for ceph-common
>>> > ii  ceph-fuse                                          17.2.5-1~bpo11+1
>>> >           amd64        FUSE-based client for the Ceph distributed file
>>> > system
>>> > ii  ceph-fuse-dbg                                      17.2.5-1~bpo11+1
>>> >           amd64        debugging symbols for ceph-fuse
>>> > ii  ceph-mds                                           17.2.5-1~bpo11+1
>>> >           amd64        metadata server for the ceph distributed file
>>> system
>>> > ii  ceph-mds-dbg                                       17.2.5-1~bpo11+1
>>> >           amd64        debugging symbols for ceph-mds
>>> > ii  ceph-mgr                                           17.2.5-1~bpo11+1
>>> >           amd64        manager for the ceph distributed storage system
>>> > ii  ceph-mgr-cephadm                                   17.2.5-1~bpo11+1
>>> >           all          cephadm orchestrator module for ceph-mgr
>>> > ii  ceph-mgr-dashboard                                 17.2.5-1~bpo11+1
>>> >           all          dashboard module for ceph-mgr
>>> > ii  ceph-mgr-dbg                                       17.2.5-1~bpo11+1
>>> >           amd64        debugging symbols for ceph-mgr
>>> > ii  ceph-mgr-diskprediction-local                      17.2.5-1~bpo11+1
>>> >           all          diskprediction-local module for ceph-mgr
>>> > ii  ceph-mgr-k8sevents                                 17.2.5-1~bpo11+1
>>> >           all          kubernetes events module for ceph-mgr
>>> > ii  ceph-mgr-modules-core                              17.2.5-1~bpo11+1
>>> >           all          ceph manager modules which are always enabled
>>> > ii  ceph-mon                                           17.2.5-1~bpo11+1
>>> >           amd64        monitor server for the ceph storage system
>>> > ii  ceph-mon-dbg                                       17.2.5-1~bpo11+1
>>> >           amd64        debugging symbols for ceph-mon
>>> > ii  ceph-osd                                           17.2.5-1~bpo11+1
>>> >           amd64        OSD server for the ceph storage system
>>> > ii  ceph-osd-dbg                                       17.2.5-1~bpo11+1
>>> >           amd64        debugging symbols for ceph-osd
>>> > ii  ceph-volume                                        17.2.5-1~bpo11+1
>>> >           all          tool to facilidate OSD deployment
>>> > ii  cephadm                                            17.2.5-1~bpo11+1
>>> >           amd64        cephadm utility to bootstrap ceph daemons with
>>> > systemd and containers
>>> > ii  libcephfs2                                         17.2.5-1~bpo11+1
>>> >           amd64        Ceph distributed file system client library
>>> > ii  libcephfs2-dbg                                     17.2.5-1~bpo11+1
>>> >           amd64        debugging symbols for libcephfs2
>>> > ii  libsqlite3-mod-ceph                                17.2.5-1~bpo11+1
>>> >           amd64        SQLite3 VFS for Ceph
>>> > ii  libsqlite3-mod-ceph-dbg                            17.2.5-1~bpo11+1
>>> >           amd64        debugging symbols for libsqlite3-mod-ceph
>>> > ii  python3-ceph-argparse                              17.2.5-1~bpo11+1
>>> >           all          Python 3 utility libraries for Ceph CLI
>>> > ii  python3-ceph-common                                17.2.5-1~bpo11+1
>>> >           all          Python 3 utility libraries for Ceph
>>> > ii  python3-cephfs                                     17.2.5-1~bpo11+1
>>> >           amd64        Python 3 libraries for the Ceph libcephfs
>>> library
>>> > ii  python3-cephfs-dbg                                 17.2.5-1~bpo11+1
>>> >           amd64        Python 3 libraries for the Ceph libcephfs
>>> library
>>> > root@ss-joe-01(bash):/home/hydrauser#
>>> > _______________________________________________
>>> > ceph-users mailing list -- ceph-users@xxxxxxx
>>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>>
>>>
>>>
>>> --
>>> Cheers,
>>> Venky
>>>
>>>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux