Ceph as a Fileserver for 3D Content Production

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Martin,

The Samsung Evo is just my system drive on my test host. This is my server setup

3 Nodes: (original those where render workstations that are not in use right now)

Each Node is MON MGM OSD

Mainboard: ASRock TRX40 Creator
CPU: AMD Ryzen Threadripper 3960X, 24 Cores, 3.8Ghz
RAM: 2 x Samsung 32 GB 2 x 8 DDR4 2666 MHz 288-pin DIMM, Unregistered, ECC (64 GB Total) NIC Public: OnBoard Aquantia 107, 10Gbit NIC Ceph:
Intel XXV710-DA2, 2x SFP28, 25Gbit System Drive: 2x Samsung SSD 860 PRO 256GB, SATA, ZFS Raid 1
System: Proxmox VE 6.2, Debian Buster, Ceph Nautilus
HBA: Broadcom SAS 9305-16i

OSDs:
6x Seagate Exos, 16TB, 7.200 rpm, 12Gb SAS

Cache:
1x Micron 9300 MAX 3.2TB U.2 NVME



__________________________________

EventArt GmbH
Kirchenstraße 13,
3243 St. Leonhard am Forst
mobile: +43(0)6641951196
mail: office@xxxxxxxxxxx<mailto:office@xxxxxxxxxxx>
UID: ATU 64417269

http://www.eventart.at II http://www.clamlive.at

Von: Martin Verges <martin.verges@xxxxxxxx>
Gesendet: Sonntag, 17. Mai 2020 18:28
An: Moritz Wilhelm <moritz@xxxxxxxxxxx>
Cc: Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx>; ceph-users <ceph-users@xxxxxxx>
Betreff: Re:  Re: Ceph as a Fileserver for 3D Content Production

Hello Moritz,

drop the EVO disk and use a SSD that is working with Ceph. For example just use PM883 / PM983 from the same vendor and you will have a huge performance increase.

--
Martin Verges
Managing director

Mobile: +49 174 9335695
E-Mail: martin.verges@xxxxxxxx<mailto:martin.verges@xxxxxxxx>
Chat: https://t.me/MartinVerges

croit GmbH, Freseniusstr. 31h, 81247 Munich
CEO: Martin Verges - VAT-ID: DE310638492
Com. register: Amtsgericht Munich HRB 231263

Web: https://croit.io
YouTube: https://goo.gl/PGE1Bx


Am So., 17. Mai 2020 um 15:12 Uhr schrieb Moritz Wilhelm <moritz@xxxxxxxxxxx<mailto:moritz@xxxxxxxxxxx>>:
Hi Marc,

thank you very much for your feedback, actually that is what I am looking for (Design advices and feedback). I also wanted to get in touch with the community because right now I am on my own with this project with no experience at all. But I also wanted to get into it first and learn the basics and setup a system before bothering other people with it and to be able to even hold a conversation about it.

So we are a 3D Content Production Company with around 10-12 Workstations and 6 Render stations. We work with software like Autodesk Maya or Adobe After effects. All our Projects are stored on the NAS and we work directly with those files on the NAS.

This is our current System:

Synology RS4017xs+
Intel Xeon D-1541 2,1 Ghz, 8 Cores
32GB Ram
16x 8TB WD Red Drives
2  TB SSD Cache
2 x 10 Gb SFP+ Nics


About the Bcache I just did tests with it as I found the idea very interesting and also performance was better. If this project is ever getting into real-life production I will stick to the most common setups for sure.

Here are some Results of Diskmark on one Computer. (I Had only 5 OSDs per Node up and running at this test because I am still waiting for the new Seagate drives) For Sequential Reads and Writes I sometimes maxed out the 10Gbe host connection so the data is not very useful but the other numbers are quite interesting:

For Comparison:

Hosts Local NVME Drive (Samsung 970 EVO Pro 1TB):
        Read:                                                                                   Write:
        Sequential 1MiB (Q=  8, T= 1):  3507.700 MB/s [   3345.2 IOPS] <  2389.61 us>           Sequential 1MiB (Q=  8, T= 1):  2548.063 MB/s [   2430.0 IOPS] <  3284.79 us>
        Sequential 1MiB (Q=  1, T= 1):  2368.308 MB/s [   2258.6 IOPS] <   442.07 us>           Sequential 1MiB (Q=  1, T= 1):  2456.471 MB/s [   2342.7 IOPS] <   426.10 us>
        Random 4KiB (Q= 32, T=16):  1555.565 MB/s [ 379776.6 IOPS] <  1312.45 us>                Random 4KiB (Q= 32, T=16):    16.580 MB/s [   4047.9 IOPS] <124267.88 us>
        Random 4KiB (Q=  1, T= 1):    51.666 MB/s [  12613.8 IOPS] <    78.82 us>               Random 4KiB (Q=  1, T= 1):   108.983 MB/s [  26607.2 IOPS] <    37.13 us>

Current Synology NAS (SMB):
        Read:                                                                                   Write:
        Sequential 1MiB (Q=  8, T= 1):  1045.594 MB/s [    997.2 IOPS] <  7990.33 us>           Sequential 1MiB (Q=  8, T= 1):  1101.007 MB/s [   1050.0 IOPS] <  7588.25 us>
        Sequential 1MiB (Q=  1, T= 1):   953.709 MB/s [    909.5 IOPS] <  1098.57 us>           Sequential 1MiB (Q=  1, T= 1):   847.847 MB/s [    808.6 IOPS] <  1235.26 us>
        Random 4KiB (Q= 32, T=16):     4.198 MB/s [   1024.9 IOPS] <380158.65 us>               Random 4KiB (Q= 32, T=16):   188.827 MB/s [  46100.3 IOPS] < 11076.14 us>
        Random 4KiB (Q=  1, T= 1):     2.486 MB/s [    606.9 IOPS] <  1637.08 us>                       Random 4KiB (Q=  1, T= 1):     7.177 MB/s [   1752.2 IOPS] <   570.16 us>

Ceph With WAL/DB on NVME total of 5 old SATA HDD OSDs (SMB):
        Read:                                                                                   Write:
        Sequential 1MiB (Q=  8, T= 1):   534.050 MB/s [    509.3 IOPS] < 15628.63 us>           Sequential 1MiB (Q=  8, T= 1):   198.420 MB/s [    189.2 IOPS] < 42020.67 us>
        Sequential 1MiB (Q=  1, T= 1):   340.580 MB/s [    324.8 IOPS] <  2921.17 us>           Sequential 1MiB (Q=  1, T= 1):   184.329 MB/s [    175.8 IOPS] <  5603.99 us>
        Random 4KiB (Q= 32, T=16):     3.172 MB/s [    774.4 IOPS] <398622.73 us>               Random 4KiB (Q= 32, T=16):     8.639 MB/s [   2109.1 IOPS] <222699.43 us>
        Random 4KiB (Q=  1, T= 1):     1.907 MB/s [    465.6 IOPS] <  2139.08 us>                       Random 4KiB (Q=  1, T= 1):     7.294 MB/s [   1780.8 IOPS] <   560.91 us>

Ceph With Bcache total of 5 old SATA HDD OSDs (SMB):
        Sequential 1MiB (Q=  8, T= 1):   967.386 MB/s [    922.6 IOPS] <  8660.59 us>           Sequential 1MiB (Q=  8, T= 1):   882.882 MB/s [    842.0 IOPS] <  9444.23 us>
        Sequential 1MiB (Q=  1, T= 1):   418.157 MB/s [    398.8 IOPS] <  2448.00 us>           Sequential 1MiB (Q=  1, T= 1):   793.417 MB/s [    756.7 IOPS] <  1320.37 us>
        Random 4KiB (Q= 32, T=16):    39.756 MB/s [   9706.1 IOPS] < 51788.10 us>               Random 4KiB (Q= 32, T=16):   128.303 MB/s [  31324.0 IOPS] < 16298.94 us>
        Random 4KiB (Q=  1, T= 1):    29.527 MB/s [   7208.7 IOPS] <   138.36 us>               Random 4KiB (Q=  1, T= 1):    32.334 MB/s [   7894.0 IOPS] <   126.25 us>

I will do an Update as soon as I have all the 18 OSDs running.

I could also need some advice on how to do proper testing.





__________________________________

EventArt GmbH
Kirchenstraße 13,
3243 St. Leonhard am Forst
mobile: +43(0)6641951196
mail: office@xxxxxxxxxxx<mailto:office@xxxxxxxxxxx>
UID: ATU 64417269

http://www.eventart.at II http://www.clamlive.at

-----Ursprüngliche Nachricht-----
Von: Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx<mailto:M.Roos@xxxxxxxxxxxxxxxxx>>
Gesendet: Sonntag, 17. Mai 2020 14:33
An: ceph-users <ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx>>; Moritz Wilhelm <moritz@xxxxxxxxxxx<mailto:moritz@xxxxxxxxxxx>>
Betreff: RE:  Ceph as a Fileserver for 3D Content Production



I was just reading your post, and started wondering why you posted it. I do not see clear question, and you also do not share test results (from your nas vs cephfs smb). So maybe you like some attention in this covid social distancing time? ;)

Anyway, I have been 'testing' with ceph for 2,5 years or so. (I am telling you so, you know how to value my info) If you start playing with ceph keep it simple and stick to what is recommended. Do not start looking for unconventional ways making ceph faster, because you are not able to do it better than the developers. You do not know the ins and outs, and you are more likely to shoot yourself in the foot. At least ask first. Eg this bcache, I am not 100% sure what it is, but if it is sitting between the osd process and the disk, it could be getting nasty with a reset/power outage, when ceph thinks data is written to disk, while it is not.

I would start by identifying what your minimum performance requirements are, maybe post them, and ask if someone has realized them with a ceph setup.




-----Original Message-----
Sent: 17 May 2020 13:42
To: ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx>
Subject:  Ceph as a Fileserver for 3D Content Production

Hi,

my Name is Moritz and I am working for a 3D production company. Because of the corona virus I have too much time left and also to much unused hardware. That is why I started playing around with Ceph as a fileserver for us. Here I want to share my experience for all those who are interested. To start of here is my actual running test system. I am interested in the thoughts of the community and also on more suggestions on what to try out with my available Hardware. I don’t know how to test it right now because I am a newbie to ceph and our production file server is a super user-friendly but high performance Synology NAS 😉.
All I have done so far was running Crystal disk benchmark on 1 Windows machine on the SMB Share.

3 Nodes: (original those where render workstations that are not in use right now)

Each Node is MON MGM OSD

Mainboard: ASRock TRX40 Creator
CPU: AMD Ryzen Threadripper 3960X, 24 Cores, 3.8Ghz
RAM: 2 x Samsung 32 GB 2 x 8 DDR4 2666 MHz 288-pin DIMM, Unregistered, ECC (64 GB Total) NIC Public: OnBoard Aquantia 107, 10Gbit NIC Ceph:
Intel XXV710-DA2, 2x SFP28, 25Gbit System Drive: 2x Samsung SSD 860 PRO 256GB, SATA, ZFS Raid 1
System: Proxmox VE 6.2, Debian Buster, Ceph Nautilus
HBA: Broadcom SAS 9305-16i

OSDs:
6x Seagate Exos, 16TB, 7.200 rpm, 12Gb SAS

Cache:
1x Micron 9300 MAX 3.2TB U.2 NVME

I Played around with setting it up as a WAL/DB Device. Right now I have configured the Micron NVME as a BCache Infront of the 6 Seagate Drives in writeback mode.
Because in this configuration BCache takes care of translating random writes to sequential ones for the HDDs I turned the Ceph WAL LOG off. I think Bcache gives more options to tune the System for my use case instead of just putting WAL/DB on the NVME. And also I can easily add cache drives or remove them without touching osds.

I set up SMB Shares with the vfs_ceph module. I still have to add CTDB to distribute Samba to all nodes.

My Next steps are to keep playing around in tuning the system and testing stability and performance. After that I want to put the Ceph cluster infront of our production NAS. Because our data is not super critical I thought of setting the replicas to 2 and running Rsync overnight to our NAS. That way I can switch to the old NAS at any time and wouldn’t loose more than 1 Day of work which is acceptable for us.
This is how I could compare the two solutions side by side with real-life workload.

I know that ceph might not be the best solution right now but if I am able to get at least similar performance to our Synology HDD NAS out of it, it would give a super scalable Solution in size and performance to grow with our needs. And who knows what performance improvements we get with ceph in the next 3 years.

I am happy to hear your thoughts and ideas. And please I know this might be kind of a crazy setup but I have fun with it and I learned a lot the last few weeks. If my experiment fails I will go back to my original
plan: Put FreeNas on two of the Nodes with overnight replication and put the third Node back to his render-friends. 😃

By the way I also have a spare Dell Server: 2x Xeon E5-2630 v3 2,40GHz, 128G Ram. I just don’t have an idea on how to utilize it. Maybe as extra OSD Node or as a separate Samba Server to get the SMB traffic away from the Public Ceph Network.




Moritz Wilhelm


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> To unsubscribe send an email to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx>
To unsubscribe send an email to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux