Re: Incredibly poor performance of mdraid-1 with 2 SSD Samsung 840 PRO

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

1. free -m
root [~]# free -m
             total       used       free     shared    buffers cached
Mem:         15921      15542        379          0 1063      11870
-/+ buffers/cache:       2608      13313
Swap:         2046        100       1946

2. Yes, you understood correctly regarding the raid array (all 3 of them are raid 1):

root@gts6 [~]# cat /proc/mdstat
Personalities : [raid1]
md0 : active raid1 sdb2[1] sda2[0]
      204736 blocks super 1.0 [2/2] [UU]

md2 : active raid1 sdb3[1] sda3[0]
      404750144 blocks super 1.0 [2/2] [UU]

md1 : active raid1 sdb1[1] sda1[0]
      2096064 blocks super 1.1 [2/2] [UU]

unused devices: <none>

md0 is boot.
md1 is swap.
md2 is /

3. df

root@gts6 [~]# df -h
Filesystem            Size  Used Avail Use% Mounted on
/dev/md2              380G  246G  116G  68% /
tmpfs                 7.8G     0  7.8G   0% /dev/shm
/dev/md0              194M   47M  137M  26% /boot
/usr/tmpDSK           3.6G  1.2G  2.2G  36% /tmp

4. pvs

root [~]# pvs -a
  PV         VG   Fmt Attr PSize PFree
  /dev/loop0          ---     0     0
  /dev/md0            ---     0     0
  /dev/md1            ---     0     0
  /dev/ram0           ---     0     0
  /dev/ram1           ---     0     0
  /dev/ram10          ---     0     0
  /dev/ram11          ---     0     0
  /dev/ram12          ---     0     0
  /dev/ram13          ---     0     0
  /dev/ram14          ---     0     0
  /dev/ram15          ---     0     0
  /dev/ram2           ---     0     0
  /dev/ram3           ---     0     0
  /dev/ram4           ---     0     0
  /dev/ram5           ---     0     0
  /dev/ram6           ---     0     0
  /dev/ram7           ---     0     0
  /dev/ram8           ---     0     0
  /dev/ram9           ---     0     0
  /dev/root           ---     0     0

5. lvs (No volume groups).

Thanks!

On 24/04/2013 12:12 PM, Adam Goryachev wrote:
On 24/04/13 18:26, Andrei Banu wrote:
Hello,

I am sorry for the irrelevant feedback. Where I misunderstood your
request, I filled in the blanks (poorly).

1. SWAP
root [~]# blkid | grep cef1d19d-2578-43db-9ffc-b6b70e227bfa
/dev/md1: UUID="cef1d19d-2578-43db-9ffc-b6b70e227bfa" TYPE="swap"

So yes, swap is on md1. This *md1 has a size of 2GB*. Isn't this way
too low for a system with 16GB of memory?

Provide the output of "free", if there is RAM available, then it isn't
too small (that is my personal opinion, but at least it won't affect
performance/operations until you are using most of that swap space).

3. root [~]# fdisk -lu /dev/sd*

My mistake, I should have said:
fdisk -lu /dev/sd?

In any case, all of the relevant information was included, so no harm done.
Disk /dev/sda: 512.1 GB, 512110190592 bytes
255 heads, 63 sectors/track, 62260 cylinders, total 1000215216 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00026d59

    Device Boot      Start         End      Blocks   Id  System
/dev/sda1            2048     4196351     2097152   fd  Linux raid
autodetect
Partition 1 does not end on cylinder boundary.
/dev/sda2   *     4196352     4605951      204800   fd  Linux raid
autodetect
Partition 2 does not end on cylinder boundary.
/dev/sda3         4605952   814106623   404750336   fd  Linux raid
autodetect

Disk /dev/sdb: 512.1 GB, 512110190592 bytes
255 heads, 63 sectors/track, 62260 cylinders, total 1000215216 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x0003dede

    Device Boot      Start         End      Blocks   Id  System
/dev/sdb1            2048     4196351     2097152   fd  Linux raid
autodetect
Partition 1 does not end on cylinder boundary.
/dev/sdb2   *     4196352     4605951      204800   fd  Linux raid
autodetect
Partition 2 does not end on cylinder boundary.
/dev/sdb3         4605952   814106623   404750336   fd  Linux raid
autodetect

I'm assuming from this you have three md RAID1 arrays where sda1/sdb1
are a pair, sda2/sdb2 are a pair and sda3/sdb3 are a pair?

Can you describe what is on each of these arrays?
Output of
cat /proc/mdstat
df
pvs
lvs

Might be helpful....

Regards,
Adam


--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux