I checked the machine.
Here is the df -hT output
[jenkins@slave29 ~]$ cat /etc/fstab
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
/dev/xvda1 / ext3 defaults,noatime,barrier=0 1 1
tmpfs /dev/shm tmpfs defaults 0 0
devpts /dev/pts devpts gid=5,mode=620 0 0
sysfs /sys sysfs defaults 0 0
proc /proc proc defaults 0 0
#/dev/xvdc1 none swap sw 0 0
We don't see a xfs device mounted at /d and / is of type ext3 which does not support fallocate. The uptime of the machine is 73 days though. I don't know how the /d xfs partition vanished.
On Tue, Jul 12, 2016 at 4:54 PM, Atin Mukherjee <amukherj@xxxxxxxxxx> wrote:
--On Tue, Jul 12, 2016 at 4:42 PM, Atin Mukherjee <amukherj@xxxxxxxxxx> wrote:On Tue, Jul 12, 2016 at 4:36 PM, Avra Sengupta <asengupt@xxxxxxxxxx> wrote:Hi Atin,
Please check the testcase result in the console. It clearly states the reason of the failure. A quick search of 30815, as shown in the testcase shows that the error that is generated is a thinp issue, and we can see fallocate failing and lvm not properly being setup in the environment.While this is valid for my docker containers, I am just wondering why did this happen in jenkins slave?
Regards,
Avra
P.S Here are the logs from the console stating so.
02:50:34 [09:50:34] Running tests in file ./tests/basic/op_errnos.t 02:50:41 fallocate: /d/backends/patchy_snap_vhd: fallocate failed: Operation not supported 02:50:41 losetup: /d/backends/patchy_snap_vhd: warning: file smaller than 512 bytes, the loop device maybe be useless or invisible for system tools. 02:50:41 Device /d/backends/patchy_snap_loop not found (or ignored by filtering). 02:50:41 Device /d/backends/patchy_snap_loop not found (or ignored by filtering). 02:50:41 Unable to add physical volume '/d/backends/patchy_snap_loop' to volume group 'patchy_snap_vg_1'. 02:50:41 Volume group "patchy_snap_vg_1" not found 02:50:41 Cannot process volume group patchy_snap_vg_1 02:50:42 Volume group "patchy_snap_vg_1" not found 02:50:42 Cannot process volume group patchy_snap_vg_1 02:50:42 /dev/patchy_snap_vg_1/brick_lvm: No such file or directory 02:50:42 Usage: mkfs.xfs 02:50:42 /* blocksize */ [-b log=n|size=num] 02:50:42 /* data subvol */ [-d agcount=n,agsize=n,file,name=xxx,size=num, 02:50:42 (sunit=value,swidth=value|su=num,sw=num), 02:50:42 sectlog=n|sectsize=num 02:50:42 /* inode size */ [-i log=n|perblock=n|size=num,maxpct=n,attr=0|1|2, 02:50:42 projid32bit=0|1] 02:50:42 /* log subvol */ [-l agnum=n,internal,size=num,logdev=xxx,version=n 02:50:42 sunit=value|su=num,sectlog=n|sectsize=num, 02:50:42 lazy-count=0|1] 02:50:42 /* label */ [-L label (maximum 12 characters)] 02:50:42 /* naming */ [-n log=n|size=num,version=2|ci] 02:50:42 /* prototype file */ [-p fname] 02:50:42 /* quiet */ [-q] 02:50:42 /* realtime subvol */ [-r extsize=num,size=num,rtdev=xxx] 02:50:42 /* sectorsize */ [-s log=n|size=num] 02:50:42 /* version */ [-V] 02:50:42 devicename 02:50:42 <devicename> is required unless -d name=xxx is given. 02:50:42 <num> is xxx (bytes), xxxs (sectors), xxxb (fs blocks), xxxk (xxx KiB), 02:50:42 xxxm (xxx MiB), xxxg (xxx GiB), xxxt (xxx TiB) or xxxp (xxx PiB). 02:50:42 <value> is xxx (512 byte blocks). 02:50:42 mount: special device /dev/patchy_snap_vg_1/brick_lvm does not exist 02:50:53 ./tests/basic/op_errnos.t .. 02:50:53 1..21 02:50:53 ok 1, LINENUM:12 02:50:53 ok 2, LINENUM:13 02:50:53 ok 3, LINENUM:14 02:50:53 ok 4, LINENUM:16 02:50:53 ok 5, LINENUM:18 02:50:53 ok 6, LINENUM:19 02:50:53 ok 7, LINENUM:20On 07/12/2016 03:47 PM, Atin Mukherjee wrote:
Hi Avra,
The above fails locally as well along with few regression failures I observed and one of them are at [1]
not ok 12 Got " 30807" instead of "30809", LINENUM:26
FAILED COMMAND: 30809 get-op_errno-xml snapshot restore snap1
not ok 17 Got " 30815" instead of "30812", LINENUM:31
FAILED COMMAND: 30812 get-op_errno-xml snapshot create snap1 patchy no-timestamp
[1] https://build.gluster.org/job/rackspace-regression-2GB-triggered/22154/console
--Atin
----Atin--Atin
_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel
_______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-devel