Hi All, Please let me know if anyone has any suggestion on this issue. I am stuck on these errors and really looking forward for your help. Thanks, LP On Thu, Jun 14, 2018 at 11:08 PM, learn pvt <learnpvt2050@xxxxxxxxx> wrote: > Hi All, > > I am getting same errors while running fio using libhdfs on > open-source hadoop-2.6.0. > > I also installed the 'hadoop-2.6.0-src' package and built it using the > Maven (mvn compile) > > After compiling, I got the following executables created under > /hadoop/sources/hadoop-2.6.0-src/hadoop-hdfs-project/hadoop-hdfs/target/native/: > test_libhdfs_ops test_libhdfs_threaded test_libhdfs_write > test_native_mini_dfs test_libhdfs_read test_libhdfs_vecsum > test_libhdfs_zerocopy > > Using these binaries I was able to write/read files in my hadoop-cluster. > /usr/local/hadoop/sources/hadoop-2.6.0-src/hadoop-hdfs-project/hadoop-hdfs/target/native$./test_libhdfs_write > test-file 100M 10M > > But if I try to run FIO using 'ioengine=libhdfs' then it fails with > the same errors mentioned in earlier email. > > As executables are able to utilize the 'libhdfs.so' correctly, it > seems like FIO is not utilizing it as it should. > > I have configured the classpaths and other required settings as > mentioned in earlier email. > > Please let me know if some additional configurations are required to > run FIO using libhdfs engine. > > Thanks, > LP > > On Tue, Jun 5, 2018 at 1:09 AM, Sitsofe Wheeler <sitsofe@xxxxxxxxx> wrote: >> CC'ing Fabrice and Manish... >> >> On 4 June 2018 at 19:36, learn pvt <learnpvt2050@xxxxxxxxx> wrote: >>> Hi, >>> >>> I am trying to configure the 'libhdfs' in my CHD5.14.0 cluster but >>> getting some issues. >>> >>> I have two node CDH-cluster and its working fine as I checked with TestDFSIO. >>> I set the FIO_LIBHDFS_INCLUDE, FIO_LIBHDFS_LIB, 'hadoop classpath' in >>> environment. >>> >>> Installed FIO on Master and configured it with enabling libhdfs while >>> 'configure'. >>> >>> Created hdfs.fio file as >>> [global] >>> runtime=10 >>> [hdfs] >>> rw=read >>> namenode=xx.xx.x.xxx >>> hostname=server1 >>> port=9000 >>> hdfsdirectory=/tmp >>> chunk_size=1m >>> hdfs_use_direct=1 >>> ioengine=libhdfs >>> size=20m >>> bs=256k >>> single_instance=1 >>> >>> Ran the fio job(sudo fio hdfs.fio) which resulted into errors as >>> >>> fio-3.6-33-g540e >>> Starting 1 process >>> loadFileSystems error >>> hdfsBuilderConnect forceNewInstance=0, nn=server1, port=9000, >>> kerbTicketCachePath=(NULL), userName=(NULL) error >>> hdfsExists: constructNewObjectOfPath error >>> hdfs: invalid working directory /tmp: Unknown error 255 >>> fio: failed to init engine data: 255 >>> Run status group 0 (all jobs) >>> >>> I tried many things but could not configure libhdfs to work correctly >>> with FIO on my CDH. Any suggestion would be appreciated! >>> Thanks in advance. >>> Note: /tmp directory is present at the hdfs level. >>> >>> Thanks, >>> LP >> >> -- >> Sitsofe | http://sucs.org/~sits/ -- To unsubscribe from this list: send the line "unsubscribe fio" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html