Re: fio libhdfs - loadFileSystems error while running FIO in CDH5.14.0

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi All,

Please let me know if anyone has any suggestion on this issue. I am
stuck on these errors and really looking forward for your help.

Thanks,
LP

On Thu, Jun 14, 2018 at 11:08 PM, learn pvt <learnpvt2050@xxxxxxxxx> wrote:
> Hi All,
>
> I am getting same errors while running fio using libhdfs on
> open-source hadoop-2.6.0.
>
> I also installed the 'hadoop-2.6.0-src' package and built it using the
> Maven (mvn compile)
>
> After compiling, I got the following executables created under
> /hadoop/sources/hadoop-2.6.0-src/hadoop-hdfs-project/hadoop-hdfs/target/native/:
> test_libhdfs_ops   test_libhdfs_threaded   test_libhdfs_write
> test_native_mini_dfs   test_libhdfs_read   test_libhdfs_vecsum
> test_libhdfs_zerocopy
>
> Using these binaries I was able to write/read files in my hadoop-cluster.
> /usr/local/hadoop/sources/hadoop-2.6.0-src/hadoop-hdfs-project/hadoop-hdfs/target/native$./test_libhdfs_write
> test-file 100M 10M
>
> But if I try to run FIO using 'ioengine=libhdfs' then it fails with
> the same errors mentioned in earlier email.
>
> As executables are able to utilize the 'libhdfs.so' correctly, it
> seems like FIO is not utilizing it as it should.
>
> I have configured the classpaths and other required settings as
> mentioned in earlier email.
>
> Please let me know if some additional configurations are required to
> run FIO using libhdfs engine.
>
> Thanks,
> LP
>
> On Tue, Jun 5, 2018 at 1:09 AM, Sitsofe Wheeler <sitsofe@xxxxxxxxx> wrote:
>> CC'ing Fabrice and Manish...
>>
>> On 4 June 2018 at 19:36, learn pvt <learnpvt2050@xxxxxxxxx> wrote:
>>> Hi,
>>>
>>> I am trying to configure the 'libhdfs' in my CHD5.14.0 cluster but
>>> getting some issues.
>>>
>>> I have two node CDH-cluster and its working fine as I checked with TestDFSIO.
>>> I set the FIO_LIBHDFS_INCLUDE, FIO_LIBHDFS_LIB, 'hadoop classpath' in
>>> environment.
>>>
>>> Installed FIO on Master and configured it with enabling libhdfs while
>>> 'configure'.
>>>
>>> Created hdfs.fio file as
>>> [global]
>>> runtime=10
>>> [hdfs]
>>> rw=read
>>> namenode=xx.xx.x.xxx
>>> hostname=server1
>>> port=9000
>>> hdfsdirectory=/tmp
>>> chunk_size=1m
>>> hdfs_use_direct=1
>>> ioengine=libhdfs
>>> size=20m
>>> bs=256k
>>> single_instance=1
>>>
>>> Ran the fio job(sudo fio hdfs.fio) which resulted into errors as
>>>
>>> fio-3.6-33-g540e
>>> Starting 1 process
>>> loadFileSystems error
>>> hdfsBuilderConnect forceNewInstance=0, nn=server1, port=9000,
>>> kerbTicketCachePath=(NULL), userName=(NULL) error
>>> hdfsExists: constructNewObjectOfPath error
>>> hdfs: invalid working directory /tmp: Unknown error 255
>>> fio: failed to init engine data: 255
>>> Run status group 0 (all jobs)
>>>
>>> I tried many things but could not configure libhdfs to work correctly
>>> with FIO on my CDH.  Any suggestion would be appreciated!
>>> Thanks in advance.
>>> Note: /tmp directory is present at the hdfs level.
>>>
>>> Thanks,
>>> LP
>>
>> --
>> Sitsofe | http://sucs.org/~sits/
--
To unsubscribe from this list: send the line "unsubscribe fio" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux Kernel]     [Linux SCSI]     [Linux IDE]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux SCSI]

  Powered by Linux