Build failed in Jenkins: 389-DS-NIGHTLY #123

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



See <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/123/>

------------------------------------------
[...truncated 9931 lines...]


INFO:lib389:Found entry dn: cn=dc\3Dpass2\2Cdc\3Dthru,cn=mapping tree,cn=config
cn: dc=pass2,dc=thru
nsslapd-backend: PASS2
nsslapd-state: backend
objectClass: top
objectClass: extensibleObject
objectClass: nsMappingTree


INFO:plugin_tests:test_passthru: PASS

INFO:plugin_tests:Testing referential integrity postoperation...
INFO:plugin_tests:test_referint: PASS

INFO:plugin_tests:Testing Retro Changelog Plugin...
INFO:plugin_tests:test_retrocl: PASS

INFO:plugin_tests:Testing RootDN Access Control...
INFO:plugin_tests:test_rootdn: PASS

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Successfully Tested Dynamic Plugins Functionality (no replication).
INFO:test_dynamic_plugins:####################################################################

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Testing Dynamic Plugins for Memory Corruption (no replication)...
INFO:test_dynamic_plugins:####################################################################

INFO:plugin_tests:Testing Account Policy Plugin...
INFO:plugin_tests:test_acctpolicy: PASS

INFO:plugin_tests:Testing attribute uniqueness...
INFO:plugin_tests:test_attruniq: PASS

INFO:plugin_tests:Testing Auto Membership Plugin...
INFO:plugin_tests:test_automember: PASS

INFO:plugin_tests:Testing Distributed Numeric Assignment Plugin...
INFO:plugin_tests:test_dna: PASS

INFO:plugin_tests:Testing Linked Attributes...
INFO:plugin_tests:test_linkedattrs: PASS

INFO:plugin_tests:Testing MemberOf Plugin...
INFO:plugin_tests:test_memberof: PASS

INFO:plugin_tests:Testing Managed Entries...
INFO:plugin_tests:test_mep: PASS

INFO:plugin_tests:Testing Pass Through Authentication...
INFO:lib389:List backend with suffix=dc=pass2,dc=thru
INFO:lib389:Creating a local backend
INFO:lib389:List backend cn=PASS2,cn=ldbm database,cn=plugins,cn=config
INFO:lib389:Found entry dn: cn=PASS2,cn=ldbm database,cn=plugins,cn=config
cn: PASS2
nsslapd-cachememsize: 10485760
nsslapd-cachesize: -1
nsslapd-directory: /var/lib/dirsrv/slapd-passthru/db/PASS2
nsslapd-dncachememsize: 10485760
nsslapd-readonly: off
nsslapd-require-index: off
nsslapd-suffix: dc=pass2,dc=thru
objectClass: top
objectClass: extensibleObject
objectClass: nsBackendInstance


INFO:lib389:Entry dn: cn="dc=pass2,dc=thru",cn=mapping tree,cn=config
cn: dc=pass2,dc=thru
nsslapd-backend: PASS2
nsslapd-state: backend
objectclass: top
objectclass: extensibleObject
objectclass: nsMappingTree


INFO:lib389:Found entry dn: cn=dc\3Dpass2\2Cdc\3Dthru,cn=mapping tree,cn=config
cn: dc=pass2,dc=thru
nsslapd-backend: PASS2
nsslapd-state: backend
objectClass: top
objectClass: extensibleObject
objectClass: nsMappingTree


INFO:plugin_tests:test_passthru: PASS

INFO:plugin_tests:Testing referential integrity postoperation...
INFO:plugin_tests:test_referint: PASS

INFO:plugin_tests:Testing Retro Changelog Plugin...
INFO:plugin_tests:test_retrocl: PASS

INFO:plugin_tests:Testing RootDN Access Control...
INFO:plugin_tests:test_rootdn: PASS

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Successfully Tested Dynamic Plugins for Memory Corruption (no replication).
INFO:test_dynamic_plugins:####################################################################

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Stressing Dynamic Plugins (no replication)...
INFO:test_dynamic_plugins:####################################################################

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Running stress test (no replication).  Run (1/5)...
INFO:test_dynamic_plugins:####################################################################

INFO:stress_tests:AddUsers - Adding 250 entries (employee)...
INFO:stress_tests:AddUsers - Adding 250 entries (entry)...
INFO:stress_tests:AddUsers - Adding 250 entries (person)...
INFO:stress_tests:AddUsers - Finished adding 250 entries (employee).
INFO:stress_tests:AddUsers - Finished adding 250 entries (entry).
INFO:stress_tests:AddUsers - Finished adding 250 entries (person).
INFO:stress_tests:DelUsers - Deleting 250 entries (employee)...
INFO:stress_tests:DelUsers - Deleting 250 entries (person)...
INFO:stress_tests:DelUsers - Deleting 250 entries (entry)...
INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (person).
INFO:plugin_tests:Testing referential integrity postoperation...
INFO:plugin_tests:test_referint: PASS

INFO:plugin_tests:Testing MemberOf Plugin...
INFO:plugin_tests:test_memberof: PASS

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication).  Completed (1/5)
INFO:test_dynamic_plugins:####################################################################

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Running stress test (no replication).  Run (2/5)...
INFO:test_dynamic_plugins:####################################################################

INFO:stress_tests:AddUsers - Adding 250 entries (employee)...
INFO:stress_tests:AddUsers - Adding 250 entries (person)...
INFO:stress_tests:AddUsers - Adding 250 entries (entry)...
INFO:stress_tests:AddUsers - Finished adding 250 entries (employee).
INFO:stress_tests:AddUsers - Finished adding 250 entries (entry).
INFO:stress_tests:AddUsers - Finished adding 250 entries (person).
INFO:stress_tests:DelUsers - Deleting 250 entries (employee)...
INFO:stress_tests:DelUsers - Deleting 250 entries (entry)...
INFO:stress_tests:DelUsers - Deleting 250 entries (person)...
INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (person).
INFO:plugin_tests:Testing referential integrity postoperation...
INFO:plugin_tests:test_referint: PASS

INFO:plugin_tests:Testing MemberOf Plugin...
INFO:plugin_tests:test_memberof: PASS

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication).  Completed (2/5)
INFO:test_dynamic_plugins:####################################################################

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Running stress test (no replication).  Run (3/5)...
INFO:test_dynamic_plugins:####################################################################

INFO:stress_tests:AddUsers - Adding 250 entries (entry)...
INFO:stress_tests:AddUsers - Adding 250 entries (employee)...
INFO:stress_tests:AddUsers - Adding 250 entries (person)...
INFO:stress_tests:AddUsers - Finished adding 250 entries (employee).
INFO:stress_tests:AddUsers - Finished adding 250 entries (entry).
INFO:stress_tests:AddUsers - Finished adding 250 entries (person).
INFO:stress_tests:DelUsers - Deleting 250 entries (employee)...
INFO:stress_tests:DelUsers - Deleting 250 entries (person)...
INFO:stress_tests:DelUsers - Deleting 250 entries (entry)...
INFO:stress_tests:DelUsers - Finished deleting 250 entries (person).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee).
INFO:plugin_tests:Testing referential integrity postoperation...
INFO:plugin_tests:test_referint: PASS

INFO:plugin_tests:Testing MemberOf Plugin...
INFO:plugin_tests:test_memberof: PASS

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication).  Completed (3/5)
INFO:test_dynamic_plugins:####################################################################

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Running stress test (no replication).  Run (4/5)...
INFO:test_dynamic_plugins:####################################################################

INFO:stress_tests:AddUsers - Adding 250 entries (employee)...
INFO:stress_tests:AddUsers - Adding 250 entries (entry)...
INFO:stress_tests:AddUsers - Adding 250 entries (person)...
INFO:stress_tests:AddUsers - Finished adding 250 entries (person).
INFO:stress_tests:AddUsers - Finished adding 250 entries (entry).
INFO:stress_tests:AddUsers - Finished adding 250 entries (employee).
INFO:stress_tests:DelUsers - Deleting 250 entries (employee)...
INFO:stress_tests:DelUsers - Deleting 250 entries (entry)...
INFO:stress_tests:DelUsers - Deleting 250 entries (person)...
INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (person).
INFO:plugin_tests:Testing referential integrity postoperation...
INFO:plugin_tests:test_referint: PASS

INFO:plugin_tests:Testing MemberOf Plugin...
INFO:plugin_tests:test_memberof: PASS

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication).  Completed (4/5)
INFO:test_dynamic_plugins:####################################################################

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Running stress test (no replication).  Run (5/5)...
INFO:test_dynamic_plugins:####################################################################

INFO:stress_tests:AddUsers - Adding 250 entries (employee)...
INFO:stress_tests:AddUsers - Adding 250 entries (entry)...
INFO:stress_tests:AddUsers - Adding 250 entries (person)...
INFO:stress_tests:AddUsers - Finished adding 250 entries (entry).
INFO:stress_tests:AddUsers - Finished adding 250 entries (person).
INFO:stress_tests:AddUsers - Finished adding 250 entries (employee).
INFO:stress_tests:DelUsers - Deleting 250 entries (entry)...
INFO:stress_tests:DelUsers - Deleting 250 entries (person)...
INFO:stress_tests:DelUsers - Deleting 250 entries (employee)...
INFO:stress_tests:DelUsers - Finished deleting 250 entries (person).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee).
INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry).
INFO:plugin_tests:Testing referential integrity postoperation...
INFO:plugin_tests:test_referint: PASS

INFO:plugin_tests:Testing MemberOf Plugin...
INFO:plugin_tests:test_memberof: PASS

INFO:test_dynamic_plugins:####################################################################
INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication).  Completed (5/5)
INFO:test_dynamic_plugins:####################################################################

INFO:test_dynamic_plugins:Setting up replication, and rerunning the tests...

============== 24 failed, 471 passed, 27 error in 8991.82 seconds =============='
+ '[' 1 -ne 0 ']'
+ echo CI Tests 'FAILED!'
CI Tests FAILED!
+ echo ============================= test session starts ============================== platform linux2 -- Python 2.7.12, pytest-2.9.2, py-1.4.31, pluggy-0.3.1 -- /usr/bin/python2 cachedir: .cache rootdir: <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests,> inifile: plugins: sourceorder-0.5, multihost-1.0 collecting ... collected 522 items tickets/ticket1347760_test.py::test_ticket1347760 PASSED tickets/ticket365_test.py::test_ticket365 PASSED tickets/ticket397_test.py::test_397 PASSED tickets/ticket47313_test.py::test_ticket47313_run PASSED tickets/ticket47384_test.py::test_ticket47384 PASSED tickets/ticket47431_test.py::test_ticket47431_0 PASSED tickets/ticket47431_test.py::test_ticket47431_1 PASSED tickets/ticket47431_test.py::test_ticket47431_2 PASSED tickets/ticket47431_test.py::test_ticket47431_3 PASSED tickets/ticket47462_test.py::test_ticket47462 PASSED tickets/ticket47490_test.py::test_ticket47490_init PASSED tickets/ticket47490_test.py::test_ticket47490_one PASSED tickets/ticket47490_test.py::test_ticket47490_two PASSED tickets/ticket47490_test.py::test_ticket47490_three PASSED tickets/ticket47490_test.py::test_ticket47490_four PASSED tickets/ticket47490_test.py::test_ticket47490_five PASSED tickets/ticket47490_test.py::test_ticket47490_six PASSED tickets/ticket47490_test.py::test_ticket47490_seven PASSED tickets/ticket47490_test.py::test_ticket47490_eight PASSED tickets/ticket47490_test.py::test_ticket47490_nine PASSED tickets/ticket47536_test.py::test_ticket47536 PASSED tickets/ticket47553_test.py::test_ticket47553 PASSED tickets/ticket47560_test.py::test_ticket47560 PASSED tickets/ticket47573_test.py::test_ticket47573_init PASSED tickets/ticket47573_test.py::test_ticket47573_one PASSED tickets/ticket47573_test.py::test_ticket47573_two PASSED tickets/ticket47573_test.py::test_ticket47573_three PASSED tickets/ticket47619_test.py::test_ticket47619_init PASSED tickets/ticket47619_test.py::test_ticket47619_create_index PASSED tickets/ticket47619_test.
py::test_ticket47619_reindex PASSED tickets/ticket47619_test.py::test_ticket47619_check_indexed_search PASSED tickets/ticket47640_test.py::test_ticket47640 PASSED tickets/ticket47653MMR_test.py::test_ticket47653_init PASSED tickets/ticket47653MMR_test.py::test_ticket47653_add PASSED tickets/ticket47653MMR_test.py::test_ticket47653_modify PASSED tickets/ticket47653_test.py::test_ticket47653_init PASSED tickets/ticket47653_test.py::test_ticket47653_add PASSED tickets/ticket47653_test.py::test_ticket47653_search PASSED tickets/ticket47653_test.py::test_ticket47653_modify PASSED tickets/ticket47653_test.py::test_ticket47653_delete PASSED tickets/ticket47669_test.py::test_ticket47669_init FAILED tickets/ticket47669_test.py::test_ticket47669_changelog_maxage FAILED tickets/ticket47669_test.py::test_ticket47669_changelog_triminterval FAILED tickets/ticket47669_test.py::test_ticket47669_changelog_compactdbinterval FAILED tickets/ticket47669_test.py::test_ticket47669_retrochangelog_maxage FAILED tickets/ticket47676_test.py::test_ticket47676_init PASSED tickets/ticket47676_test.py::test_ticket47676_skip_oc_at PASSED tickets/ticket47676_test.py::test_ticket47676_reject_action PASSED tickets/ticket47714_test.py::test_ticket47714_init PASSED tickets/ticket47714_test.py::test_ticket47714_run_0 PASSED tickets/ticket47714_test.py::test_ticket47714_run_1 PASSED tickets/ticket47721_test.py::test_ticket47721_init PASSED tickets/ticket47721_test.py::test_ticket47721_0 PASSED tickets/ticket47721_test.py::test_ticket47721_1 PASSED tickets/ticket47721_test.py::test_ticket47721_2 PASSED tickets/ticket47721_test.py::test_ticket47721_3 PASSED tickets/ticket47721_test.py::test_ticket47721_4 PASSED tickets/ticket47781_test.py::test_ticket47781 PASSED tickets/ticket47787_test.py::test_ticket47787_init PASSED tickets/ticket47787_test.py::test_ticket47787_2 PASSED tickets/ticket47808_test.py::test_ticket47808_run PASSED tickets/ticket47815_test.py::test_ticket47815 PASSED tickets/ticket47819_test.py::test_ticket47819 PASSED tickets/ticket47823
_test.py::test_ticket47823_init PASSED tickets/ticket47823_test.py::test_ticket47823_one_container_add PASSED tickets/ticket47823_test.py::test_ticket47823_one_container_mod PASSED tickets/ticket47823_test.py::test_ticket47823_one_container_modrdn PASSED tickets/ticket47823_test.py::test_ticket47823_multi_containers_add PASSED tickets/ticket47823_test.py::test_ticket47823_multi_containers_mod PASSED tickets/ticket47823_test.py::test_ticket47823_multi_containers_modrdn PASSED tickets/ticket47823_test.py::test_ticket47823_across_multi_containers_add PASSED tickets/ticket47823_test.py::test_ticket47823_across_multi_containers_mod PASSED tickets/ticket47823_test.py::test_ticket47823_across_multi_containers_modrdn PASSED tickets/ticket47823_test.py::test_ticket47823_invalid_config_1 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_2 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_3 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_4 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_5 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_6 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_7 FAILED tickets/ticket47828_test.py::test_ticket47828_init PASSED tickets/ticket47828_test.py::test_ticket47828_run_0 PASSED tickets/ticket47828_test.py::test_ticket47828_run_1 PASSED tickets/ticket47828_test.py::test_ticket47828_run_2 PASSED tickets/ticket47828_test.py::test_ticket47828_run_3 PASSED tickets/ticket47828_test.py::test_ticket47828_run_4 PASSED tickets/ticket47828_test.py::test_ticket47828_run_5 PASSED tickets/ticket47828_test.py::test_ticket47828_run_6 PASSED tickets/ticket47828_test.py::test_ticket47828_run_7 PASSED tickets/ticket47828_test.py::test_ticket47828_run_8 PASSED tickets/ticket47828_test.py::test_ticket47828_run_9 PASSED tickets/ticket47828_test.py::test_ticket47828_run_10 PASSED tickets/ticket47828_test.py::test_ticket47828_run_11 PASSED tickets/ticket47828_test.py::test_ticket47828_run_12 PASSED tick
ets/ticket47828_test.py::test_ticket47828_run_13 PASSED tickets/ticket47828_test.py::test_ticket47828_run_14 PASSED tickets/ticket47828_test.py::test_ticket47828_run_15 PASSED tickets/ticket47828_test.py::test_ticket47828_run_16 PASSED tickets/ticket47828_test.py::test_ticket47828_run_17 PASSED tickets/ticket47828_test.py::test_ticket47828_run_18 PASSED tickets/ticket47828_test.py::test_ticket47828_run_19 PASSED tickets/ticket47828_test.py::test_ticket47828_run_20 PASSED tickets/ticket47828_test.py::test_ticket47828_run_21 PASSED tickets/ticket47828_test.py::test_ticket47828_run_22 PASSED tickets/ticket47828_test.py::test_ticket47828_run_23 PASSED tickets/ticket47828_test.py::test_ticket47828_run_24 PASSED tickets/ticket47828_test.py::test_ticket47828_run_25 PASSED tickets/ticket47828_test.py::test_ticket47828_run_26 PASSED tickets/ticket47828_test.py::test_ticket47828_run_27 PASSED tickets/ticket47828_test.py::test_ticket47828_run_28 PASSED tickets/ticket47828_test.py::test_ticket47828_run_29 PASSED tickets/ticket47828_test.py::test_ticket47828_run_30 PASSED tickets/ticket47828_test.py::test_ticket47828_run_31 PASSED tickets/ticket47829_test.py::test_ticket47829_init PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_2 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_3 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_2 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_3 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_out_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_out_user_2 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_out_user_3 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_modrdn_active_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_modrdn_stage_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mo
d_active_user_modrdn_out_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_modrdn_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_modrdn_active_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_modrdn_stage_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_indirect_active_group_1 PASSED tickets/ticket47829_test.py::test_ticket47829_indirect_active_group_2 PASSED tickets/ticket47829_test.py::test_ticket47829_indirect_active_group_3 PASSED tickets/ticket47829_test.py::test_ticket47829_indirect_active_group_4 PASSED tickets/ticket47833_test.py::test_ticket47829_init PASSED tickets/ticket47833_test.py::test_ticket47829_mod_stage_user_modrdn_stage_user_1 PASSED tickets/ticket47869MMR_test.py::test_ticket47869_init PASSED tickets/ticket47869MMR_test.py::test_ticket47869_check PASSED tickets/ticket47871_test.py::test_ticket47871_init PASSED tickets/ticket47871_test.py::test_ticket47871_1 PASSED tickets/ticket47871_test.py::test_ticket47871_2 PASSED tickets/ticket47900_test.py::test_ticket47900 PASSED tickets/ticket47910_test.py::test_ticket47910_logconv_start_end_positive PASSED tickets/ticket47910_test.py::test_ticket47910_logconv_start_end_negative PASSED tickets/ticket47910_test.py::test_ticket47910_logconv_start_end_invalid PASSED tickets/ticket47910_test.py::test_ticket47910_logconv_noaccesslogs PASSED tickets/ticket47920_test.py::test_ticket47920_init PASSED tickets/ticket47920_test.py::test_ticket47920_mod_readentry_ctrl PASSED tickets/ticket47921_test.py::test_ticket47921 PASSED tickets/ticket47927_test.py::test_ticket47927_init PASSED tickets/ticket47927_test.py::test_ticket47927_one PASSED tickets/ticket47927_test.py::test_ticket47927_two PASSED tickets/ticket47927_test.py::test_ticket47927_three PASSED tickets/ticket47927_test.py::test_ticket47927_four PASSED tickets/ticket47927_test.py::test_ticket47927_five PASSED tickets/ticket47927_test.py::test_ticket47927_six PASSED tickets/ticket47931_test.py::test_ticket47931 PASSED tickets/ticket4
7937_test.py::test_ticket47937 PASSED tickets/ticket47950_test.py::test_ticket47950 PASSED tickets/ticket47953_test.py::test_ticket47953 PASSED tickets/ticket47963_test.py::test_ticket47963 PASSED tickets/ticket47966_test.py::test_ticket47966 PASSED tickets/ticket47970_test.py::test_ticket47970 PASSED tickets/ticket47973_test.py::test_ticket47973 PASSED tickets/ticket47976_test.py::test_ticket47976_init PASSED tickets/ticket47976_test.py::test_ticket47976_1 PASSED tickets/ticket47976_test.py::test_ticket47976_2 PASSED tickets/ticket47976_test.py::test_ticket47976_3 PASSED tickets/ticket47980_test.py::test_ticket47980 PASSED tickets/ticket47981_test.py::test_ticket47981 PASSED tickets/ticket47988_test.py::test_ticket47988_init PASSED tickets/ticket47988_test.py::test_ticket47988_1 PASSED tickets/ticket47988_test.py::test_ticket47988_2 PASSED tickets/ticket47988_test.py::test_ticket47988_3 PASSED tickets/ticket47988_test.py::test_ticket47988_4 PASSED tickets/ticket47988_test.py::test_ticket47988_5 PASSED tickets/ticket47988_test.py::test_ticket47988_6 PASSED tickets/ticket48005_test.py::test_ticket48005_setup PASSED tickets/ticket48005_test.py::test_ticket48005_memberof PASSED tickets/ticket48005_test.py::test_ticket48005_automember PASSED tickets/ticket48005_test.py::test_ticket48005_syntaxvalidate PASSED tickets/ticket48005_test.py::test_ticket48005_usn PASSED tickets/ticket48005_test.py::test_ticket48005_schemareload PASSED tickets/ticket48013_test.py::test_ticket48013 PASSED tickets/ticket48026_test.py::test_ticket48026 PASSED tickets/ticket48109_test.py::test_ticket48109 PASSED tickets/ticket48170_test.py::test_ticket48170 PASSED tickets/ticket48194_test.py::test_init PASSED tickets/ticket48194_test.py::test_run_0 PASSED tickets/ticket48194_test.py::test_run_1 PASSED tickets/ticket48194_test.py::test_run_2 PASSED tickets/ticket48194_test.py::test_run_3 PASSED tickets/ticket48194_test.py::test_run_4 PASSED tickets/ticket48194_test.py::test_run_5 PASSED tickets/ticket48194_test.py::test_run_6 PASSED tickets/tick
et48194_test.py::test_run_7 PASSED tickets/ticket48194_test.py::test_run_8 PASSED tickets/ticket48194_test.py::test_run_9 PASSED tickets/ticket48194_test.py::test_run_10 PASSED tickets/ticket48194_test.py::test_run_11 PASSED tickets/ticket48212_test.py::test_ticket48212 PASSED tickets/ticket48214_test.py::test_ticket48214_run PASSED tickets/ticket48226_test.py::test_ticket48226_set_purgedelay PASSED tickets/ticket48226_test.py::test_ticket48226_1 FAILED tickets/ticket48228_test.py::test_ticket48228_test_global_policy PASSED tickets/ticket48228_test.py::test_ticket48228_test_subtree_policy PASSED tickets/ticket48233_test.py::test_ticket48233 PASSED tickets/ticket48234_test.py::test_ticket48234 PASSED tickets/ticket48252_test.py::test_ticket48252_setup PASSED tickets/ticket48252_test.py::test_ticket48252_run_0 PASSED tickets/ticket48252_test.py::test_ticket48252_run_1 PASSED tickets/ticket48265_test.py::test_ticket48265_test PASSED tickets/ticket48266_test.py::test_ticket48266_fractional PASSED tickets/ticket48266_test.py::test_ticket48266_check_repl_desc PASSED tickets/ticket48266_test.py::test_ticket48266_count_csn_evaluation FAILED tickets/ticket48270_test.py::test_ticket48270_init PASSED tickets/ticket48270_test.py::test_ticket48270_homeDirectory_indexed_cis FAILED tickets/ticket48270_test.py::test_ticket48270_homeDirectory_mixed_value PASSED tickets/ticket48270_test.py::test_ticket48270_extensible_search PASSED tickets/ticket48272_test.py::test_ticket48272 PASSED tickets/ticket48294_test.py::test_48294_init PASSED tickets/ticket48294_test.py::test_48294_run_0 PASSED tickets/ticket48294_test.py::test_48294_run_1 PASSED tickets/ticket48294_test.py::test_48294_run_2 PASSED tickets/ticket48295_test.py::test_48295_init PASSED tickets/ticket48295_test.py::test_48295_run PASSED tickets/ticket48312_test.py::test_ticket48312 PASSED tickets/ticket48325_test.py::test_ticket48325 PASSED tickets/ticket48342_test.py::test_ticket4026 PASSED tickets/ticket48354_test.py::test_ticket48354 PASSED tickets/ticket48362_test.py::tes
t_ticket48362 PASSED tickets/ticket48366_test.py::test_ticket48366_init PASSED tickets/ticket48366_test.py::test_ticket48366_search_user PASSED tickets/ticket48366_test.py::test_ticket48366_search_dm PASSED tickets/ticket48370_test.py::test_ticket48370 PASSED tickets/ticket48383_test.py::test_ticket48383 FAILED tickets/ticket48497_test.py::test_ticket48497_init PASSED tickets/ticket48497_test.py::test_ticket48497_homeDirectory_mixed_value PASSED tickets/ticket48497_test.py::test_ticket48497_extensible_search PASSED tickets/ticket48497_test.py::test_ticket48497_homeDirectory_index_cfg PASSED tickets/ticket48497_test.py::test_ticket48497_homeDirectory_index_run FAILED tickets/ticket48637_test.py::test_ticket48637 PASSED tickets/ticket48665_test.py::test_ticket48665 PASSED tickets/ticket48745_test.py::test_ticket48745_init PASSED tickets/ticket48745_test.py::test_ticket48745_homeDirectory_indexed_cis FAILED tickets/ticket48745_test.py::test_ticket48745_homeDirectory_mixed_value PASSED tickets/ticket48745_test.py::test_ticket48745_extensible_search_after_index PASSED tickets/ticket48746_test.py::test_ticket48746_init PASSED tickets/ticket48746_test.py::test_ticket48746_homeDirectory_indexed_cis FAILED tickets/ticket48746_test.py::test_ticket48746_homeDirectory_mixed_value PASSED tickets/ticket48746_test.py::test_ticket48746_extensible_search_after_index PASSED tickets/ticket48746_test.py::test_ticket48746_homeDirectory_indexed_ces FAILED tickets/ticket48755_test.py::test_ticket48755 PASSED tickets/ticket48759_test.py::test_ticket48759 PASSED tickets/ticket48784_test.py::test_ticket48784 PASSED tickets/ticket48798_test.py::test_ticket48798 PASSED tickets/ticket48799_test.py::test_ticket48799 PASSED tickets/ticket48808_test.py::test_ticket48808 PASSED tickets/ticket48844_test.py::test_ticket48844_init PASSED tickets/ticket48844_test.py::test_ticket48844_bitwise_on PASSED tickets/ticket48844_test.py::test_ticket48844_bitwise_off PASSED tickets/ticket48891_test.py::test_ticket48891_setup PASSED tickets/ticket48893_test.p
y::test_ticket48893 PASSED tickets/ticket48896_test.py::test_ticket48896 PASSED tickets/ticket48906_test.py::test_ticket48906_setup PASSED tickets/ticket48906_test.py::test_ticket48906_dblock_default PASSED tickets/ticket48906_test.py::test_ticket48906_dblock_ldap_update FAILED tickets/ticket48906_test.py::test_ticket48906_dblock_edit_update FAILED tickets/ticket48906_test.py::test_ticket48906_dblock_robust FAILED tickets/ticket48916_test.py::test_ticket48916 ERROR tickets/ticket48956_test.py::test_ticket48956 PASSED tickets/ticket548_test.py::test_ticket548_test_with_no_policy PASSED tickets/ticket548_test.py::test_ticket548_test_global_policy PASSED tickets/ticket548_test.py::test_ticket548_test_subtree_policy PASSED suites/acct_usability_plugin/acct_usability_test.py::test_acct_usability_init PASSED suites/acct_usability_plugin/acct_usability_test.py::test_acct_usability_ PASSED suites/acctpolicy_plugin/acctpolicy_test.py::test_acctpolicy_init PASSED suites/acctpolicy_plugin/acctpolicy_test.py::test_acctpolicy_ PASSED 'suites/acl/acl_test.py::test_aci_attr_subtype_targetattr[lang-ja]' ERROR 'suites/acl/acl_test.py::test_aci_attr_subtype_targetattr[binary]' ERROR 'suites/acl/acl_test.py::test_aci_attr_subtype_targetattr[phonetic]' ERROR suites/acl/acl_test.py::test_mode_default_add_deny ERROR suites/acl/acl_test.py::test_mode_default_delete_deny ERROR 'suites/acl/acl_test.py::test_moddn_staging_prod[0-cn=staged' 'user,dc=example,dc=com-cn=accounts,dc=example,dc=com-False]' ERROR 'suites/acl/acl_test.py::test_moddn_staging_prod[1-cn=staged' 'user,dc=example,dc=com-cn=accounts,dc=example,dc=com-False]' ERROR 'suites/acl/acl_test.py::test_moddn_staging_prod[2-cn=staged' 'user,dc=example,dc=com-cn=bad*,dc=example,dc=com-True]' ERROR 'suites/acl/acl_test.py::test_moddn_staging_prod[3-cn=st*,dc=example,dc=com-cn=accounts,dc=example,dc=com-False]' ERROR 'suites/acl/acl_test.py::test_moddn_staging_prod[4-cn=bad*,dc=example,dc=com-cn=accounts,dc=example,dc=com-True]' ERROR 'suites/acl/acl_test.py::test_moddn_staging_pro
d[5-cn=st*,dc=example,dc=com-cn=ac*,dc=example,dc=com-False]' ERROR 'suites/acl/acl_test.py::test_moddn_staging_prod[6-None-cn=ac*,dc=example,dc=com-False]' ERROR 'suites/acl/acl_test.py::test_moddn_staging_prod[7-cn=st*,dc=example,dc=com-None-False]' ERROR 'suites/acl/acl_test.py::test_moddn_staging_prod[8-None-None-False]' ERROR suites/acl/acl_test.py::test_moddn_staging_prod_9 ERROR suites/acl/acl_test.py::test_moddn_prod_staging ERROR suites/acl/acl_test.py::test_check_repl_M2_to_M1 ERROR suites/acl/acl_test.py::test_moddn_staging_prod_except ERROR suites/acl/acl_test.py::test_mode_default_ger_no_moddn ERROR suites/acl/acl_test.py::test_mode_default_ger_with_moddn ERROR suites/acl/acl_test.py::test_mode_switch_default_to_legacy ERROR suites/acl/acl_test.py::test_mode_legacy_ger_no_moddn1 ERROR suites/acl/acl_test.py::test_mode_legacy_ger_no_moddn2 ERROR suites/acl/acl_test.py::test_mode_legacy_ger_with_moddn ERROR suites/acl/acl_test.py::test_rdn_write_get_ger ERROR suites/acl/acl_test.py::test_rdn_write_modrdn_anonymous ERROR suites/attr_encryption/attr_encrypt_test.py::test_attr_encrypt_init PASSED suites/attr_encryption/attr_encrypt_test.py::test_attr_encrypt_ PASSED suites/attr_uniqueness_plugin/attr_uniqueness_test.py::test_attr_uniqueness_init PASSED suites/attr_uniqueness_plugin/attr_uniqueness_test.py::test_attr_uniqueness PASSED suites/automember_plugin/automember_test.py::test_automember_init PASSED suites/automember_plugin/automember_test.py::test_automember_ PASSED suites/basic/basic_test.py::test_basic_ops PASSED suites/basic/basic_test.py::test_basic_import_export PASSED suites/basic/basic_test.py::test_basic_backup PASSED suites/basic/basic_test.py::test_basic_acl PASSED suites/basic/basic_test.py::test_basic_searches PASSED suites/basic/basic_test.py::test_basic_referrals PASSED suites/basic/basic_test.py::test_basic_systemctl PASSED suites/basic/basic_test.py::test_basic_ldapagent PASSED suites/basic/basic_test.py::test_basic_dse PASSED 'suites/basic/basic_test.py::test_def_rootdse_attr[namin
gContexts]' PASSED 'suites/basic/basic_test.py::test_def_rootdse_attr[supportedLDAPVersion]' PASSED 'suites/basic/basic_test.py::test_def_rootdse_attr[supportedControl]' PASSED 'suites/basic/basic_test.py::test_def_rootdse_attr[supportedExtension]' PASSED 'suites/basic/basic_test.py::test_def_rootdse_attr[supportedSASLMechanisms]' PASSED 'suites/basic/basic_test.py::test_def_rootdse_attr[vendorName]' PASSED 'suites/basic/basic_test.py::test_def_rootdse_attr[vendorVersion]' PASSED 'suites/basic/basic_test.py::test_mod_def_rootdse_attr[namingContexts]' PASSED 'suites/basic/basic_test.py::test_mod_def_rootdse_attr[supportedLDAPVersion]' PASSED 'suites/basic/basic_test.py::test_mod_def_rootdse_attr[supportedControl]' PASSED 'suites/basic/basic_test.py::test_mod_def_rootdse_attr[supportedExtension]' PASSED 'suites/basic/basic_test.py::test_mod_def_rootdse_attr[supportedSASLMechanisms]' PASSED 'suites/basic/basic_test.py::test_mod_def_rootdse_attr[vendorName]' PASSED 'suites/basic/basic_test.py::test_mod_def_rootdse_attr[vendorVersion]' PASSED suites/betxns/betxn_test.py::test_betxn_init PASSED suites/betxns/betxn_test.py::test_betxt_7bit PASSED suites/betxns/betxn_test.py::test_betxn_attr_uniqueness PASSED suites/betxns/betxn_test.py::test_betxn_memberof PASSED suites/chaining_plugin/chaining_test.py::test_chaining_init PASSED suites/chaining_plugin/chaining_test.py::test_chaining_ PASSED suites/clu/clu_test.py::test_clu_init PASSED suites/clu/clu_test.py::test_clu_pwdhash PASSED suites/clu/db2ldif_test.py::test_db2ldif_init PASSED suites/collation_plugin/collatation_test.py::test_collatation_init PASSED suites/collation_plugin/collatation_test.py::test_collatation_ PASSED suites/config/config_test.py::test_maxbersize_repl PASSED suites/config/config_test.py::test_config_listen_backport_size PASSED suites/config/config_test.py::test_config_deadlock_policy PASSED suites/cos_plugin/cos_test.py::test_cos_init PASSED suites/cos_plugin/cos_test.py::test_cos_ PASSED suites/deref_plugin/deref_test.py::test_deref_init PASSED 
suites/deref_plugin/deref_test.py::test_deref_ PASSED suites/disk_monitoring/disk_monitor_test.py::test_disk_monitor_init PASSED suites/disk_monitoring/disk_monitor_test.py::test_disk_monitor_ PASSED suites/distrib_plugin/distrib_test.py::test_distrib_init PASSED suites/distrib_plugin/distrib_test.py::test_distrib_ PASSED suites/dna_plugin/dna_test.py::test_dna_init PASSED suites/dna_plugin/dna_test.py::test_dna_ PASSED suites/ds_logs/ds_logs_test.py::test_ds_logs_init PASSED suites/ds_logs/ds_logs_test.py::test_ds_logs_ PASSED suites/dynamic-plugins/test_dynamic_plugins.py::test_dynamic_plugins FAILED suites/filter/filter_test.py::test_filter_init PASSED suites/filter/filter_test.py::test_filter_escaped PASSED suites/filter/filter_test.py::test_filter_search_original_attrs PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_supported_features PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-False-oper_attr_list0]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-False-oper_attr_list0-*]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-False-oper_attr_list0-objectClass]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-True-oper_attr_list1]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-True-oper_attr_list1-*]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-True-oper_attr_list1-objectClass]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-False-oper_attr_list2]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-False-oper_attr_list2-*]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-False-oper_attr_list2-objectClass]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-True-oper_attr_list3]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basi
c[ou=people,dc=example,dc=com-True-oper_attr_list3-*]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-True-oper_attr_list3-objectClass]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-False-oper_attr_list4]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-False-oper_attr_list4-*]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-False-oper_attr_list4-objectClass]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-True-oper_attr_list5]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-True-oper_attr_list5-*]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-True-oper_attr_list5-objectClass]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[cn=config-False-oper_attr_list6]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[cn=config-False-oper_attr_list6-*]' PASSED 'suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[cn=config-False-oper_attr_list6-objectClass]' PASSED suites/get_effective_rights/ger_test.py::test_ger_init PASSED suites/get_effective_rights/ger_test.py::test_ger_ PASSED suites/gssapi_repl/gssapi_repl_test.py::test_gssapi_repl PASSED suites/ldapi/ldapi_test.py::test_ldapi_init PASSED suites/ldapi/ldapi_test.py::test_ldapi_ PASSED suites/linkedattrs_plugin/linked_attrs_test.py::test_linked_attrs_init PASSED suites/linkedattrs_plugin/linked_attrs_test.py::test_linked_attrs_ PASSED suites/mapping_tree/mapping_tree_test.py::test_mapping_tree_init PASSED suites/mapping_tree/mapping_tree_test.py::test_mapping_tree_ PASSED suites/memberof_plugin/memberof_test.py::test_memberof_auto_ad
d_oc PASSED suites/memory_leaks/range_search_test.py::test_range_search_init PASSED suites/memory_leaks/range_search_test.py::test_range_search PASSED suites/monitor/monitor_test.py::test_monitor_init PASSED suites/monitor/monitor_test.py::test_monitor_ PASSED 'suites/paged_results/paged_results_test.py::test_search_success[6-5]' PASSED 'suites/paged_results/paged_results_test.py::test_search_success[5-5]' PASSED 'suites/paged_results/paged_results_test.py::test_search_success[5-25]' PASSED 'suites/paged_results/paged_results_test.py::test_search_limits_fail[50-200-cn=config,cn=ldbm' 'database,cn=plugins,cn=config-nsslapd-idlistscanlimit-100-UNWILLING_TO_PERFORM]' PASSED 'suites/paged_results/paged_results_test.py::test_search_limits_fail[5-15-cn=config-nsslapd-timelimit-20-UNAVAILABLE_CRITICAL_EXTENSION]' PASSED 'suites/paged_results/paged_results_test.py::test_search_limits_fail[21-50-cn=config-nsslapd-sizelimit-20-SIZELIMIT_EXCEEDED]' PASSED 'suites/paged_results/paged_results_test.py::test_search_limits_fail[21-50-cn=config-nsslapd-pagedsizelimit-5-SIZELIMIT_EXCEEDED]' PASSED 'suites/paged_results/paged_results_test.py::test_search_limits_fail[5-50-cn=config,cn=ldbm' 'database,cn=plugins,cn=config-nsslapd-lookthroughlimit-20-ADMINLIMIT_EXCEEDED]' PASSED suites/paged_results/paged_results_test.py::test_search_sort_success PASSED suites/paged_results/paged_results_test.py::test_search_abandon PASSED suites/paged_results/paged_results_test.py::test_search_with_timelimit PASSED 'suites/paged_results/paged_results_test.py::test_search_dns_ip_aci[dns' = '"localhost.localdomain"]' PASSED 'suites/paged_results/paged_results_test.py::test_search_dns_ip_aci[ip' = '"::1"' or ip = '"127.0.0.1"]' PASSED suites/paged_results/paged_results_test.py::test_search_multiple_paging PASSED 'suites/paged_results/paged_results_test.py::test_search_invalid_cookie[1000]' PASSED 'suites/paged_results/paged_results_test.py::test_search_invalid_cookie[-1]' PASSED suites/paged_results/paged_results_test.py::test_search_abandon_with_zero_s
ize PASSED suites/paged_results/paged_results_test.py::test_search_pagedsizelimit_success PASSED 'suites/paged_results/paged_results_test.py::test_search_nspagedsizelimit[5-15-PASS]' PASSED 'suites/paged_results/paged_results_test.py::test_search_nspagedsizelimit[15-5-SIZELIMIT_EXCEEDED]' PASSED 'suites/paged_results/paged_results_test.py::test_search_paged_limits[conf_attr_values0-ADMINLIMIT_EXCEEDED]' PASSED 'suites/paged_results/paged_results_test.py::test_search_paged_limits[conf_attr_values1-PASS]' PASSED 'suites/paged_results/paged_results_test.py::test_search_paged_user_limits[conf_attr_values0-ADMINLIMIT_EXCEEDED]' PASSED 'suites/paged_results/paged_results_test.py::test_search_paged_user_limits[conf_attr_values1-PASS]' PASSED suites/paged_results/paged_results_test.py::test_ger_basic PASSED suites/paged_results/paged_results_test.py::test_multi_suffix_search PASSED 'suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_success[None]' PASSED 'suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_success[-1]' PASSED 'suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_success[1000]' PASSED 'suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_failure[0]' PASSED 'suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_failure[1]' PASSED suites/pam_passthru_plugin/pam_test.py::test_pam_init PASSED suites/pam_passthru_plugin/pam_test.py::test_pam_ PASSED suites/passthru_plugin/passthru_test.py::test_passthru_init PASSED suites/passthru_plugin/passthru_test.py::test_passthru_ PASSED suites/password/password_test.py::test_password_init PASSED suites/password/password_test.py::test_password_delete_specific_password PASSED suites/password/pwdAdmin_test.py::test_pwdAdmin_init PASSED suites/password/pwdAdmin_test.py::test_pwdAdmin PASSED suites/password/pwdAdmin_test.py::test_pwdAdmin_config_validation PASSED 'suites/password/pwdPolicy_attribute_test.py::test_change_pwd[on-off-UNWILLING_TO_PERFORM]' PASSED 'suites/passwo
rd/pwdPolicy_attribute_test.py::test_change_pwd[off-off-UNWILLING_TO_PERFORM]' PASSED 'suites/password/pwdPolicy_attribute_test.py::test_change_pwd[off-on-None]' PASSED 'suites/password/pwdPolicy_attribute_test.py::test_change_pwd[on-on-None]' PASSED suites/password/pwdPolicy_attribute_test.py::test_pwd_min_age PASSED 'suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_no_restrictions[off-off]' PASSED 'suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_no_restrictions[on-off]' PASSED 'suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_no_restrictions[off-on]' PASSED 'suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_restrictions[cn=config]' PASSED 'suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_restrictions[cn="cn=nsPwPolicyEntry,ou=People,dc=example,dc=com",cn=nsPwPolicyContainer,ou=People,dc=example,dc=com]' PASSED suites/password/pwdPolicy_syntax_test.py::test_pwdPolicy_syntax PASSED 'suites/password/pwdPolicy_warning_test.py::test_different_values[' ']' PASSED 'suites/password/pwdPolicy_warning_test.py::test_different_values[junk123]' PASSED 'suites/password/pwdPolicy_warning_test.py::test_different_values[on]' PASSED 'suites/password/pwdPolicy_warning_test.py::test_different_values[off]' PASSED suites/password/pwdPolicy_warning_test.py::test_expiry_time PASSED 'suites/password/pwdPolicy_warning_test.py::test_password_warning[passwordSendExpiringTime-off]' PASSED 'suites/password/pwdPolicy_warning_test.py::test_password_warning[passwordWarning-3600]' PASSED suites/password/pwdPolicy_warning_test.py::test_with_different_password_states PASSED suites/password/pwdPolicy_warning_test.py::test_default_behavior PASSED suites/password/pwdPolicy_warning_test.py::test_with_local_policy PASSED suites/password/pwd_algo_test.py::test_pwd_algo_test PASSED suites/password/pwp_history_test.py::test_pwp_history_test PASSED suites/posix_winsync_plugin/posix_winsync_test.py::test_posix_winsync_init PASSED suites/posix_winsync_plugin/posix_winsync_test.p
y::test_posix_winsync_ PASSED suites/psearch/psearch_test.py::test_psearch_init PASSED suites/psearch/psearch_test.py::test_psearch_ PASSED suites/referint_plugin/referint_test.py::test_referint_init PASSED suites/referint_plugin/referint_test.py::test_referint_ PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_init PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_clean PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_clean_restart PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_clean_force PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_abort PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_abort_restart PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_abort_certify PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_stress_clean PASSED suites/replication/wait_for_async_feature_test.py::test_not_int_value PASSED suites/replication/wait_for_async_feature_test.py::test_multi_value PASSED 'suites/replication/wait_for_async_feature_test.py::test_value_check[waitfor_async_attr0]' PASSED 'suites/replication/wait_for_async_feature_test.py::test_value_check[waitfor_async_attr1]' PASSED 'suites/replication/wait_for_async_feature_test.py::test_value_check[waitfor_async_attr2]' PASSED 'suites/replication/wait_for_async_feature_test.py::test_value_check[waitfor_async_attr3]' PASSED 'suites/replication/wait_for_async_feature_test.py::test_behavior_with_value[waitfor_async_attr0]' PASSED 'suites/replication/wait_for_async_feature_test.py::test_behavior_with_value[waitfor_async_attr1]' PASSED 'suites/replication/wait_for_async_feature_test.py::test_behavior_with_value[waitfor_async_attr2]' PASSED 'suites/replication/wait_for_async_feature_test.py::test_behavior_with_value[waitfor_async_attr3]' PASSED suites/replsync_plugin/repl_sync_test.py::test_repl_sync_init PASSED suites/replsync_plugin/repl_sync_test.py::test_repl_sync_ PASSED suites/resource_limits/res_limits_test.py::test_res_limits_init PASSED suites/re
source_limits/res_limits_test.py::test_res_limits_ PASSED suites/retrocl_plugin/retrocl_test.py::test_retrocl_init PASSED suites/retrocl_plugin/retrocl_test.py::test_retrocl_ PASSED suites/reverpwd_plugin/reverpwd_test.py::test_reverpwd_init PASSED suites/reverpwd_plugin/reverpwd_test.py::test_reverpwd_ PASSED suites/roles_plugin/roles_test.py::test_roles_init PASSED suites/roles_plugin/roles_test.py::test_roles_ PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_init PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_specific_time PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_day_of_week PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_denied_ip PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_denied_host PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_allowed_ip PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_allowed_host PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_config_validate PASSED suites/sasl/sasl_test.py::test_sasl_init PASSED suites/sasl/sasl_test.py::test_sasl_ PASSED suites/schema/test_schema.py::test_schema_comparewithfiles PASSED suites/schema_reload_plugin/schema_reload_test.py::test_schema_reload_init PASSED suites/schema_reload_plugin/schema_reload_test.py::test_schema_reload_ PASSED suites/snmp/snmp_test.py::test_snmp_init PASSED suites/snmp/snmp_test.py::test_snmp_ PASSED suites/ssl/ssl_test.py::test_ssl_init PASSED suites/ssl/ssl_test.py::test_ssl_ PASSED suites/syntax_plugin/syntax_test.py::test_syntax_init PASSED suites/syntax_plugin/syntax_test.py::test_syntax_ PASSED suites/usn_plugin/usn_test.py::test_usn_init PASSED suites/usn_plugin/usn_test.py::test_usn_ PASSED suites/views_plugin/views_test.py::test_views_init PASSED suites/views_plugin/views_test.py::test_views_ PASSED suites/vlv/vlv_test.py::test_vlv_init PASSED suites/vlv/vlv_test.py::test_vlv_ PASSED suites/whoami_plugin/whoami_test.py::test_whoami_init PASSED suites/whoam
i_plugin/whoami_test.py::test_whoami_ PASSED ==================================== ERRORS ==================================== ______________________ ERROR at setup of test_ticket48916 ______________________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_ticket48916'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""Create' Replication 'Deployment"""' '#' Creating master 1... if DEBUGGING: master1 = 'DirSrv(verbose=True)' else: master1 = 'DirSrv(verbose=False)' 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 'args_instance[SER_CREATION_SUFFIX]' = DEFAULT_SUFFIX args_master = 'args_instance.copy()' 'master1.allocate(args_master)' instance_master1 = 'master1.exists()' if instance_master1: 'master1.delete()' 'master1.create()' 'master1.open()' 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' '#' Creating master 2... if DEBUGGING: master2 = 'DirSrv(verbose=True)' else: master2 = 'DirSrv(verbose=False)' 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 'args_instance[SER_CREATION_SUFFIX]' = DEFAULT_SUFFIX args_master = 'args_instance.copy()' 'master2.allocate(args_master)' instance_master2 = 'master2.exists()' if instance_master2: 'master2.delete()' 'master2.create()' 'master2.open()' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' '#' Create all the agreements '#' '#' Creating agreement from master 1 to master 2 properties = '{RA_NAME:' ''\''meTo_'\''' + master2.host + ''\'':'\''' + 'str(master2.port),' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' m1_m2_agmt = 'master1.agreement.create(suffix=SUFFIX,' h
ost=master2.host, port=master2.port, 'properties=properties)' if not m1_m2_agmt: 'log.fatal("Fail' to create a master '->' master replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'm1_m2_agmt)' '#' Creating agreement from master 2 to master 1 properties = '{RA_NAME:' ''\''meTo_'\''' + master1.host + ''\'':'\''' + 'str(master1.port),' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' m2_m1_agmt = 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' if not m2_m1_agmt: 'log.fatal("Fail' to create a master '->' master replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'm2_m1_agmt)' '#' Allow the replicas to get situated with the new agreements... 'time.sleep(5)' '#' '#' Initialize all the agreements '#' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' 'master1.waitForReplInit(m1_m2_agmt)' '#' Check replication is working... if 'master1.testReplication(DEFAULT_SUFFIX,' 'master2):' 'log.info('\''Replication' is 'working.'\'')' else: 'log.fatal('\''Replication' is not 'working.'\'')' '>' assert False E assert False <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48916_test.py>:116: AssertionError ---------------------------- Captured stdout setup ----------------------------- OK group dirsrv exists OK user dirsrv exists OK group dirsrv exists OK user dirsrv exists '('\''Update' succeeded: status ''\'',' ''\''0' Total update 'succeeded'\'')' ---------------------------- Captured stderr setup ----------------------------- INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Found entry dn: cn=replrepl,cn=config cn: bind dn pseudo user cn: replrepl objectClass: top objectClass: person sn: bind dn pseudo user userPassword: '{SSHA512}q7nvw6jOxgHn8newYgDvaqOo8Q2gIlTAWxkQo15O
ZF1SY6mRD9RMglTwVOF9kTPYfoVztiHR1dBQy1OT3JiSe3UtbMHVeGkI' INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Found entry dn: cn=replrepl,cn=config cn: bind dn pseudo user cn: replrepl objectClass: top objectClass: person sn: bind dn pseudo user userPassword: '{SSHA512}M95yCD/bO/csc8UcPdQaLbIAKHVv6aUDSOluZNoE7sjM41rHjAtplXEfx7EZ3++wNJPNLEGjLcYvRt03SGaAxiVbkndDQtPb' INFO:lib389:Starting total init 'cn=meTo_localhost.localdomain:38942,cn=replica,cn=dc\3Dexample\2Cdc\3Dcom,cn=mapping' tree,cn=config 'CRITICAL:lib389:testReplication()' failed to modify '(dc=example,dc=com),' error '({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"})' CRITICAL:tickets.ticket48916_test:Replication is not working. _________ ERROR at setup of 'test_aci_attr_subtype_targetattr[lang-ja]' __________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_mast
er1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-08
1.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7
fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ---------------------------- Captured stdout setup ----------------------------- Instance slapd-master_1 removed. Instance slapd-master_2 removed. OK group dirsrv exists OK user dirsrv exists OK group dirsrv exists OK user dirsrv exists ---------------------------- Captured stderr setup ----------------------------- INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Found entry dn: cn=replrepl,cn=config cn: bind dn pseudo user cn: replrepl objectClass: top objectClass: person sn: bind dn pseudo user userPassword: '{SSHA512}CpXNtBaOtj2EfWZS8ofFNVaoXZoSm/C9kng/Nnj2odQdbD6DnakmpHkNRye3V8WUJuPOq80FTmoVar7FPXLFG738x/lvSYTA' INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Found entry dn: cn=replrepl,cn=config cn: bind dn pseudo user cn: replrepl objectClass: top objectClass: person sn: bind dn pseudo user userPassword: '{SSHA512}7+rYVDUDoyNWc6ZB+gOBvvNhwNQtK8ZF0hLxtc3wuo8rQnSKLsyYNamt4ZVmCFJwOytdCh/78zeYSHr3la1MpY9LE6rzs0PJ' 'DEBUG:suites.acl.acl_test:cn=meTo_$host:$port,cn=replica,cn=dc\3Dexample\2Cdc\3Dcom,cn=mapping' tree,cn=config created INFO:lib389:Starting total init 'cn=meTo_$host:$port,cn=replica,cn=dc\3Dexamp
le\2Cdc\3Dcom,cn=mapping' tree,cn=config __________ ERROR at setup of 'test_aci_attr_subtype_targetattr[binary]' __________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'de
faultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/li
b64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_
message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _________ ERROR at setup of 'test_aci_attr_subtype_targetattr[phonetic]' _________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,
' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <htt
p://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,fun
c.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _________________ ERROR at setup of test_mode_default_add_deny _________________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable repl
ication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <h
ttp://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __de
bug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_mode_default_delete_deny ________________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Cr
eate the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.i
dm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_
call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of 'test_moddn_staging_prod[0-cn=staged' 'user,dc=example,dc=com-cn=accounts,dc=example,dc=com-False]' request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if
 it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(ag
mtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result
4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of 'test_moddn_staging_prod[1-cn=staged' 'user,dc=example,dc=com-cn=accounts,dc=example,dc=com-False]' request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[S
ER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in 
inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of 'test_moddn_staging_prod[2-cn=staged' 'user,dc=example,dc=com-cn=bad*,dc=example,dc=com-True]' request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'D
irSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(r
epl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_c
all(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of 'test_moddn_staging_prod[3-cn=st*,dc=example,dc=com-cn=accounts,dc=example,dc=com-False]' request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PO
RT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2
.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/3
89-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of 'test_moddn_staging_prod[4-cn=bad*,dc=example,dc=com-cn=accounts,dc=example,dc=com-True]' request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=Fals
e)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_
BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner retu
rn 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of 'test_moddn_staging_prod[5-cn=st*,dc=example,dc=com-cn=ac*,dc=example,dc=com-False]' request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.'
 The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to c
reate a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, r
esp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of 'test_moddn_staging_prod[6-None-cn=ac*,dc=example,dc=com-False]' request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_ta
rgetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 
repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/sour
ce/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERRO
R at setup of 'test_moddn_staging_prod[7-cn=st*,dc=example,dc=com-None-False]' request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BIN
DPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.p
y:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result =
 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _________ ERROR at setup of 'test_moddn_staging_prod[8-None-None-False]' _________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICA
ID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com
:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))
' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _________________ ERROR at setup of test_moddn_staging_prod_9 __________________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(su
ffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.c
om:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trac
e_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN __________________ ERROR at setup of test_moddn_prod_staging ___________________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master
1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIG
HTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrappe
r method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN __________________ ERROR at setup of test_check_repl_M2_to_M1 __________________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exist
s()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/w
s/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{
},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_moddn_staging_prod_except _______________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' '
master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/li
b389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ sel
f = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_mode_default_ger_no_moddn _______________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instan
ce[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _
 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY
/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ______________ ERROR at setup of test_mode_default_ger_with_moddn ______________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployemen
t master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.wai
tForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result =
 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _____________ ERROR at setup of test_mode_switch_default_to_legacy _____________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 '
args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.cr
eate(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY
/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_mode_legacy_ger_no_moddn1 _______________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_insta
nce[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties
[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/pyt
hon2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_mode_legacy_ger_no_moddn2 _______________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' globa
l installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'c
reated"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <htt
p://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ______________ ERROR at setup of test_mode_legacy_ger_with_moddn _______________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def '
topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=
master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '*
*kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ___________________ ERROR at setup of test_rdn_write_get_ger _______
____________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'default
Properties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(
msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t
' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ______________ ERROR at setup of test_rdn_write_modrdn_anonymous _______________ request = '<SubRequest' ''\''topology'\''' for '<Function' ''\''test_aci_attr_subtype_targetattr[lang-ja]'\''>>' '@pytest.fixture(scope="module")' def 'topology(request):' '"""This' fixture is used to create a replicated topology for the ''\''module'\''.' The replicated topology is MASTER1 '<->' Master2. '"""' global installation1_prefix global installation2_prefix '#' allocate master1 on a given deployement master1 = 'DirSrv(verbose=False)' if installation1_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation1_prefix '#' Args for the master1 instance 'args_instance[SER_HOST]' = HOST_MASTER_1 'args_instance[SER_PORT]' = PORT_MASTER_1 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_1 args_master = 'args_instance.copy()' 'master1.allocate(args_master)' '#' allocate master1 on a given deployement master2 = 'DirSrv(verbose=False)' if installation2_prefix: 'args_instance[SER_DEPLOYED_DIR]' = installation2_prefix '#' Args for the consumer instance 'args_instance[SER_HOST]' = HOST_MASTER_2 'args_instance[SER_PORT]' = PORT_MASTER_2 'args_instance[SER_SERVERID_PROP]' = SERVERID_MASTER_2 args_master = 'args_instance.copy()' 'master2.allocate(args_master)' '#' Get the status of the instance and restart it if it exists instance_master1 = 'master1.exists()' instance_master2 = 'master2.exists()' '#' Remove all the instances if instance_master1: 'master1.delete()' if instance_master2: 'master2.delete()' '#' Create the instances 'master1.create()' 'master1.open()' 'master2.create()' 'master2.open()' '#' '#' Now prepare the Master-Consumer topology '#' '#' First Enable replication 'master1.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_1)' 'master2.replica.enableReplication(suffix=SUFFIX,' role=REPLICAROLE_MASTER, 'replicaId=REPLICAID_MASTER_2)' '#' Initialize the 'supplier->consumer' properties = '
{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' repl_agreement = 'master1.agreement.create(suffix=SUFFIX,' host=master2.host, port=master2.port, 'properties=properties)' if not repl_agreement: 'log.fatal("Fail' to create a replica 'agreement")' 'sys.exit(1)' 'log.debug("%s' 'created"' % 'repl_agreement)' properties = '{RA_NAME:' 'r'\''meTo_$host:$port'\'',' RA_BINDDN: 'defaultProperties[REPLICATION_BIND_DN],' RA_BINDPW: 'defaultProperties[REPLICATION_BIND_PW],' RA_METHOD: 'defaultProperties[REPLICATION_BIND_METHOD],' RA_TRANSPORT_PROT: 'defaultProperties[REPLICATION_TRANSPORT]}' 'master2.agreement.create(suffix=SUFFIX,' host=master1.host, port=master1.port, 'properties=properties)' 'master1.agreement.init(SUFFIX,' HOST_MASTER_2, 'PORT_MASTER_2)' '>' 'master1.waitForReplInit(repl_agreement)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return 'self.replica.wait_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = 'self.check_init(agmtdn)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, '"(objectclass=*)",' 'attrlist)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = 'self.result(res)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: i
n inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f164fd40>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa977418>' args = '(17,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self.
_trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN =================================== FAILURES =================================== ____________________________ test_ticket47669_init _____________________________ topology = '<tickets.ticket47669_test.TopologyStandalone' object at '0x7fa7fa745990>' def 'test_ticket47669_init(topology):' '"""' Add cn=changelog5,cn=config Enable cn=Retro Changelog Plugin,cn=plugins,cn=config '"""' 'log.info('\''Testing' Ticket 47669 - Test duration syntax in the 'changelogs'\'')' '#' bind as directory manager 'topology.standalone.log.info("Bind' as '%s"' % 'DN_DM)' 'topology.standalone.simple_bind_s(DN_DM,' 'PASSWORD)' try: changelogdir = '"%s/changelog"' % topology.standalone.dbdir 'topology.standalone.add_s(Entry((CHANGELOG,' '{'\''objectclass'\'':' ''\''top' 'extensibleObject'\''.split(),' ''\''nsslapd-changelogdir'\'':' 'changelogdir})))' except ldap.LDAPError as e: 'log.error('\''Failed' to add \' + CHANGELOG + ''\'':' error \' + 'e.message['\''desc'\''])' assert False try: 'topology.standalone.modify_s(RETROCHANGELOG,' '[(ldap.MOD_REPLACE,' ''\''nsslapd-pluginEnabled'\'',' ''\''on'\'')])' except ldap.LDAPError as e: 'log.error('\''Failed' to enable \' + RETROCHANGELOG + ''\'':' error \' + 'e.message['\''desc'\''])' assert False '#' restart the server '>' 'topology.standalone.restart(timeout=10)' tickets/ticket47669_test.py:103: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:1215: in restart 'self.start(timeout)' ../../../lib389/lib389/__init__.py:1096: in start '"dirsrv@%s"' % 'self.serverid])' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ popenargs = '(['\''/usr/bin/systemctl'\'',' ''\''start'\'',' ''\''dirsrv@standalone'\''],),' kwargs = '{}' retcode = 1, cmd = '['\''
/usr/bin/systemctl'\'',' ''\''start'\'',' ''\''dirsrv@standalone'\'']' def 'check_call(*popenargs,' '**kwargs):' '"""Run' command with arguments. Wait for command to complete. If the exit code was zero then return, otherwise raise CalledProcessError. The CalledProcessError object will have the return code in the returncode attribute. The arguments are the same as for the Popen constructor. Example: 'check_call(["ls",' '"-l"])' '"""' retcode = 'call(*popenargs,' '**kwargs)' if retcode: cmd = 'kwargs.get("args")' if cmd is None: cmd = 'popenargs[0]' '>' raise 'CalledProcessError(retcode,' 'cmd)' E CalledProcessError: Command ''\''['\''/usr/bin/systemctl'\'',' ''\''start'\'',' ''\''dirsrv@standalone'\'']'\''' returned non-zero exit status 1 /usr/lib64/python2.7/subprocess.py:541: CalledProcessError ---------------------------- Captured stdout setup ----------------------------- OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:Testing Ticket 47669 - Test duration syntax in the changelogs INFO:lib389:Bind as cn=Directory Manager Job for dirsrv@standalone.service failed because the control process exited with error code. See '"systemctl' status 'dirsrv@standalone.service"' and '"journalctl' '-xe"' for details. ______________________ test_ticket47669_changelog_maxage _______________________ topology = '<tickets.ticket47669_test.TopologyStandalone' object at '0x7fa7fa745990>' def 'test_ticket47669_changelog_maxage(topology):' '"""' Test nsslapd-changelogmaxage in cn=changelog5,cn=config '"""' 'log.info('\''1.' Test nsslapd-changelogmaxage in 'cn=changelog5,cn=config'\'')' '#' bind as directory manager 'topology.standalone.log.info("Bind' as '%s"' % 'DN_DM)' '>' 'topology.standalone.simple_bind_s(DN_DM,' 'PASSWORD)' tickets/ticket47669_test.py:159: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/p
ython2.7/site-packages/ldap/ldapobject.py:223: in simple_bind_s resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all=1,timeout=self.timeout)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f9dbb5a8>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa8c2e68>' args = '(13,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:1. Test nsslapd-changelogmaxage in cn=changelog5,cn=config INFO:lib389:Bind as cn=Directory Manager ___________________ test_ticket47669_changelog_triminterval ____________________ topol
ogy = '<tickets.ticket47669_test.TopologyStandalone' object at '0x7fa7fa745990>' def 'test_ticket47669_changelog_triminterval(topology):' '"""' Test nsslapd-changelogtrim-interval in cn=changelog5,cn=config '"""' 'log.info('\''2.' Test nsslapd-changelogtrim-interval in 'cn=changelog5,cn=config'\'')' '#' bind as directory manager 'topology.standalone.log.info("Bind' as '%s"' % 'DN_DM)' '>' 'topology.standalone.simple_bind_s(DN_DM,' 'PASSWORD)' tickets/ticket47669_test.py:179: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:222: in simple_bind_s msgid = 'self.simple_bind(who,cred,serverctrls,clientctrls)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:216: in simple_bind return 'self._ldap_call(self._l.simple_bind,who,cred,RequestControlTuples(serverctrls),RequestControlTuples(clientctrls))' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f9dbb5a8>' func = '<built-in' method simple_bind of LDAP object at '0x7fa7fa8c2e68>' args = '('\''cn=Directory' 'Manager'\'',' ''\''password'\'',' None, 'None),' kwargs = '{}' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_messa
ge_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:2. Test nsslapd-changelogtrim-interval in cn=changelog5,cn=config INFO:lib389:Bind as cn=Directory Manager _________________ test_ticket47669_changelog_compactdbinterval _________________ topology = '<tickets.ticket47669_test.TopologyStandalone' object at '0x7fa7fa745990>' def 'test_ticket47669_changelog_compactdbinterval(topology):' '"""' Test nsslapd-changelogcompactdb-interval in cn=changelog5,cn=config '"""' 'log.info('\''3.' Test nsslapd-changelogcompactdb-interval in 'cn=changelog5,cn=config'\'')' '#' bind as directory manager 'topology.standalone.log.info("Bind' as '%s"' % 'DN_DM)' '>' 'topology.standalone.simple_bind_s(DN_DM,' 'PASSWORD)' tickets/ticket47669_test.py:199: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:222: in simple_bind_s msgid = 'self.simple_bind(who,cred,serverctrls,clientctrls)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:216: in simple_bind return 'self._ldap_call(self._l.simple_bind,who,cred,RequestControlTuples(serverctrls),RequestControlTuples(clientctrls))' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f9dbb5a8>' func = '<built-in' method simple_bind of LDAP object at '0x7fa7fa8c2e68>' args = '('\''cn=Directory' 'Manager'\'',' ''\''password'\'',' None, 'None),' kwargs = '{}' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'ser
ver"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:3. Test nsslapd-changelogcompactdb-interval in cn=changelog5,cn=config INFO:lib389:Bind as cn=Directory Manager ____________________ test_ticket47669_retrochangelog_maxage ____________________ topology = '<tickets.ticket47669_test.TopologyStandalone' object at '0x7fa7fa745990>' def 'test_ticket47669_retrochangelog_maxage(topology):' '"""' Test nsslapd-changelogmaxage in cn=Retro Changelog Plugin,cn=plugins,cn=config '"""' 'log.info('\''4.' Test nsslapd-changelogmaxage in cn=Retro Changelog 'Plugin,cn=plugins,cn=config'\'')' '#' bind as directory manager 'topology.standalone.log.info("Bind' as '%s"' % 'DN_DM)' '>' 'topology.standalone.simple_bind_s(DN_DM,' 'PASSWORD)' tickets/ticket47669_test.py:219: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:222: in simple_bind_s msgid = 'self.simple_bind(who,cred,serverctrls,clientctrls)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:216: in simple_bind return 'self._ldap_call(self._l.simple_bind,who,cred,R
equestControlTuples(serverctrls),RequestControlTuples(clientctrls))' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f9dbb5a8>' func = '<built-in' method simple_bind of LDAP object at '0x7fa7fa8c2e68>' args = '('\''cn=Directory' 'Manager'\'',' ''\''password'\'',' None, 'None),' kwargs = '{}' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:4. Test nsslapd-changelogmaxage in cn=Retro Changelog Plugin,cn=plugins,cn=config INFO:lib389:Bind as cn=Directory Manager ______________________ test_ticket47823_invalid_config_1 _______________________ topology = '<tickets.ticket47823_test.TopologyStandalone' object at '0x7fa7f9e8f910>' def 'test_ticket47823_invalid_config_1(topology):' ''\'''\'''\''' Check that an invalid config is detected. No uniqueness enforced Using old config: arg0 is missing ''\'''\'''\''' '_header(topology,' '"Invalid' config '(old):' arg0 is 'missing")' '_config_file(topology,' 'action='\''save'\'')' '#' create an invalid config without arg0 config = '_build_config(topology,' 'attr_n
ame='\''cn'\'',' subtree_1=ACTIVE_DN, subtree_2=None, 'type_config='\''old'\'',' 'across_subtrees=False)' del 'config.data['\''nsslapd-pluginarg0'\'']' '#' replace ''\''cn'\''' uniqueness entry try: 'topology.standalone.delete_s(config.dn)' except ldap.NO_SUCH_OBJECT: pass 'topology.standalone.add_s(config)' 'topology.standalone.getEntry(config.dn,' ldap.SCOPE_BASE, '"(objectclass=nsSlapdPlugin)",' 'ALL_CONFIG_ATTRS)' '#' Check the server did not restart 'topology.standalone.modify_s(DN_CONFIG,' '[(ldap.MOD_REPLACE,' ''\''nsslapd-errorlog-level'\'',' ''\''65536'\'')])' try: '>' 'topology.standalone.restart(timeout=5)' tickets/ticket47823_test.py:636: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:1215: in restart 'self.start(timeout)' ../../../lib389/lib389/__init__.py:1096: in start '"dirsrv@%s"' % 'self.serverid])' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ popenargs = '(['\''/usr/bin/systemctl'\'',' ''\''start'\'',' ''\''dirsrv@standalone'\''],),' kwargs = '{}' retcode = 1, cmd = '['\''/usr/bin/systemctl'\'',' ''\''start'\'',' ''\''dirsrv@standalone'\'']' def 'check_call(*popenargs,' '**kwargs):' '"""Run' command with arguments. Wait for command to complete. If the exit code was zero then return, otherwise raise CalledProcessError. The CalledProcessError object will have the return code in the returncode attribute. The arguments are the same as for the Popen constructor. Example: 'check_call(["ls",' '"-l"])' '"""' retcode = 'call(*popenargs,' '**kwargs)' if retcode: cmd = 'kwargs.get("args")' if cmd is None: cmd = 'popenargs[0]' '>' raise 'CalledProcessError(retcode,' 'cmd)' E CalledProcessError: Command ''\''['\''/usr/bin/systemctl'\'',' ''\''start'\'',' ''\''dirsrv@standalone'\'']'\''' returned non-zero exit status 1 /usr/lib64/python2.7/subprocess.py:541: CalledProcessError ----------------------------- Captured stderr call ----------------------------- INFO:lib389: '#####################################
##########' INFO:lib389:####### INFO:lib389:####### Invalid config '(old):' arg0 is missing INFO:lib389:####### INFO:lib389:############################################### Job for dirsrv@standalone.service failed because the control process exited with error code. See '"systemctl' status 'dirsrv@standalone.service"' and '"journalctl' '-xe"' for details. ______________________ test_ticket47823_invalid_config_2 _______________________ topology = '<tickets.ticket47823_test.TopologyStandalone' object at '0x7fa7f9e8f910>' def 'test_ticket47823_invalid_config_2(topology):' ''\'''\'''\''' Check that an invalid config is detected. No uniqueness enforced Using old config: arg1 is missing ''\'''\'''\''' '_header(topology,' '"Invalid' config '(old):' arg1 is 'missing")' '_config_file(topology,' 'action='\''save'\'')' '#' create an invalid config without arg0 '>' config = '_build_config(topology,' 'attr_name='\''cn'\'',' subtree_1=ACTIVE_DN, subtree_2=None, 'type_config='\''old'\'',' 'across_subtrees=False)' tickets/ticket47823_test.py:672: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:124: in _build_config config = '_uniqueness_config_entry(topology,' 'attr_name)' tickets/ticket47823_test.py:112: in _uniqueness_config_entry ''\''nsslapd-pluginDescription'\''])' ../../../lib389/lib389/__init__.py:1574: in getEntry restype, obj = 'self.result(res)' ../../../lib389/lib389/__init__.py:127: in inner objtype, data = 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = 'self.result2(msgid,all,timeout)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = 'self.result3(msgid,all,timeout)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_
ctrl_classes ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = 'self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f969eea8>' func = '<built-in' method result4 of LDAP object at '0x7fa7fa8b4800>' args = '(15,' 1, -1, 0, 0, '0),' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: '###############################################' INFO:lib389:####### INFO:lib389:####### Invalid config '(old):' arg1 is missing INFO:lib389:####### INFO:lib389:############################################### ______________________ test_ticket47823_invalid_config_3 _______________________ topology = '<tickets.ticket47823_test.TopologyStandalone' object at '0x7fa7f9e8f910>' def 'test_ticket47823_invalid_config_3(topology):' ''\'''\'''\''' Check that an invalid config is detected. No uniqueness enforced Using old config: arg0 is m
issing ''\'''\'''\''' '_header(topology,' '"Invalid' config '(old):' arg0 is missing but new config attrname 'exists")' '_config_file(topology,' 'action='\''save'\'')' '#' create an invalid config without arg0 '>' config = '_build_config(topology,' 'attr_name='\''cn'\'',' subtree_1=ACTIVE_DN, subtree_2=None, 'type_config='\''old'\'',' 'across_subtrees=False)' tickets/ticket47823_test.py:723: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:124: in _build_config config = '_uniqueness_config_entry(topology,' 'attr_name)' tickets/ticket47823_test.py:112: in _uniqueness_config_entry ''\''nsslapd-pluginDescription'\''])' ../../../lib389/lib389/__init__.py:1573: in getEntry res = 'self.search(*args,' '**kwargs)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:594: in search return 'self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f969eea8>' func = '<built-in' method search_ext of LDAP object at '0x7fa7fa8b4800>' args = '('\''cn=attribute' 'uniqueness,cn=plugins,cn=config'\'',' 0, ''\''(objectclass=nsSlapdPlugin)'\'',' '['\''objectClass'\'',' ''\''cn'\'',' ''\''nsslapd-pluginPath'\'',' ''\''nsslapd-pluginInitfunc'\'',' ''\''nsslapd-pluginType'\'',' ''\''nsslapd-pluginEnabled'\'',' '...],' 0, None, '...)' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __d
ebug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: '###############################################' INFO:lib389:####### INFO:lib389:####### Invalid config '(old):' arg0 is missing but new config attrname exists INFO:lib389:####### INFO:lib389:############################################### ______________________ test_ticket47823_invalid_config_4 _______________________ topology = '<tickets.ticket47823_test.TopologyStandalone' object at '0x7fa7f9e8f910>' def 'test_ticket47823_invalid_config_4(topology):' ''\'''\'''\''' Check that an invalid config is detected. No uniqueness enforced Using old config: arg1 is missing ''\'''\'''\''' '_header(topology,' '"Invalid' config '(old):' arg1 is missing but new config 'exist")' '_config_file(topology,' 'action='\''save'\'')' '#' create an invalid config without arg0 '>' config = '_build_config(topology,' 'attr_name='\''cn'\'',' subtree_1=ACTIVE_DN, subtree_2=None, 'type_config='\''old'\'',' 'across_subtrees=False)' tickets/ticket47823_test.py:776: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:124: in _build_config config = '_uniqueness_config_entry(topology,' 'attr_name)' tickets/ticket47823_test.py:112: in _uniqueness_config_entry ''\''nsslapd-pluginDescription'\''])' ../../../lib389/lib389/__init__.py:1573: in getEntry res = 'self.search(*args,' '**kwargs)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /us
r/lib64/python2.7/site-packages/ldap/ldapobject.py:594: in search return 'self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f969eea8>' func = '<built-in' method search_ext of LDAP object at '0x7fa7fa8b4800>' args = '('\''cn=attribute' 'uniqueness,cn=plugins,cn=config'\'',' 0, ''\''(objectclass=nsSlapdPlugin)'\'',' '['\''objectClass'\'',' ''\''cn'\'',' ''\''nsslapd-pluginPath'\'',' ''\''nsslapd-pluginInitfunc'\'',' ''\''nsslapd-pluginType'\'',' ''\''nsslapd-pluginEnabled'\'',' '...],' 0, None, '...)' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: '###############################################' INFO:lib389:####### INFO:lib389:####### Invalid config '(old):' arg1 is missing but new config exist INFO:lib389:####### INFO:lib389:#######################################
######## ______________________ test_ticket47823_invalid_config_5 _______________________ topology = '<tickets.ticket47823_test.TopologyStandalone' object at '0x7fa7f9e8f910>' def 'test_ticket47823_invalid_config_5(topology):' ''\'''\'''\''' Check that an invalid config is detected. No uniqueness enforced Using new config: uniqueness-attribute-name is missing ''\'''\'''\''' '_header(topology,' '"Invalid' config '(new):' uniqueness-attribute-name is 'missing")' '_config_file(topology,' 'action='\''save'\'')' '#' create an invalid config without arg0 '>' config = '_build_config(topology,' 'attr_name='\''cn'\'',' subtree_1=ACTIVE_DN, subtree_2=None, 'type_config='\''new'\'',' 'across_subtrees=False)' tickets/ticket47823_test.py:828: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:131: in _build_config config = '_uniqueness_config_entry(topology,' 'attr_name)' tickets/ticket47823_test.py:112: in _uniqueness_config_entry ''\''nsslapd-pluginDescription'\''])' ../../../lib389/lib389/__init__.py:1573: in getEntry res = 'self.search(*args,' '**kwargs)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:594: in search return 'self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f969eea8>' func = '<built-in' method search_ext of LDAP object at '0x7fa7fa8b4800>' args = '('\''cn=attribute' 'uniqueness,cn=plugins,cn=config'\'',' 0, ''\''(objectclass=nsSlapdPlugin)'\'',' '['\''objectClass'\'',' ''\''cn'\'',' ''\''nsslapd-pluginPath'\'',' ''\''nsslapd-pluginInitfunc'\'',' ''\''nsslapd-pluginType'\'',' ''\''ns
slapd-pluginEnabled'\'',' '...],' 0, None, '...)' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: '###############################################' INFO:lib389:####### INFO:lib389:####### Invalid config '(new):' uniqueness-attribute-name is missing INFO:lib389:####### INFO:lib389:############################################### ______________________ test_ticket47823_invalid_config_6 _______________________ topology = '<tickets.ticket47823_test.TopologyStandalone' object at '0x7fa7f9e8f910>' def 'test_ticket47823_invalid_config_6(topology):' ''\'''\'''\''' Check that an invalid config is detected. No uniqueness enforced Using new config: uniqueness-subtrees is missing ''\'''\'''\''' '_header(topology,' '"Invalid' config '(new):' uniqueness-subtrees is 'missing")' '_config_file(topology,' 'action='\''save'\'')' '#' create an invalid config without arg0 '>' config = '_build_config(topology,' 'attr_name='\''cn'\'',' subtree_1=ACTIVE_DN, subtree_2=None, 'type_config='\''new'\'',' 'across_subtrees=False)' tickets/ticket47823_test.py:879: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:131: in _b
uild_config config = '_uniqueness_config_entry(topology,' 'attr_name)' tickets/ticket47823_test.py:112: in _uniqueness_config_entry ''\''nsslapd-pluginDescription'\''])' ../../../lib389/lib389/__init__.py:1573: in getEntry res = 'self.search(*args,' '**kwargs)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:594: in search return 'self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f969eea8>' func = '<built-in' method search_ext of LDAP object at '0x7fa7fa8b4800>' args = '('\''cn=attribute' 'uniqueness,cn=plugins,cn=config'\'',' 0, ''\''(objectclass=nsSlapdPlugin)'\'',' '['\''objectClass'\'',' ''\''cn'\'',' ''\''nsslapd-pluginPath'\'',' ''\''nsslapd-pluginInitfunc'\'',' ''\''nsslapd-pluginType'\'',' ''\''nsslapd-pluginEnabled'\'',' '...],' 0, None, '...)' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/lda
pobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: '###############################################' INFO:lib389:####### INFO:lib389:####### Invalid config '(new):' uniqueness-subtrees is missing INFO:lib389:####### INFO:lib389:############################################### ______________________ test_ticket47823_invalid_config_7 _______________________ topology = '<tickets.ticket47823_test.TopologyStandalone' object at '0x7fa7f9e8f910>' def 'test_ticket47823_invalid_config_7(topology):' ''\'''\'''\''' Check that an invalid config is detected. No uniqueness enforced Using new config: uniqueness-subtrees is missing ''\'''\'''\''' '_header(topology,' '"Invalid' config '(new):' uniqueness-subtrees are 'invalid")' '_config_file(topology,' 'action='\''save'\'')' '#' create an invalid config without arg0 '>' config = '_build_config(topology,' 'attr_name='\''cn'\'',' 'subtree_1="this_is' dummy 'DN",' 'subtree_2="an' other=dummy 'DN",' 'type_config='\''new'\'',' 'across_subtrees=False)' tickets/ticket47823_test.py:930: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:131: in _build_config config = '_uniqueness_config_entry(topology,' 'attr_name)' tickets/ticket47823_test.py:112: in _uniqueness_config_entry ''\''nsslapd-pluginDescription'\''])' ../../../lib389/lib389/__init__.py:1573: in getEntry res = 'self.search(*args,' '**kwargs)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:594: in search return 'self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None)' ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return 'f(*args,' '**kwargs)' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389
.DirSrv' instance at '0x7fa7f969eea8>' func = '<built-in' method search_ext of LDAP object at '0x7fa7fa8b4800>' args = '('\''cn=attribute' 'uniqueness,cn=plugins,cn=config'\'',' 0, ''\''(objectclass=nsSlapdPlugin)'\'',' '['\''objectClass'\'',' ''\''cn'\'',' ''\''nsslapd-pluginPath'\'',' ''\''nsslapd-pluginInitfunc'\'',' ''\''nsslapd-pluginType'\'',' ''\''nsslapd-pluginEnabled'\'',' '...],' 0, None, '...)' kwargs = '{},' diagnostic_message_success = None e = 'SERVER_DOWN({'\''desc'\'':' '"Can'\''t' contact LDAP 'server"},)' def '_ldap_call(self,func,*args,**kwargs):' '"""' Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs '"""' 'self._ldap_object_lock.acquire()' if __debug__: if 'self._trace_level>=1:' 'self._trace_file.write('\''***' %s %s - '%s\n%s\n'\''' % '(' 'repr(self),' self._uri, ''\''.'\''.join((self.__class__.__name__,func.__name__)),' 'pprint.pformat((args,kwargs))' '))' if 'self._trace_level>=9:' 'traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file)' diagnostic_message_success = None try: try: '>' result = 'func(*args,**kwargs)' E SERVER_DOWN: '{'\''desc'\'':' '"Can'\''t' contact LDAP 'server"}' /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: '###############################################' INFO:lib389:####### INFO:lib389:####### Invalid config '(new):' uniqueness-subtrees are invalid INFO:lib389:####### INFO:lib389:############################################### ______________________________ test_ticket48226_1 ______________________________ topology = '<tickets.ticket48226_test.TopologyReplication' object at '0x7fa7f9e2eb50>' def 'test_ticket48226_1(topology):' name = ''\''test_entry'\''' dn = '"cn=%s,%s"' % '(name,' 'SUFFIX)' 'topology.master1.add_s(Entry((dn,' '{'\''objectclass'\'':' '"top' 'person".split(),' ''\''sn'\'':' name, ''\''cn'\'':' 'name})))' '#' First do an update that is replicated mods = '[(ldap.MOD_ADD,' ''\''descripti
on'\'',' ''\''5'\'')]' 'topology.master1.modify_s(dn,' 'mods)' nbtry = 0 while '(nbtry' '<=' '10):' try: ent = 'topology.master2.getEntry(dn,' ldap.SCOPE_BASE, '"(objectclass=*)",' '['\''description'\''])' if 'ent.hasAttr('\''description'\'')' and 'ent.getValue('\''description'\'')' == ''\''5'\'':' break except ldap.NO_SUCH_OBJECT: pass nbtry = nbtry + 1 'time.sleep(1)' assert nbtry '<=' 10 '#' Stop M2 so that it will not receive the next update 'topology.master2.stop(10)' '#' ADD a new value that is not replicated mods = '[(ldap.MOD_DELETE,' ''\''description'\'',' ''\''5'\'')]' 'topology.master1.modify_s(dn,' 'mods)' '#' Stop M1 so that it will keep del ''\''5'\''' that is unknown from master2 'topology.master1.stop(10)' '#' Get the sbin directory so we know where to replace ''\''ns-slapd'\''' sbin_dir = 'topology.master2.get_sbin_dir()' '#' Enable valgrind if not 'topology.master2.has_asan():' 'valgrind_enable(sbin_dir)' '#' start M2 to do the next updates 'topology.master2.start()' '#' ADD ''\''description'\''' by ''\''5'\''' mods = '[(ldap.MOD_DELETE,' ''\''description'\'',' ''\''5'\'')]' 'topology.master2.modify_s(dn,' 'mods)' '#' DEL ''\''description'\''' by ''\''5'\''' mods = '[(ldap.MOD_ADD,' ''\''description'\'',' ''\''5'\'')]' 'topology.master2.modify_s(dn,' 'mods)' '#' sleep of purge delay so that the next update will purge the CSN_7 'time.sleep(6)' '#' ADD ''\''description'\''' by ''\''6'\''' that purge the state info mods = '[(ldap.MOD_ADD,' ''\''description'\'',' ''\''6'\'')]' 'topology.master2.modify_s(dn,' 'mods)' '#' Restart master1 '#topology.master1.start(30)' if not 'topology.master2.has_asan():' results_file = 'valgrind_get_results_file(topology.master2)' '#' Stop master2 'topology.master2.stop(30)' '#' Check for leak if not 'topology.master2.has_asan():' '>' if 'valgrind_check_file(results_file,' VALGRIND_LEAK_STR, ''\''csnset_dup'\''):' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48226_test.py>:220: _ _ _ _ _ _ _ _ _
 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ results_file = ''\''valgrind'\''' patterns = '('\''' blocks are definitely lost in loss record ''\'',' ''\''csnset_dup'\'')' found = False, pattern_count = 2, matched_count = 0 def 'valgrind_check_file(results_file,' '*patterns):' ''\'''\'''\''' Check the valgrind results file for the all the patterns @param result_file - valgrind results file '(must' be read after server is 'stopped)' @param patterns - A plain text or regex pattern string args that should be searched for @return True/False - Return true if one if the patterns match a stack trace @raise IOError ''\'''\'''\''' '#' Verify results file if not results_file: assert False '#' Check the result file fo the leak text results_file = 'results_file.replace('\''\n'\'',' ''\'''\'')' found = False pattern_count = 'len(patterns)' matched_count = 0 '>' vlog = 'open(results_file)' E IOError: '[Errno' '2]' No such file or directory: ''\''valgrind'\''' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/utils.py>:354: IOError ----------------------------- Captured stderr call ----------------------------- INFO:lib389.utils:Valgrind is already enabled. ____________________ test_ticket48266_count_csn_evaluation _____________________ topology = '<tickets.ticket48266_test.TopologyReplication' object at '0x7fa7f9e086d0>' entries = None def 'test_ticket48266_count_csn_evaluation(topology,' 'entries):' ents = 'topology.master1.agreement.list(suffix=SUFFIX)' assert 'len(ents)' == 1 '>' first_csn = '_get_first_not_replicated_csn(topology)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48266_test.py>:328: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ topology = '<tickets.ticket48266_test.TopologyReplication' object at '0x7fa7f9e086d0>' def '_get_first_not_replicated_csn(topology):' name = '"cn=%s2,%s"' % '(NEW_ACCOUNT,' 'SUFFIX)' '#' read the first CSN that will n
ot be replicated mod = '[(ldap.MOD_REPLACE,' ''\''telephonenumber'\'',' 'str(123456))]' 'topology.master1.modify_s(name,' 'mod)' msgid = 'topology.master1.search_ext(name,' ldap.SCOPE_SUBTREE, ''\''objectclass=*'\'',' '['\''nscpentrywsi'\''])' rtype, rdata, rmsgid = 'topology.master1.result2(msgid)' attrs = None for dn, raw_attrs in rdata: 'topology.master1.log.info("dn:' '%s"' % 'dn)' if ''\''nscpentrywsi'\''' in raw_attrs: attrs = 'raw_attrs['\''nscpentrywsi'\'']' assert attrs for attr in attrs: if 'attr.lower().startswith('\''telephonenumber'\''):' break assert attr '#' now retrieve the CSN of the operation we are looking for csn = None 'topology.master1.stop(timeout=10)' file_path = 'os.path.join(topology.master1.prefix,' '"var/log/dirsrv/slapd-%s/access"' % 'topology.master1.serverid)' '>' file_obj = 'open(file_path,' '"r")' E IOError: '[Errno' '2]' No such file or directory: ''\''/usr/var/log/dirsrv/slapd-master_1/access'\''' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48266_test.py>:276: IOError ----------------------------- Captured stderr call ----------------------------- INFO:lib389:dn: cn=new_account2,dc=example,dc=com __________________ test_ticket48270_homeDirectory_indexed_cis __________________ topology = '<tickets.ticket48270_test.TopologyStandalone' object at '0x7fa7f0f75f90>' def 'test_ticket48270_homeDirectory_indexed_cis(topology):' 'log.info("\n\nindex' homeDirectory in caseIgnoreIA5Match and 'caseExactIA5Match")' try: ent = 'topology.standalone.getEntry(HOMEDIRECTORY_INDEX,' 'ldap.SCOPE_BASE)' except ldap.NO_SUCH_OBJECT: 'topology.standalone.add_s(Entry((HOMEDIRECTORY_INDEX,' '{' ''\''objectclass'\'':' '"top' 'nsIndex".split(),' ''\''cn'\'':' HOMEDIRECTORY_CN, ''\''nsSystemIndex'\'':' ''\''false'\'',' ''\''nsIndexType'\'':' ''\''eq'\''})))' '#log.info("attach' 'debugger")' '#time.sleep(60)' 'IGNORE_MR_NAME='\''caseIgnoreIA5Match'\''' 'EXACT_MR_NAME='\''caseExactIA5Match'\''' mod = '[(ldap.MOD_REPLACE,' MATCHINGRULE, 
'(IGNORE_MR_NAME,' 'EXACT_MR_NAME))]' 'topology.standalone.modify_s(HOMEDIRECTORY_INDEX,' 'mod)' '#topology.standalone.stop(timeout=10)' 'log.info("successfully' checked that filter with exact mr , a filter with lowercase eq is 'failing")' '#assert' 'topology.standalone.db2index(bename=DEFAULT_BENAME,' suffixes=None, 'attrs=['\''homeDirectory'\''])' '#topology.standalone.start(timeout=10)' args = '{TASK_WAIT:' 'True}' 'topology.standalone.tasks.reindex(suffix=SUFFIX,' 'attrname='\''homeDirectory'\'',' 'args=args)' 'log.info("Check' indexing succeeded with a specified matching 'rule")' file_path = 'os.path.join(topology.standalone.prefix,' '"var/log/dirsrv/slapd-%s/errors"' % 'topology.standalone.serverid)' '>' file_obj = 'open(file_path,' '"r")' E IOError: '[Errno' '2]' No such file or directory: ''\''/usr/var/log/dirsrv/slapd-standalone/errors'\''' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48270_test.py>:100: IOError ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket48270_test: index homeDirectory in caseIgnoreIA5Match and caseExactIA5Match INFO:tickets.ticket48270_test:successfully checked that filter with exact mr , a filter with lowercase eq is failing INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_013221 completed successfully INFO:tickets.ticket48270_test:Check indexing succeeded with a specified matching rule _______________________________ test_ticket48383 _______________________________ topology = '<tickets.ticket48383_test.TopologyStandalone' object at '0x7fa7f0f9d450>' def 'test_ticket48383(topology):' '"""' This test case will check that we re-alloc buffer sizes on import.c We achieve this by setting the servers dbcachesize to a stupid small value and adding huge objects to ds. Then when we run db2index, either: data stress suites tickets tmp If we are not using the re-alloc code, it will FAIL '(Bad)' data stress suit
es tickets tmp If we re-alloc properly, it all works regardless. '"""' 'topology.standalone.config.set('\''nsslapd-maxbersize'\'',' ''\''200000000'\'')' 'topology.standalone.restart()' '#' Create some stupid huge objects / attributes in DS. '#' seeAlso is indexed by default. Lets do 'that!' '#' This will take a while ... data = '[random.choice(string.letters)' for x in 'xrange(10000000)]' s = '"".join(data)' '#' This was here for an iteration test. i = 1 USER_DN = ''\''uid=user%s,ou=people,%s'\''' % '(i,' 'DEFAULT_SUFFIX)' padding = '['\''%s'\''' % n for n in 'range(400)]' user = 'Entry((USER_DN,' '{' ''\''objectclass'\'':' ''\''top' posixAccount person 'extensibleObject'\''.split(),' ''\''uid'\'':' ''\''user%s'\''' % '(i),' ''\''cn'\'':' ''\''user%s'\''' % '(i),' ''\''uidNumber'\'':' ''\''%s'\''' % '(i),' ''\''gidNumber'\'':' ''\''%s'\''' % '(i),' ''\''homeDirectory'\'':' ''\''/home/user%s'\''' % '(i),' ''\''description'\'':' ''\''user' 'description'\'',' ''\''sn'\''' : s , ''\''padding'\''' : padding , '}))' try: 'topology.standalone.add_s(user)' except ldap.LDAPError as e: 'log.fatal('\''test' 48383: Failed to user%s: error %s \' % '(i,' 'e.message['\''desc'\'']))' assert False '#' Set the dbsize really low. try: 'topology.standalone.modify_s(DEFAULT_BENAME,' '[(ldap.MOD_REPLACE,' ''\''nsslapd-cachememsize'\'',' ''\''1'\'')])' except ldap.LDAPError as e: 'log.fatal('\''Failed' to change nsslapd-cachememsize \' + 'e.message['\''desc'\''])' '##' Does ds try and set a minimum possible value for 'this?' '##' Yes: '[16/Feb/2016:16:39:18' '+1000]' - WARNING: cache too small, increasing to 500K bytes '#' Given the formula, by default, this means DS will make the buffsize 400k '#' So an object with a 1MB attribute should break indexing '#' stop the server 'topology.standalone.stop(timeout=30)' '#' Now export and import the DB. 'It'\''s' easier than db2index ... 'topology.standalone.db2ldif(bename=DEFAULT_BENAME,' 'suffixes=[DEFAULT_SUFFIX],' 'excludeSuffixes=[],' encrypt=False, '\' repl_data=True, 'outputfile='\''%s/l
dif/%s.ldif'\''' % '(topology.standalone.dbdir,SERVERID_STANDALONE' '))' result = 'topology.standalone.ldif2db(DEFAULT_BENAME,' None, None, False, ''\''%s/ldif/%s.ldif'\''' % '(topology.standalone.dbdir,SERVERID_STANDALONE' '))' '>' 'assert(result)' E assert False <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48383_test.py>:123: AssertionError ---------------------------- Captured stdout setup ----------------------------- OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stdout call ----------------------------- OK group dirsrv exists OK user dirsrv exists Exported ldif file: /var/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stderr call ----------------------------- CRITICAL:tickets.ticket48383_test:Failed to change nsslapd-cachememsize No such object INFO:lib389:Running script: /usr/sbin/db2ldif -Z standalone -n userRoot -s dc=example,dc=com -a /var/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif -r '[06/Nov/2016:01:45:01.114302980' '+0100]' - DEBUG - ldbm_back_start - userRoot: entry cache size: 10485760 'B;' db size: 10321920 B '[06/Nov/2016:01:45:01.122790760' '+0100]' - DEBUG - ldbm_back_start - total cache size: 20971520 'B;' '[06/Nov/2016:01:45:01.127497484' '+0100]' - DEBUG - ldbm_back_start - Total entry cache size: 20971520 'B;' dbcache size: 10000000 'B;' available memory size: 2061860864 'B;' ldiffile: /var/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif '[06/Nov/2016:01:45:01.176685068' '+0100]' - ERR - ldbm_back_ldbm2ldif - db2ldif: 'can'\''t' open /var/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif: 2 '(No' such file or 'directory)' '[06/Nov/2016:01:45:01.186710413' '+0100]' - INFO - dblayer_pre_close - Waiting for 4 database threads to stop '[06/Nov/2016:01:45:02.138980064' '+0100]' - INFO - dblayer_pre_close - All database threads now stopped ERROR:lib389:ldif2db: 'Can'\''t' find file: /v
ar/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif ___________________ test_ticket48497_homeDirectory_index_run ___________________ topology = '<tickets.ticket48497_test.TopologyStandalone' object at '0x7fa7f9e9c050>' def 'test_ticket48497_homeDirectory_index_run(topology):' args = '{TASK_WAIT:' 'True}' 'topology.standalone.tasks.reindex(suffix=SUFFIX,' 'attrname='\''homeDirectory'\'',' 'args=args)' 'log.info("Check' indexing succeeded with a specified matching 'rule")' file_path = 'os.path.join(topology.standalone.prefix,' '"var/log/dirsrv/slapd-%s/errors"' % 'topology.standalone.serverid)' '>' file_obj = 'open(file_path,' '"r")' E IOError: '[Errno' '2]' No such file or directory: ''\''/usr/var/log/dirsrv/slapd-standalone/errors'\''' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48497_test.py>:139: IOError ----------------------------- Captured stderr call ----------------------------- INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_014532 completed successfully INFO:tickets.ticket48497_test:Check indexing succeeded with a specified matching rule __________________ test_ticket48745_homeDirectory_indexed_cis __________________ topology = '<tickets.ticket48745_test.TopologyStandalone' object at '0x7fa7f0faae10>' def 'test_ticket48745_homeDirectory_indexed_cis(topology):' 'log.info("\n\nindex' homeDirectory in caseIgnoreIA5Match and 'caseExactIA5Match")' try: ent = 'topology.standalone.getEntry(HOMEDIRECTORY_INDEX,' 'ldap.SCOPE_BASE)' except ldap.NO_SUCH_OBJECT: 'topology.standalone.add_s(Entry((HOMEDIRECTORY_INDEX,' '{' ''\''objectclass'\'':' '"top' 'nsIndex".split(),' ''\''cn'\'':' HOMEDIRECTORY_CN, ''\''nsSystemIndex'\'':' ''\''false'\'',' ''\''nsIndexType'\'':' ''\''eq'\''})))' '#log.info("attach' 'debugger")' '#time.sleep(60)' 'IGNORE_MR_NAME='\''caseIgnoreIA5Match'\''' 'EXACT_MR_NAME='\''caseExactIA5Match'\''' mod = '[(ldap.MOD_REPLACE,' MATCHINGRULE, '(IGNORE_MR_NAME,' 'EXAC
T_MR_NAME))]' 'topology.standalone.modify_s(HOMEDIRECTORY_INDEX,' 'mod)' '#topology.standalone.stop(timeout=10)' 'log.info("successfully' checked that filter with exact mr , a filter with lowercase eq is 'failing")' '#assert' 'topology.standalone.db2index(bename=DEFAULT_BENAME,' suffixes=None, 'attrs=['\''homeDirectory'\''])' '#topology.standalone.start(timeout=10)' args = '{TASK_WAIT:' 'True}' 'topology.standalone.tasks.reindex(suffix=SUFFIX,' 'attrname='\''homeDirectory'\'',' 'args=args)' 'log.info("Check' indexing succeeded with a specified matching 'rule")' file_path = 'os.path.join(topology.standalone.prefix,' '"var/log/dirsrv/slapd-%s/errors"' % 'topology.standalone.serverid)' '>' file_obj = 'open(file_path,' '"r")' E IOError: '[Errno' '2]' No such file or directory: ''\''/usr/var/log/dirsrv/slapd-standalone/errors'\''' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48745_test.py>:110: IOError ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket48745_test: index homeDirectory in caseIgnoreIA5Match and caseExactIA5Match INFO:tickets.ticket48745_test:successfully checked that filter with exact mr , a filter with lowercase eq is failing INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_014702 completed successfully INFO:tickets.ticket48745_test:Check indexing succeeded with a specified matching rule __________________ test_ticket48746_homeDirectory_indexed_cis __________________ topology = '<tickets.ticket48746_test.TopologyStandalone' object at '0x7fa7f9e08050>' def 'test_ticket48746_homeDirectory_indexed_cis(topology):' 'log.info("\n\nindex' homeDirectory in caseIgnoreIA5Match and 'caseExactIA5Match")' try: ent = 'topology.standalone.getEntry(HOMEDIRECTORY_INDEX,' 'ldap.SCOPE_BASE)' except ldap.NO_SUCH_OBJECT: 'topology.standalone.add_s(Entry((HOMEDIRECTORY_INDEX,' '{' ''\''objectclass'\'':' '"top' 'nsIndex".split(),' ''\''cn'\'':' HO
MEDIRECTORY_CN, ''\''nsSystemIndex'\'':' ''\''false'\'',' ''\''nsIndexType'\'':' ''\''eq'\''})))' '#log.info("attach' 'debugger")' '#time.sleep(60)' 'IGNORE_MR_NAME='\''caseIgnoreIA5Match'\''' 'EXACT_MR_NAME='\''caseExactIA5Match'\''' mod = '[(ldap.MOD_REPLACE,' MATCHINGRULE, '(IGNORE_MR_NAME,' 'EXACT_MR_NAME))]' 'topology.standalone.modify_s(HOMEDIRECTORY_INDEX,' 'mod)' '#topology.standalone.stop(timeout=10)' 'log.info("successfully' checked that filter with exact mr , a filter with lowercase eq is 'failing")' '#assert' 'topology.standalone.db2index(bename=DEFAULT_BENAME,' suffixes=None, 'attrs=['\''homeDirectory'\''])' '#topology.standalone.start(timeout=10)' args = '{TASK_WAIT:' 'True}' 'topology.standalone.tasks.reindex(suffix=SUFFIX,' 'attrname='\''homeDirectory'\'',' 'args=args)' 'log.info("Check' indexing succeeded with a specified matching 'rule")' file_path = 'os.path.join(topology.standalone.prefix,' '"var/log/dirsrv/slapd-%s/errors"' % 'topology.standalone.serverid)' '>' file_obj = 'open(file_path,' '"r")' E IOError: '[Errno' '2]' No such file or directory: ''\''/usr/var/log/dirsrv/slapd-standalone/errors'\''' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48746_test.py>:108: IOError ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket48746_test: index homeDirectory in caseIgnoreIA5Match and caseExactIA5Match INFO:tickets.ticket48746_test:successfully checked that filter with exact mr , a filter with lowercase eq is failing INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_014731 completed successfully INFO:tickets.ticket48746_test:Check indexing succeeded with a specified matching rule __________________ test_ticket48746_homeDirectory_indexed_ces __________________ topology = '<tickets.ticket48746_test.TopologyStandalone' object at '0x7fa7f9e08050>' def 'test_ticket48746_homeDirectory_indexed_ces(topology):' 'log.info("\n\ninde
x' homeDirectory in caseExactIA5Match, this would trigger the 'crash")' try: ent = 'topology.standalone.getEntry(HOMEDIRECTORY_INDEX,' 'ldap.SCOPE_BASE)' except ldap.NO_SUCH_OBJECT: 'topology.standalone.add_s(Entry((HOMEDIRECTORY_INDEX,' '{' ''\''objectclass'\'':' '"top' 'nsIndex".split(),' ''\''cn'\'':' HOMEDIRECTORY_CN, ''\''nsSystemIndex'\'':' ''\''false'\'',' ''\''nsIndexType'\'':' ''\''eq'\''})))' '#' 'log.info("attach' 'debugger")' '#' 'time.sleep(60)' 'EXACT_MR_NAME='\''caseExactIA5Match'\''' mod = '[(ldap.MOD_REPLACE,' MATCHINGRULE, '(EXACT_MR_NAME))]' 'topology.standalone.modify_s(HOMEDIRECTORY_INDEX,' 'mod)' '#topology.standalone.stop(timeout=10)' 'log.info("successfully' checked that filter with exact mr , a filter with lowercase eq is 'failing")' '#assert' 'topology.standalone.db2index(bename=DEFAULT_BENAME,' suffixes=None, 'attrs=['\''homeDirectory'\''])' '#topology.standalone.start(timeout=10)' args = '{TASK_WAIT:' 'True}' 'topology.standalone.tasks.reindex(suffix=SUFFIX,' 'attrname='\''homeDirectory'\'',' 'args=args)' 'log.info("Check' indexing succeeded with a specified matching 'rule")' file_path = 'os.path.join(topology.standalone.prefix,' '"var/log/dirsrv/slapd-%s/errors"' % 'topology.standalone.serverid)' '>' file_obj = 'open(file_path,' '"r")' E IOError: '[Errno' '2]' No such file or directory: ''\''/usr/var/log/dirsrv/slapd-standalone/errors'\''' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48746_test.py>:172: IOError ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket48746_test: index homeDirectory in caseExactIA5Match, this would trigger the crash INFO:tickets.ticket48746_test:successfully checked that filter with exact mr , a filter with lowercase eq is failing INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_014733 completed successfully INFO:tickets.ticket48746_test:Check indexing succeeded with a specified m
atching rule _____________________ test_ticket48906_dblock_ldap_update ______________________ topology = '<tickets.ticket48906_test.TopologyStandalone' object at '0x7fa7ef30aa90>' def 'test_ticket48906_dblock_ldap_update(topology):' 'topology.standalone.log.info('\''###################################'\'')' 'topology.standalone.log.info('\''###'\'')' 'topology.standalone.log.info('\''###' Check that after ldap 'update'\'')' 'topology.standalone.log.info('\''###' - monitor contains 'DEFAULT'\'')' 'topology.standalone.log.info('\''###' - configured contains 'DBLOCK_LDAP_UPDATE'\'')' 'topology.standalone.log.info('\''###' - After stop dse.ldif contains 'DBLOCK_LDAP_UPDATE'\'')' 'topology.standalone.log.info('\''###' - After stop guardian contains 'DEFAULT'\'')' 'topology.standalone.log.info('\''###' In fact guardian should differ from config to recreate the 'env'\'')' 'topology.standalone.log.info('\''###' Check that after restart '(DBenv' 'recreated)'\'')' 'topology.standalone.log.info('\''###' - monitor contains DBLOCK_LDAP_UPDATE ''\'')' 'topology.standalone.log.info('\''###' - configured contains 'DBLOCK_LDAP_UPDATE'\'')' 'topology.standalone.log.info('\''###' - dse.ldif contains 'DBLOCK_LDAP_UPDATE'\'')' 'topology.standalone.log.info('\''###'\'')' 'topology.standalone.log.info('\''###################################'\'')' 'topology.standalone.modify_s(ldbm_config,' '[(ldap.MOD_REPLACE,' DBLOCK_ATTR_CONFIG, 'DBLOCK_LDAP_UPDATE)])' '_check_monitored_value(topology,' 'DBLOCK_DEFAULT)' '_check_configured_value(topology,' attr=DBLOCK_ATTR_CONFIG, expected_value=DBLOCK_LDAP_UPDATE, 'required=True)' 'topology.standalone.stop(timeout=10)' '_check_dse_ldif_value(topology,' attr=DBLOCK_ATTR_CONFIG, 'expected_value=DBLOCK_LDAP_UPDATE)' '>' '_check_guardian_value(topology,' attr=DBLOCK_ATTR_GUARDIAN, 'expected_value=DBLOCK_DEFAULT)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:218: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ _ topology = '<tickets.ticket48906_test.TopologyStandalone' object at '0x7fa7ef30aa90>' attr = ''\''locks'\'',' expected_value = ''\''10000'\''' def '_check_guardian_value(topology,' attr=DBLOCK_ATTR_CONFIG, 'expected_value=None):' guardian_file = topology.standalone.dbdir + ''\''/db/guardian'\''' '>' 'assert(os.path.exists(guardian_file))' E assert '<function' exists at '0x7fa80a67a050>('\''/var/lib/dirsrv/slapd-standalone/db/db/guardian'\'')' E + where '<function' exists at '0x7fa80a67a050>' = '<module' ''\''posixpath'\''' from ''\''/usr/lib64/python2.7/posixpath.pyc'\''>.exists' E + where '<module' ''\''posixpath'\''' from ''\''/usr/lib64/python2.7/posixpath.pyc'\''>' = os.path <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:164: AssertionError ----------------------------- Captured stderr call ----------------------------- INFO:lib389:################################### INFO:lib389:### INFO:lib389:### Check that after ldap update INFO:lib389:### - monitor contains DEFAULT INFO:lib389:### - configured contains DBLOCK_LDAP_UPDATE INFO:lib389:### - After stop dse.ldif contains DBLOCK_LDAP_UPDATE INFO:lib389:### - After stop guardian contains DEFAULT INFO:lib389:### In fact guardian should differ from config to recreate the env INFO:lib389:### Check that after restart '(DBenv' 'recreated)' INFO:lib389:### - monitor contains DBLOCK_LDAP_UPDATE INFO:lib389:### - configured contains DBLOCK_LDAP_UPDATE INFO:lib389:### - dse.ldif contains DBLOCK_LDAP_UPDATE INFO:lib389:### INFO:lib389:################################### _____________________ test_ticket48906_dblock_edit_update ______________________ topology = '<tickets.ticket48906_test.TopologyStandalone' object at '0x7fa7ef30aa90>' def 'test_ticket48906_dblock_edit_update(topology):' 'topology.standalone.log.info('\''###################################'\'')' 'topology.standalone.log.info('\''###'\'')' 'topology.standalone.log.info('\''###' Check that after '
stop'\'')' 'topology.standalone.log.info('\''###' - dse.ldif contains 'DBLOCK_LDAP_UPDATE'\'')' 'topology.standalone.log.info('\''###' - guardian contains 'DBLOCK_LDAP_UPDATE'\'')' 'topology.standalone.log.info('\''###' Check that edit 'dse+restart'\'')' 'topology.standalone.log.info('\''###' - monitor contains 'DBLOCK_EDIT_UPDATE'\'')' 'topology.standalone.log.info('\''###' - configured contains 'DBLOCK_EDIT_UPDATE'\'')' 'topology.standalone.log.info('\''###' Check that after 'stop'\'')' 'topology.standalone.log.info('\''###' - dse.ldif contains 'DBLOCK_EDIT_UPDATE'\'')' 'topology.standalone.log.info('\''###' - guardian contains 'DBLOCK_EDIT_UPDATE'\'')' 'topology.standalone.log.info('\''###'\'')' 'topology.standalone.log.info('\''###################################'\'')' 'topology.standalone.stop(timeout=10)' '_check_dse_ldif_value(topology,' attr=DBLOCK_ATTR_CONFIG, 'expected_value=DBLOCK_LDAP_UPDATE)' '>' '_check_guardian_value(topology,' attr=DBLOCK_ATTR_GUARDIAN, 'expected_value=DBLOCK_LDAP_UPDATE)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:243: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ topology = '<tickets.ticket48906_test.TopologyStandalone' object at '0x7fa7ef30aa90>' attr = ''\''locks'\'',' expected_value = ''\''20000'\''' def '_check_guardian_value(topology,' attr=DBLOCK_ATTR_CONFIG, 'expected_value=None):' guardian_file = topology.standalone.dbdir + ''\''/db/guardian'\''' '>' 'assert(os.path.exists(guardian_file))' E assert '<function' exists at '0x7fa80a67a050>('\''/var/lib/dirsrv/slapd-standalone/db/db/guardian'\'')' E + where '<function' exists at '0x7fa80a67a050>' = '<module' ''\''posixpath'\''' from ''\''/usr/lib64/python2.7/posixpath.pyc'\''>.exists' E + where '<module' ''\''posixpath'\''' from ''\''/usr/lib64/python2.7/posixpath.pyc'\''>' = os.path <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket4
8906_test.py>:164: AssertionError ----------------------------- Captured stderr call ----------------------------- INFO:lib389:################################### INFO:lib389:### INFO:lib389:### Check that after stop INFO:lib389:### - dse.ldif contains DBLOCK_LDAP_UPDATE INFO:lib389:### - guardian contains DBLOCK_LDAP_UPDATE INFO:lib389:### Check that edit dse+restart INFO:lib389:### - monitor contains DBLOCK_EDIT_UPDATE INFO:lib389:### - configured contains DBLOCK_EDIT_UPDATE INFO:lib389:### Check that after stop INFO:lib389:### - dse.ldif contains DBLOCK_EDIT_UPDATE INFO:lib389:### - guardian contains DBLOCK_EDIT_UPDATE INFO:lib389:### INFO:lib389:################################### ________________________ test_ticket48906_dblock_robust ________________________ topology = '<tickets.ticket48906_test.TopologyStandalone' object at '0x7fa7ef30aa90>' def 'test_ticket48906_dblock_robust(topology):' 'topology.standalone.log.info('\''###################################'\'')' 'topology.standalone.log.info('\''###'\'')' 'topology.standalone.log.info('\''###' Check that the following values are 'rejected'\'')' 'topology.standalone.log.info('\''###' - negative 'value'\'')' 'topology.standalone.log.info('\''###' - insuffisant 'value'\'')' 'topology.standalone.log.info('\''###' - invalid 'value'\'')' 'topology.standalone.log.info('\''###' Check that minimum value is 'accepted'\'')' 'topology.standalone.log.info('\''###'\'')' 'topology.standalone.log.info('\''###################################'\'')' 'topology.standalone.start(timeout=10)' '>' '_check_monitored_value(topology,' 'DBLOCK_EDIT_UPDATE)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:291: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ topology = '<tickets.ticket48906_test.TopologyStandalone' object at '0x7fa7ef30aa90>' expected_value = ''\''40000'\''' def '_check_monitored_value(topology,' 'expected_value):' entries = 'topology.standalone.search
_s(ldbm_monitor,' ldap.SCOPE_BASE, ''\''(objectclass=*)'\'')' '>' 'assert(entries[0].hasValue(DBLOCK_ATTR_MONITOR)' and 'entries[0].getValue(DBLOCK_ATTR_MONITOR)' == 'expected_value)' E assert '(True' and ''\''20000'\''' == ''\''40000'\''' E + where True = '<bound' method Entry.hasValue of dn: cn=database,cn=monitor,cn=ldbm database,cn...pd-db-txn-region-wait-rate: '0\nobjectClass:' 'top\nobjectClass:' 'extensibleObject\n\n>('\''nsslapd-db-configured-locks'\'')' E + where '<bound' method Entry.hasValue of dn: cn=database,cn=monitor,cn=ldbm database,cn...pd-db-txn-region-wait-rate: '0\nobjectClass:' 'top\nobjectClass:' 'extensibleObject\n\n>' = dn: cn=database,cn=monitor,cn=ldbm 'database,cn=plugins,cn=config\ncn:' 'database\n...apd-db-txn-region-wait-rate:' '0\nobjectClass:' 'top\nobjectClass:' 'extensibleObject\n\n.hasValue' E - 20000 E '?' '^' E + 40000 E '?' '^)' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:144: AssertionError ----------------------------- Captured stderr call ----------------------------- INFO:lib389:################################### INFO:lib389:### INFO:lib389:### Check that the following values are rejected INFO:lib389:### - negative value INFO:lib389:### - insuffisant value INFO:lib389:### - invalid value INFO:lib389:### Check that minimum value is accepted INFO:lib389:### INFO:lib389:################################### 'INFO:lib389:open():' Connecting to uri ldap://localhost.localdomain:38931/ 'INFO:lib389:open():' bound as cn=Directory Manager _____________________________ test_dynamic_plugins _____________________________ topology = '<test_dynamic_plugins.TopologyStandalone' object at '0x7fa7f157d510>' def 'test_dynamic_plugins(topology):' '"""' Test Dynamic Plugins - exercise each plugin and its main features, while changing the configuration without restarting the server. Need to test: functionality, stability, and stress. These tests need to run with replication disabled, and with replication s
etup with a second instance. Then test if replication is working, and we have same entries on each side. Functionality - Make sure that as configuration changes are made they take effect immediately. Cross plugin interaction '(e.g.' 'automember/memberOf)' needs to tested, as well as plugin tasks. Need to test plugin config 'validation(dependencies,' 'etc).' Memory Corruption - Restart the plugins many times, and in different orders and test functionality, and stability. This will excerise the internal plugin linked lists, dse callbacks, and task handlers. Stress - Put the server under load that will trigger multiple 'plugins(MO,' RI, DNA, 'etc)' Restart various plugins while these operations are going on. Perform this test 5 'times(stress_max_run).' '"""' REPLICA_PORT = 33334 RUV_FILTER = ''\''(&(nsuniqueid=ffffffff-ffffffff-ffffffff-ffffffff)(objectclass=nstombstone))'\''' master_maxcsn = 0 replica_maxcsn = 0 msg = \' '(no' 'replication)'\''' replication_run = False stress_max_runs = 5 '#' First enable dynamic plugins try: 'topology.standalone.modify_s(DN_CONFIG,' '[(ldap.MOD_REPLACE,' ''\''nsslapd-dynamic-plugins'\'',' ''\''on'\'')])' except ldap.LDAPError as e: 'ldap.fatal('\''Failed' to enable dynamic 'plugin!'\''' + 'e.message['\''desc'\''])' assert False '#' Test that critical plugins can be updated even though the change might not be applied try: 'topology.standalone.modify_s(DN_LDBM,' '[(ldap.MOD_REPLACE,' ''\''description'\'',' ''\''test'\'')])' except ldap.LDAPError as e: 'ldap.fatal('\''Failed' to apply change to critical 'plugin'\''' + 'e.message['\''desc'\''])' assert False while 1: '#' '#' First run the tests with replication disabled, then rerun them with replication set up '#' '############################################################################' '#' Test plugin functionality '############################################################################' 'log.info('\''####################################################################'\'')' 'log.info('\''Testing' Dynamic Plugins 'Functiona
lity'\''' + msg + ''\''...'\'')' 'log.info('\''####################################################################\n'\'')' 'plugin_tests.test_all_plugins(topology.standalone)' 'log.info('\''####################################################################'\'')' 'log.info('\''Successfully' Tested Dynamic Plugins 'Functionality'\''' + msg + ''\''.'\'')' 'log.info('\''####################################################################\n'\'')' '############################################################################' '#' Test the stability by exercising the internal lists, callabcks, and task handlers '############################################################################' 'log.info('\''####################################################################'\'')' 'log.info('\''Testing' Dynamic Plugins for Memory 'Corruption'\''' + msg + ''\''...'\'')' 'log.info('\''####################################################################\n'\'')' prev_plugin_test = None prev_prev_plugin_test = None for plugin_test in plugin_tests.func_tests: '#' '#' Restart the plugin several times '(and' prev 'plugins)' - work that linked list '#' 'plugin_test(topology.standalone,' '"restart")' if prev_prev_plugin_test: 'prev_prev_plugin_test(topology.standalone,' '"restart")' 'plugin_test(topology.standalone,' '"restart")' if prev_plugin_test: 'prev_plugin_test(topology.standalone,' '"restart")' 'plugin_test(topology.standalone,' '"restart")' '#' Now run the functional test 'plugin_test(topology.standalone)' '#' Set the previous tests if prev_plugin_test: prev_prev_plugin_test = prev_plugin_test prev_plugin_test = plugin_test 'log.info('\''####################################################################'\'')' 'log.info('\''Successfully' Tested Dynamic Plugins for Memory 'Corruption'\''' + msg + ''\''.'\'')' 'log.info('\''####################################################################\n'\'')' '############################################################################' '#' Stress two plugins while restarting it, and 
while restarting other plugins. '#' The goal is to not crash, and have the plugins work after stressing them. '############################################################################' 'log.info('\''####################################################################'\'')' 'log.info('\''Stressing' Dynamic 'Plugins'\''' + msg + ''\''...'\'')' 'log.info('\''####################################################################\n'\'')' 'stress_tests.configureMO(topology.standalone)' 'stress_tests.configureRI(topology.standalone)' stress_count = 0 while stress_count '<' stress_max_runs: 'log.info('\''####################################################################'\'')' 'log.info('\''Running' stress 'test'\''' + msg + ''\''.' Run '(%d/%d)...'\''' % '(stress_count' + 1, 'stress_max_runs))' 'log.info('\''####################################################################\n'\'')' try: '#' Launch three new threads to add a bunch of users add_users = 'stress_tests.AddUsers(topology.standalone,' ''\''employee'\'',' 'True)' 'add_users.start()' add_users2 = 'stress_tests.AddUsers(topology.standalone,' ''\''entry'\'',' 'True)' 'add_users2.start()' add_users3 = 'stress_tests.AddUsers(topology.standalone,' ''\''person'\'',' 'True)' 'add_users3.start()' 'time.sleep(1)' '#' While we are adding users restart the MO plugin and an idle plugin 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' 'time.sleep(1)' 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'time.sleep(1)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' 'topology.standalone.plugins.disable(name=PLUGIN_LINKED_ATTRS)' 'topology.standalone.plugins.enable(name=PLUGIN_LINKED_ATTRS)' 'time.sleep(1)' 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' 'time.sleep(2)' 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'time.sleep(1)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' 'topo
logy.standalone.plugins.disable(name=PLUGIN_LINKED_ATTRS)' 'topology.standalone.plugins.enable(name=PLUGIN_LINKED_ATTRS)' 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'time.sleep(1)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' '#' Wait for the ''\''adding'\''' threads to complete 'add_users.join()' 'add_users2.join()' 'add_users3.join()' '#' Now launch three threads to delete the users del_users = 'stress_tests.DelUsers(topology.standalone,' ''\''employee'\'')' 'del_users.start()' del_users2 = 'stress_tests.DelUsers(topology.standalone,' ''\''entry'\'')' 'del_users2.start()' del_users3 = 'stress_tests.DelUsers(topology.standalone,' ''\''person'\'')' 'del_users3.start()' 'time.sleep(1)' '#' Restart both the MO, RI plugins during these deletes, and an idle plugin 'topology.standalone.plugins.disable(name=PLUGIN_REFER_INTEGRITY)' 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' 'topology.standalone.plugins.enable(name=PLUGIN_REFER_INTEGRITY)' 'time.sleep(1)' 'topology.standalone.plugins.disable(name=PLUGIN_REFER_INTEGRITY)' 'time.sleep(1)' 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'time.sleep(1)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' 'time.sleep(1)' 'topology.standalone.plugins.enable(name=PLUGIN_REFER_INTEGRITY)' 'topology.standalone.plugins.disable(name=PLUGIN_LINKED_ATTRS)' 'topology.standalone.plugins.enable(name=PLUGIN_LINKED_ATTRS)' 'topology.standalone.plugins.disable(name=PLUGIN_REFER_INTEGRITY)' 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' 'topology.standalone.plugins.enable(name=PLUGIN_REFER_INTEGRITY)' 'time.sleep(2)' 'topology.standalone.plugins.disable(name=PLUGIN_REFER_INTEGRITY)' 'time.sleep(1)' 'topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF)' 'time.slee
p(1)' 'topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF)' 'time.sleep(1)' 'topology.standalone.plugins.enable(name=PLUGIN_REFER_INTEGRITY)' 'topology.standalone.plugins.disable(name=PLUGIN_LINKED_ATTRS)' 'topology.standalone.plugins.enable(name=PLUGIN_LINKED_ATTRS)' '#' Wait for the ''\''deleting'\''' threads to complete 'del_users.join()' 'del_users2.join()' 'del_users3.join()' '#' Now make sure both the MO and RI plugins still work correctly 'plugin_tests.func_tests[8](topology.standalone)' '#' RI plugin 'plugin_tests.func_tests[5](topology.standalone)' '#' MO plugin '#' Cleanup the stress tests 'stress_tests.cleanup(topology.standalone)' except: 'log.info('\''Stress' test 'failed!'\'')' 'repl_fail(replica_inst)' stress_count += 1 'log.info('\''####################################################################'\'')' 'log.info('\''Successfully' Stressed Dynamic 'Plugins'\''' + msg + ''\''.' Completed '(%d/%d)'\''' % '(stress_count,' 'stress_max_runs))' 'log.info('\''####################################################################\n'\'')' if replication_run: '#' 'We'\''re' done. break else: '#' '#' Enable replication and run everything one more time '#' 'log.info('\''Setting' up replication, and rerunning the 'tests...\n'\'')' '#' Create replica instance replica_inst = 'DirSrv(verbose=False)' 'args_instance[SER_HOST]' = LOCALHOST 'args_instance[SER_PORT]' = REPLICA_PORT 'args_instance[SER_SERVERID_PROP]' = ''\''replica'\''' 'args_instance[SER_CREATION_SUFFIX]' = DEFAULT_SUFFIX args_replica_inst = 'args_instance.copy()' 'replica_inst.allocate(args_replica_inst)' '>' 'replica_inst.create()' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/dynamic-plugins/test_dynamic_plugins.py>:318: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:905: in create 'self._createDirsrv()' _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = '<lib389.DirSrv' instance at '0x7fa7f131dd88>' def '_createDirsrv(self):' '"""Create' a new instance of directory server @param self - containing the set properties SER_HOST '(host)' SER_PORT '(port)' SER_SECURE_PORT '(sslport)' SER_ROOT_DN '(binddn)' SER_ROOT_PW '(bindpw)' SER_CREATION_SUFFIX '(creation_suffix)' SER_USER_ID '(userid)' SER_SERVERID_PROP '(serverid)' SER_GROUP_ID '(groupid)' SER_DEPLOYED_DIR '(prefix)' SER_BACKUP_INST_DIR '(backupdir)' SER_REALM '(krb5_realm)' @return None @raise None '}' '"""' 'DirSrvTools.lib389User(user=DEFAULT_USER)' prog = 'os.path.join(self.ds_paths.sbin_dir,' 'CMD_PATH_SETUP_DS)' if not 'os.path.isfile(prog):' 'log.error("Can'\''t' find file: %r, removing 'extension"' % 'prog)' prog = 'prog[:-3]' '#' Create and extract a service keytab args = '{SER_HOST:' self.host, SER_PORT: self.port, SER_SECURE_PORT: self.sslport, SER_ROOT_DN: self.binddn, SER_ROOT_PW: self.bindpw, SER_CREATION_SUFFIX: self.creation_suffix, SER_USER_ID: self.userid, SER_SERVERID_PROP: self.serverid, SER_GROUP_ID: self.groupid, SER_DEPLOYED_DIR: self.prefix, SER_BACKUP_INST_DIR: self.backupdir, SER_STRICT_HOSTNAME_CHECKING: 'self.strict_hostname}' content = 'formatInfData(args)' result = 'DirSrvTools.runInfProg(prog,' content, self.verbose, 'prefix=self.prefix)' if result '!=' 0: '>' raise 'Exception('\''Failed' to run 'setup-ds.pl'\'')' E Exception: Failed to run setup-ds.pl <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:861: Exception ---------------------------- Captured stdout setup ----------------------------- OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stdout call ----------------------------- OK group dirsrv exists OK user dirsrv exists Instance slapd-passthru removed. OK group dirsrv exists OK user dirsrv exists Instance slapd-passthru removed. OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stderr call 
----------------------------- INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Testing Dynamic Plugins Functionality '(no' 'replication)...' INFO:test_dynamic_plugins:#################################################################### INFO:plugin_tests:Testing Account Policy Plugin... INFO:plugin_tests:test_acctpolicy: PASS INFO:plugin_tests:Testing attribute uniqueness... INFO:plugin_tests:test_attruniq: PASS INFO:plugin_tests:Testing Auto Membership Plugin... INFO:plugin_tests:test_automember: PASS INFO:plugin_tests:Testing Distributed Numeric Assignment Plugin... INFO:plugin_tests:test_dna: PASS INFO:plugin_tests:Testing Linked Attributes... INFO:plugin_tests:test_linkedattrs: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:plugin_tests:Testing Managed Entries... INFO:plugin_tests:test_mep: PASS INFO:plugin_tests:Testing Pass Through Authentication... INFO:lib389:List backend with suffix=dc=pass2,dc=thru INFO:lib389:Creating a local backend INFO:lib389:List backend cn=PASS2,cn=ldbm database,cn=plugins,cn=config INFO:lib389:Found entry dn: cn=PASS2,cn=ldbm database,cn=plugins,cn=config cn: PASS2 nsslapd-cachememsize: 10485760 nsslapd-cachesize: -1 nsslapd-directory: /var/lib/dirsrv/slapd-passthru/db/PASS2 nsslapd-dncachememsize: 10485760 nsslapd-readonly: off nsslapd-require-index: off nsslapd-suffix: dc=pass2,dc=thru objectClass: top objectClass: extensibleObject objectClass: nsBackendInstance INFO:lib389:Entry dn: 'cn="dc=pass2,dc=thru",cn=mapping' tree,cn=config cn: dc=pass2,dc=thru nsslapd-backend: PASS2 nsslapd-state: backend objectclass: top objectclass: extensibleObject objectclass: nsMappingTree INFO:lib389:Found entry dn: 'cn=dc\3Dpass2\2Cdc\3Dthru,cn=mapping' tree,cn=config cn: dc=pass2,dc=thru nsslapd-backend: PASS2 nsslapd-state: backend objectClass: top objectClass: extensibleObject objectClass: nsMappingTree INFO:plugin_tests:test_passthru: PASS INFO:plugin_tests:Testing referentia
l integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing Retro Changelog Plugin... INFO:plugin_tests:test_retrocl: PASS INFO:plugin_tests:Testing RootDN Access Control... INFO:plugin_tests:test_rootdn: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Tested Dynamic Plugins Functionality '(no' 'replication).' INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Testing Dynamic Plugins for Memory Corruption '(no' 'replication)...' INFO:test_dynamic_plugins:#################################################################### INFO:plugin_tests:Testing Account Policy Plugin... INFO:plugin_tests:test_acctpolicy: PASS INFO:plugin_tests:Testing attribute uniqueness... INFO:plugin_tests:test_attruniq: PASS INFO:plugin_tests:Testing Auto Membership Plugin... INFO:plugin_tests:test_automember: PASS INFO:plugin_tests:Testing Distributed Numeric Assignment Plugin... INFO:plugin_tests:test_dna: PASS INFO:plugin_tests:Testing Linked Attributes... INFO:plugin_tests:test_linkedattrs: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:plugin_tests:Testing Managed Entries... INFO:plugin_tests:test_mep: PASS INFO:plugin_tests:Testing Pass Through Authentication... INFO:lib389:List backend with suffix=dc=pass2,dc=thru INFO:lib389:Creating a local backend INFO:lib389:List backend cn=PASS2,cn=ldbm database,cn=plugins,cn=config INFO:lib389:Found entry dn: cn=PASS2,cn=ldbm database,cn=plugins,cn=config cn: PASS2 nsslapd-cachememsize: 10485760 nsslapd-cachesize: -1 nsslapd-directory: /var/lib/dirsrv/slapd-passthru/db/PASS2 nsslapd-dncachememsize: 10485760 nsslapd-readonly: off nsslapd-require-index: off nsslapd-suffix: dc=pass2,dc=thru objectClass: top objectClass: extensibleObject objectClass: nsBackendInstance I
NFO:lib389:Entry dn: 'cn="dc=pass2,dc=thru",cn=mapping' tree,cn=config cn: dc=pass2,dc=thru nsslapd-backend: PASS2 nsslapd-state: backend objectclass: top objectclass: extensibleObject objectclass: nsMappingTree INFO:lib389:Found entry dn: 'cn=dc\3Dpass2\2Cdc\3Dthru,cn=mapping' tree,cn=config cn: dc=pass2,dc=thru nsslapd-backend: PASS2 nsslapd-state: backend objectClass: top objectClass: extensibleObject objectClass: nsMappingTree INFO:plugin_tests:test_passthru: PASS INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing Retro Changelog Plugin... INFO:plugin_tests:test_retrocl: PASS INFO:plugin_tests:Testing RootDN Access Control... INFO:plugin_tests:test_rootdn: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Tested Dynamic Plugins for Memory Corruption '(no' 'replication).' INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Stressing Dynamic Plugins '(no' 'replication)...' INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress test '(no' 'replication).' Run '(1/5)...' INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries '(employee)...' INFO:stress_tests:AddUsers - Adding 250 entries '(entry)...' INFO:stress_tests:AddUsers - Adding 250 entries '(person)...' INFO:stress_tests:AddUsers - Finished adding 250 entries '(employee).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(entry).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(person).' INFO:stress_tests:DelUsers - Deleting 250 entries '(employee)...' 
INFO:stress_tests:DelUsers - Deleting 250 entries '(person)...' INFO:stress_tests:DelUsers - Deleting 250 entries '(entry)...' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(employee).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(entry).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(person).' INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins '(no' 'replication).' Completed '(1/5)' INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress test '(no' 'replication).' Run '(2/5)...' INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries '(employee)...' INFO:stress_tests:AddUsers - Adding 250 entries '(person)...' INFO:stress_tests:AddUsers - Adding 250 entries '(entry)...' INFO:stress_tests:AddUsers - Finished adding 250 entries '(employee).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(entry).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(person).' INFO:stress_tests:DelUsers - Deleting 250 entries '(employee)...' INFO:stress_tests:DelUsers - Deleting 250 entries '(entry)...' INFO:stress_tests:DelUsers - Deleting 250 entries '(person)...' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(employee).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(entry).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(person).' INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO
:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins '(no' 'replication).' Completed '(2/5)' INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress test '(no' 'replication).' Run '(3/5)...' INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries '(entry)...' INFO:stress_tests:AddUsers - Adding 250 entries '(employee)...' INFO:stress_tests:AddUsers - Adding 250 entries '(person)...' INFO:stress_tests:AddUsers - Finished adding 250 entries '(employee).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(entry).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(person).' INFO:stress_tests:DelUsers - Deleting 250 entries '(employee)...' INFO:stress_tests:DelUsers - Deleting 250 entries '(person)...' INFO:stress_tests:DelUsers - Deleting 250 entries '(entry)...' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(person).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(entry).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(employee).' INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins '(no' 'replication).' Completed '(3/5)' INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress test '(no' 'replication).' Ru
n '(4/5)...' INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries '(employee)...' INFO:stress_tests:AddUsers - Adding 250 entries '(entry)...' INFO:stress_tests:AddUsers - Adding 250 entries '(person)...' INFO:stress_tests:AddUsers - Finished adding 250 entries '(person).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(entry).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(employee).' INFO:stress_tests:DelUsers - Deleting 250 entries '(employee)...' INFO:stress_tests:DelUsers - Deleting 250 entries '(entry)...' INFO:stress_tests:DelUsers - Deleting 250 entries '(person)...' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(entry).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(employee).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(person).' INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins '(no' 'replication).' Completed '(4/5)' INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress test '(no' 'replication).' Run '(5/5)...' INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries '(employee)...' INFO:stress_tests:AddUsers - Adding 250 entries '(entry)...' INFO:stress_tests:AddUsers - Adding 250 entries '(person)...' INFO:stress_tests:AddUsers - Finished adding 250 entries '(entry).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(person).' INFO:stress_tests:AddUsers - Finished adding 250 entries '(em
ployee).' INFO:stress_tests:DelUsers - Deleting 250 entries '(entry)...' INFO:stress_tests:DelUsers - Deleting 250 entries '(person)...' INFO:stress_tests:DelUsers - Deleting 250 entries '(employee)...' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(person).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(employee).' INFO:stress_tests:DelUsers - Finished deleting 250 entries '(entry).' INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins '(no' 'replication).' Completed '(5/5)' INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Setting up replication, and rerunning the tests... ============== 24 failed, 471 passed, 27 error in 8991.82 seconds ==============
============================= test session starts ============================== platform linux2 -- Python 2.7.12, pytest-2.9.2, py-1.4.31, pluggy-0.3.1 -- /usr/bin/python2 cachedir: .cache rootdir: <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests,> inifile: plugins: sourceorder-0.5, multihost-1.0 collecting ... collected 522 items tickets/ticket1347760_test.py::test_ticket1347760 PASSED tickets/ticket365_test.py::test_ticket365 PASSED tickets/ticket397_test.py::test_397 PASSED tickets/ticket47313_test.py::test_ticket47313_run PASSED tickets/ticket47384_test.py::test_ticket47384 PASSED tickets/ticket47431_test.py::test_ticket47431_0 PASSED tickets/ticket47431_test.py::test_ticket47431_1 PASSED tickets/ticket47431_test.py::test_ticket47431_2 PASSED tickets/ticket47431_test.py::test_ticket47431_3 PASSED tickets/ticket47462_test.py::test_ticket47462 PASSED tickets/ticket47490_test.py::test_ticket47490_init PASSED tickets/ticket47490_test.py::test_ticket47490_one PASSED tickets/ticket47490_test.py::test_ticket47490_two PASSED tickets/ticket47490_test.py::test_ticket47490_three PASSED tickets/ticket47490_test.py::test_ticket47490_four PASSED tickets/ticket47490_test.py::test_ticket47490_five PASSED tickets/ticket47490_test.py::test_ticket47490_six PASSED tickets/ticket47490_test.py::test_ticket47490_seven PASSED tickets/ticket47490_test.py::test_ticket47490_eight PASSED tickets/ticket47490_test.py::test_ticket47490_nine PASSED tickets/ticket47536_test.py::test_ticket47536 PASSED tickets/ticket47553_test.py::test_ticket47553 PASSED tickets/ticket47560_test.py::test_ticket47560 PASSED tickets/ticket47573_test.py::test_ticket47573_init PASSED tickets/ticket47573_test.py::test_ticket47573_one PASSED tickets/ticket47573_test.py::test_ticket47573_two PASSED tickets/ticket47573_test.py::test_ticket47573_three PASSED tickets/ticket47619_test.py::test_ticket47619_init PASSED tickets/ticket47619_test.py::test_ticket47619_create_index PASSED tickets/ticket47619_test.py::tes
t_ticket47619_reindex PASSED tickets/ticket47619_test.py::test_ticket47619_check_indexed_search PASSED tickets/ticket47640_test.py::test_ticket47640 PASSED tickets/ticket47653MMR_test.py::test_ticket47653_init PASSED tickets/ticket47653MMR_test.py::test_ticket47653_add PASSED tickets/ticket47653MMR_test.py::test_ticket47653_modify PASSED tickets/ticket47653_test.py::test_ticket47653_init PASSED tickets/ticket47653_test.py::test_ticket47653_add PASSED tickets/ticket47653_test.py::test_ticket47653_search PASSED tickets/ticket47653_test.py::test_ticket47653_modify PASSED tickets/ticket47653_test.py::test_ticket47653_delete PASSED tickets/ticket47669_test.py::test_ticket47669_init FAILED tickets/ticket47669_test.py::test_ticket47669_changelog_maxage FAILED tickets/ticket47669_test.py::test_ticket47669_changelog_triminterval FAILED tickets/ticket47669_test.py::test_ticket47669_changelog_compactdbinterval FAILED tickets/ticket47669_test.py::test_ticket47669_retrochangelog_maxage FAILED tickets/ticket47676_test.py::test_ticket47676_init PASSED tickets/ticket47676_test.py::test_ticket47676_skip_oc_at PASSED tickets/ticket47676_test.py::test_ticket47676_reject_action PASSED tickets/ticket47714_test.py::test_ticket47714_init PASSED tickets/ticket47714_test.py::test_ticket47714_run_0 PASSED tickets/ticket47714_test.py::test_ticket47714_run_1 PASSED tickets/ticket47721_test.py::test_ticket47721_init PASSED tickets/ticket47721_test.py::test_ticket47721_0 PASSED tickets/ticket47721_test.py::test_ticket47721_1 PASSED tickets/ticket47721_test.py::test_ticket47721_2 PASSED tickets/ticket47721_test.py::test_ticket47721_3 PASSED tickets/ticket47721_test.py::test_ticket47721_4 PASSED tickets/ticket47781_test.py::test_ticket47781 PASSED tickets/ticket47787_test.py::test_ticket47787_init PASSED tickets/ticket47787_test.py::test_ticket47787_2 PASSED tickets/ticket47808_test.py::test_ticket47808_run PASSED tickets/ticket47815_test.py::test_ticket47815 PASSED tickets/ticket47819_test.py::test_ticket47819 PASSED tickets/ticket47823_test.p
y::test_ticket47823_init PASSED tickets/ticket47823_test.py::test_ticket47823_one_container_add PASSED tickets/ticket47823_test.py::test_ticket47823_one_container_mod PASSED tickets/ticket47823_test.py::test_ticket47823_one_container_modrdn PASSED tickets/ticket47823_test.py::test_ticket47823_multi_containers_add PASSED tickets/ticket47823_test.py::test_ticket47823_multi_containers_mod PASSED tickets/ticket47823_test.py::test_ticket47823_multi_containers_modrdn PASSED tickets/ticket47823_test.py::test_ticket47823_across_multi_containers_add PASSED tickets/ticket47823_test.py::test_ticket47823_across_multi_containers_mod PASSED tickets/ticket47823_test.py::test_ticket47823_across_multi_containers_modrdn PASSED tickets/ticket47823_test.py::test_ticket47823_invalid_config_1 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_2 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_3 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_4 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_5 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_6 FAILED tickets/ticket47823_test.py::test_ticket47823_invalid_config_7 FAILED tickets/ticket47828_test.py::test_ticket47828_init PASSED tickets/ticket47828_test.py::test_ticket47828_run_0 PASSED tickets/ticket47828_test.py::test_ticket47828_run_1 PASSED tickets/ticket47828_test.py::test_ticket47828_run_2 PASSED tickets/ticket47828_test.py::test_ticket47828_run_3 PASSED tickets/ticket47828_test.py::test_ticket47828_run_4 PASSED tickets/ticket47828_test.py::test_ticket47828_run_5 PASSED tickets/ticket47828_test.py::test_ticket47828_run_6 PASSED tickets/ticket47828_test.py::test_ticket47828_run_7 PASSED tickets/ticket47828_test.py::test_ticket47828_run_8 PASSED tickets/ticket47828_test.py::test_ticket47828_run_9 PASSED tickets/ticket47828_test.py::test_ticket47828_run_10 PASSED tickets/ticket47828_test.py::test_ticket47828_run_11 PASSED tickets/ticket47828_test.py::test_ticket47828_run_12 PASSED tickets/tic
ket47828_test.py::test_ticket47828_run_13 PASSED tickets/ticket47828_test.py::test_ticket47828_run_14 PASSED tickets/ticket47828_test.py::test_ticket47828_run_15 PASSED tickets/ticket47828_test.py::test_ticket47828_run_16 PASSED tickets/ticket47828_test.py::test_ticket47828_run_17 PASSED tickets/ticket47828_test.py::test_ticket47828_run_18 PASSED tickets/ticket47828_test.py::test_ticket47828_run_19 PASSED tickets/ticket47828_test.py::test_ticket47828_run_20 PASSED tickets/ticket47828_test.py::test_ticket47828_run_21 PASSED tickets/ticket47828_test.py::test_ticket47828_run_22 PASSED tickets/ticket47828_test.py::test_ticket47828_run_23 PASSED tickets/ticket47828_test.py::test_ticket47828_run_24 PASSED tickets/ticket47828_test.py::test_ticket47828_run_25 PASSED tickets/ticket47828_test.py::test_ticket47828_run_26 PASSED tickets/ticket47828_test.py::test_ticket47828_run_27 PASSED tickets/ticket47828_test.py::test_ticket47828_run_28 PASSED tickets/ticket47828_test.py::test_ticket47828_run_29 PASSED tickets/ticket47828_test.py::test_ticket47828_run_30 PASSED tickets/ticket47828_test.py::test_ticket47828_run_31 PASSED tickets/ticket47829_test.py::test_ticket47829_init PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_2 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_3 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_2 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_3 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_out_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_out_user_2 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_out_user_3 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_modrdn_active_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_active_user_modrdn_stage_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_activ
e_user_modrdn_out_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_modrdn_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_modrdn_active_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_mod_stage_user_modrdn_stage_user_1 PASSED tickets/ticket47829_test.py::test_ticket47829_indirect_active_group_1 PASSED tickets/ticket47829_test.py::test_ticket47829_indirect_active_group_2 PASSED tickets/ticket47829_test.py::test_ticket47829_indirect_active_group_3 PASSED tickets/ticket47829_test.py::test_ticket47829_indirect_active_group_4 PASSED tickets/ticket47833_test.py::test_ticket47829_init PASSED tickets/ticket47833_test.py::test_ticket47829_mod_stage_user_modrdn_stage_user_1 PASSED tickets/ticket47869MMR_test.py::test_ticket47869_init PASSED tickets/ticket47869MMR_test.py::test_ticket47869_check PASSED tickets/ticket47871_test.py::test_ticket47871_init PASSED tickets/ticket47871_test.py::test_ticket47871_1 PASSED tickets/ticket47871_test.py::test_ticket47871_2 PASSED tickets/ticket47900_test.py::test_ticket47900 PASSED tickets/ticket47910_test.py::test_ticket47910_logconv_start_end_positive PASSED tickets/ticket47910_test.py::test_ticket47910_logconv_start_end_negative PASSED tickets/ticket47910_test.py::test_ticket47910_logconv_start_end_invalid PASSED tickets/ticket47910_test.py::test_ticket47910_logconv_noaccesslogs PASSED tickets/ticket47920_test.py::test_ticket47920_init PASSED tickets/ticket47920_test.py::test_ticket47920_mod_readentry_ctrl PASSED tickets/ticket47921_test.py::test_ticket47921 PASSED tickets/ticket47927_test.py::test_ticket47927_init PASSED tickets/ticket47927_test.py::test_ticket47927_one PASSED tickets/ticket47927_test.py::test_ticket47927_two PASSED tickets/ticket47927_test.py::test_ticket47927_three PASSED tickets/ticket47927_test.py::test_ticket47927_four PASSED tickets/ticket47927_test.py::test_ticket47927_five PASSED tickets/ticket47927_test.py::test_ticket47927_six PASSED tickets/ticket47931_test.py::test_ticket47931 PASSED tickets/ticket47937_te
st.py::test_ticket47937 PASSED tickets/ticket47950_test.py::test_ticket47950 PASSED tickets/ticket47953_test.py::test_ticket47953 PASSED tickets/ticket47963_test.py::test_ticket47963 PASSED tickets/ticket47966_test.py::test_ticket47966 PASSED tickets/ticket47970_test.py::test_ticket47970 PASSED tickets/ticket47973_test.py::test_ticket47973 PASSED tickets/ticket47976_test.py::test_ticket47976_init PASSED tickets/ticket47976_test.py::test_ticket47976_1 PASSED tickets/ticket47976_test.py::test_ticket47976_2 PASSED tickets/ticket47976_test.py::test_ticket47976_3 PASSED tickets/ticket47980_test.py::test_ticket47980 PASSED tickets/ticket47981_test.py::test_ticket47981 PASSED tickets/ticket47988_test.py::test_ticket47988_init PASSED tickets/ticket47988_test.py::test_ticket47988_1 PASSED tickets/ticket47988_test.py::test_ticket47988_2 PASSED tickets/ticket47988_test.py::test_ticket47988_3 PASSED tickets/ticket47988_test.py::test_ticket47988_4 PASSED tickets/ticket47988_test.py::test_ticket47988_5 PASSED tickets/ticket47988_test.py::test_ticket47988_6 PASSED tickets/ticket48005_test.py::test_ticket48005_setup PASSED tickets/ticket48005_test.py::test_ticket48005_memberof PASSED tickets/ticket48005_test.py::test_ticket48005_automember PASSED tickets/ticket48005_test.py::test_ticket48005_syntaxvalidate PASSED tickets/ticket48005_test.py::test_ticket48005_usn PASSED tickets/ticket48005_test.py::test_ticket48005_schemareload PASSED tickets/ticket48013_test.py::test_ticket48013 PASSED tickets/ticket48026_test.py::test_ticket48026 PASSED tickets/ticket48109_test.py::test_ticket48109 PASSED tickets/ticket48170_test.py::test_ticket48170 PASSED tickets/ticket48194_test.py::test_init PASSED tickets/ticket48194_test.py::test_run_0 PASSED tickets/ticket48194_test.py::test_run_1 PASSED tickets/ticket48194_test.py::test_run_2 PASSED tickets/ticket48194_test.py::test_run_3 PASSED tickets/ticket48194_test.py::test_run_4 PASSED tickets/ticket48194_test.py::test_run_5 PASSED tickets/ticket48194_test.py::test_run_6 PASSED tickets/ticket48194
_test.py::test_run_7 PASSED tickets/ticket48194_test.py::test_run_8 PASSED tickets/ticket48194_test.py::test_run_9 PASSED tickets/ticket48194_test.py::test_run_10 PASSED tickets/ticket48194_test.py::test_run_11 PASSED tickets/ticket48212_test.py::test_ticket48212 PASSED tickets/ticket48214_test.py::test_ticket48214_run PASSED tickets/ticket48226_test.py::test_ticket48226_set_purgedelay PASSED tickets/ticket48226_test.py::test_ticket48226_1 FAILED tickets/ticket48228_test.py::test_ticket48228_test_global_policy PASSED tickets/ticket48228_test.py::test_ticket48228_test_subtree_policy PASSED tickets/ticket48233_test.py::test_ticket48233 PASSED tickets/ticket48234_test.py::test_ticket48234 PASSED tickets/ticket48252_test.py::test_ticket48252_setup PASSED tickets/ticket48252_test.py::test_ticket48252_run_0 PASSED tickets/ticket48252_test.py::test_ticket48252_run_1 PASSED tickets/ticket48265_test.py::test_ticket48265_test PASSED tickets/ticket48266_test.py::test_ticket48266_fractional PASSED tickets/ticket48266_test.py::test_ticket48266_check_repl_desc PASSED tickets/ticket48266_test.py::test_ticket48266_count_csn_evaluation FAILED tickets/ticket48270_test.py::test_ticket48270_init PASSED tickets/ticket48270_test.py::test_ticket48270_homeDirectory_indexed_cis FAILED tickets/ticket48270_test.py::test_ticket48270_homeDirectory_mixed_value PASSED tickets/ticket48270_test.py::test_ticket48270_extensible_search PASSED tickets/ticket48272_test.py::test_ticket48272 PASSED tickets/ticket48294_test.py::test_48294_init PASSED tickets/ticket48294_test.py::test_48294_run_0 PASSED tickets/ticket48294_test.py::test_48294_run_1 PASSED tickets/ticket48294_test.py::test_48294_run_2 PASSED tickets/ticket48295_test.py::test_48295_init PASSED tickets/ticket48295_test.py::test_48295_run PASSED tickets/ticket48312_test.py::test_ticket48312 PASSED tickets/ticket48325_test.py::test_ticket48325 PASSED tickets/ticket48342_test.py::test_ticket4026 PASSED tickets/ticket48354_test.py::test_ticket48354 PASSED tickets/ticket48362_test.py::test_ticke
t48362 PASSED tickets/ticket48366_test.py::test_ticket48366_init PASSED tickets/ticket48366_test.py::test_ticket48366_search_user PASSED tickets/ticket48366_test.py::test_ticket48366_search_dm PASSED tickets/ticket48370_test.py::test_ticket48370 PASSED tickets/ticket48383_test.py::test_ticket48383 FAILED tickets/ticket48497_test.py::test_ticket48497_init PASSED tickets/ticket48497_test.py::test_ticket48497_homeDirectory_mixed_value PASSED tickets/ticket48497_test.py::test_ticket48497_extensible_search PASSED tickets/ticket48497_test.py::test_ticket48497_homeDirectory_index_cfg PASSED tickets/ticket48497_test.py::test_ticket48497_homeDirectory_index_run FAILED tickets/ticket48637_test.py::test_ticket48637 PASSED tickets/ticket48665_test.py::test_ticket48665 PASSED tickets/ticket48745_test.py::test_ticket48745_init PASSED tickets/ticket48745_test.py::test_ticket48745_homeDirectory_indexed_cis FAILED tickets/ticket48745_test.py::test_ticket48745_homeDirectory_mixed_value PASSED tickets/ticket48745_test.py::test_ticket48745_extensible_search_after_index PASSED tickets/ticket48746_test.py::test_ticket48746_init PASSED tickets/ticket48746_test.py::test_ticket48746_homeDirectory_indexed_cis FAILED tickets/ticket48746_test.py::test_ticket48746_homeDirectory_mixed_value PASSED tickets/ticket48746_test.py::test_ticket48746_extensible_search_after_index PASSED tickets/ticket48746_test.py::test_ticket48746_homeDirectory_indexed_ces FAILED tickets/ticket48755_test.py::test_ticket48755 PASSED tickets/ticket48759_test.py::test_ticket48759 PASSED tickets/ticket48784_test.py::test_ticket48784 PASSED tickets/ticket48798_test.py::test_ticket48798 PASSED tickets/ticket48799_test.py::test_ticket48799 PASSED tickets/ticket48808_test.py::test_ticket48808 PASSED tickets/ticket48844_test.py::test_ticket48844_init PASSED tickets/ticket48844_test.py::test_ticket48844_bitwise_on PASSED tickets/ticket48844_test.py::test_ticket48844_bitwise_off PASSED tickets/ticket48891_test.py::test_ticket48891_setup PASSED tickets/ticket48893_test.py::test
_ticket48893 PASSED tickets/ticket48896_test.py::test_ticket48896 PASSED tickets/ticket48906_test.py::test_ticket48906_setup PASSED tickets/ticket48906_test.py::test_ticket48906_dblock_default PASSED tickets/ticket48906_test.py::test_ticket48906_dblock_ldap_update FAILED tickets/ticket48906_test.py::test_ticket48906_dblock_edit_update FAILED tickets/ticket48906_test.py::test_ticket48906_dblock_robust FAILED tickets/ticket48916_test.py::test_ticket48916 ERROR tickets/ticket48956_test.py::test_ticket48956 PASSED tickets/ticket548_test.py::test_ticket548_test_with_no_policy PASSED tickets/ticket548_test.py::test_ticket548_test_global_policy PASSED tickets/ticket548_test.py::test_ticket548_test_subtree_policy PASSED suites/acct_usability_plugin/acct_usability_test.py::test_acct_usability_init PASSED suites/acct_usability_plugin/acct_usability_test.py::test_acct_usability_ PASSED suites/acctpolicy_plugin/acctpolicy_test.py::test_acctpolicy_init PASSED suites/acctpolicy_plugin/acctpolicy_test.py::test_acctpolicy_ PASSED suites/acl/acl_test.py::test_aci_attr_subtype_targetattr[lang-ja] ERROR suites/acl/acl_test.py::test_aci_attr_subtype_targetattr[binary] ERROR suites/acl/acl_test.py::test_aci_attr_subtype_targetattr[phonetic] ERROR suites/acl/acl_test.py::test_mode_default_add_deny ERROR suites/acl/acl_test.py::test_mode_default_delete_deny ERROR suites/acl/acl_test.py::test_moddn_staging_prod[0-cn=staged user,dc=example,dc=com-cn=accounts,dc=example,dc=com-False] ERROR suites/acl/acl_test.py::test_moddn_staging_prod[1-cn=staged user,dc=example,dc=com-cn=accounts,dc=example,dc=com-False] ERROR suites/acl/acl_test.py::test_moddn_staging_prod[2-cn=staged user,dc=example,dc=com-cn=bad*,dc=example,dc=com-True] ERROR suites/acl/acl_test.py::test_moddn_staging_prod[3-cn=st*,dc=example,dc=com-cn=accounts,dc=example,dc=com-False] ERROR suites/acl/acl_test.py::test_moddn_staging_prod[4-cn=bad*,dc=example,dc=com-cn=accounts,dc=example,dc=com-True] ERROR suites/acl/acl_test.py::test_moddn_staging_prod[5-cn=st*,dc=example,dc=com-c
n=ac*,dc=example,dc=com-False] ERROR suites/acl/acl_test.py::test_moddn_staging_prod[6-None-cn=ac*,dc=example,dc=com-False] ERROR suites/acl/acl_test.py::test_moddn_staging_prod[7-cn=st*,dc=example,dc=com-None-False] ERROR suites/acl/acl_test.py::test_moddn_staging_prod[8-None-None-False] ERROR suites/acl/acl_test.py::test_moddn_staging_prod_9 ERROR suites/acl/acl_test.py::test_moddn_prod_staging ERROR suites/acl/acl_test.py::test_check_repl_M2_to_M1 ERROR suites/acl/acl_test.py::test_moddn_staging_prod_except ERROR suites/acl/acl_test.py::test_mode_default_ger_no_moddn ERROR suites/acl/acl_test.py::test_mode_default_ger_with_moddn ERROR suites/acl/acl_test.py::test_mode_switch_default_to_legacy ERROR suites/acl/acl_test.py::test_mode_legacy_ger_no_moddn1 ERROR suites/acl/acl_test.py::test_mode_legacy_ger_no_moddn2 ERROR suites/acl/acl_test.py::test_mode_legacy_ger_with_moddn ERROR suites/acl/acl_test.py::test_rdn_write_get_ger ERROR suites/acl/acl_test.py::test_rdn_write_modrdn_anonymous ERROR suites/attr_encryption/attr_encrypt_test.py::test_attr_encrypt_init PASSED suites/attr_encryption/attr_encrypt_test.py::test_attr_encrypt_ PASSED suites/attr_uniqueness_plugin/attr_uniqueness_test.py::test_attr_uniqueness_init PASSED suites/attr_uniqueness_plugin/attr_uniqueness_test.py::test_attr_uniqueness PASSED suites/automember_plugin/automember_test.py::test_automember_init PASSED suites/automember_plugin/automember_test.py::test_automember_ PASSED suites/basic/basic_test.py::test_basic_ops PASSED suites/basic/basic_test.py::test_basic_import_export PASSED suites/basic/basic_test.py::test_basic_backup PASSED suites/basic/basic_test.py::test_basic_acl PASSED suites/basic/basic_test.py::test_basic_searches PASSED suites/basic/basic_test.py::test_basic_referrals PASSED suites/basic/basic_test.py::test_basic_systemctl PASSED suites/basic/basic_test.py::test_basic_ldapagent PASSED suites/basic/basic_test.py::test_basic_dse PASSED suites/basic/basic_test.py::test_def_rootdse_attr[namingContexts] PASSED suites/basic/basic_t
est.py::test_def_rootdse_attr[supportedLDAPVersion] PASSED suites/basic/basic_test.py::test_def_rootdse_attr[supportedControl] PASSED suites/basic/basic_test.py::test_def_rootdse_attr[supportedExtension] PASSED suites/basic/basic_test.py::test_def_rootdse_attr[supportedSASLMechanisms] PASSED suites/basic/basic_test.py::test_def_rootdse_attr[vendorName] PASSED suites/basic/basic_test.py::test_def_rootdse_attr[vendorVersion] PASSED suites/basic/basic_test.py::test_mod_def_rootdse_attr[namingContexts] PASSED suites/basic/basic_test.py::test_mod_def_rootdse_attr[supportedLDAPVersion] PASSED suites/basic/basic_test.py::test_mod_def_rootdse_attr[supportedControl] PASSED suites/basic/basic_test.py::test_mod_def_rootdse_attr[supportedExtension] PASSED suites/basic/basic_test.py::test_mod_def_rootdse_attr[supportedSASLMechanisms] PASSED suites/basic/basic_test.py::test_mod_def_rootdse_attr[vendorName] PASSED suites/basic/basic_test.py::test_mod_def_rootdse_attr[vendorVersion] PASSED suites/betxns/betxn_test.py::test_betxn_init PASSED suites/betxns/betxn_test.py::test_betxt_7bit PASSED suites/betxns/betxn_test.py::test_betxn_attr_uniqueness PASSED suites/betxns/betxn_test.py::test_betxn_memberof PASSED suites/chaining_plugin/chaining_test.py::test_chaining_init PASSED suites/chaining_plugin/chaining_test.py::test_chaining_ PASSED suites/clu/clu_test.py::test_clu_init PASSED suites/clu/clu_test.py::test_clu_pwdhash PASSED suites/clu/db2ldif_test.py::test_db2ldif_init PASSED suites/collation_plugin/collatation_test.py::test_collatation_init PASSED suites/collation_plugin/collatation_test.py::test_collatation_ PASSED suites/config/config_test.py::test_maxbersize_repl PASSED suites/config/config_test.py::test_config_listen_backport_size PASSED suites/config/config_test.py::test_config_deadlock_policy PASSED suites/cos_plugin/cos_test.py::test_cos_init PASSED suites/cos_plugin/cos_test.py::test_cos_ PASSED suites/deref_plugin/deref_test.py::test_deref_init PASSED suites/deref_plugin/deref_test.py::test_deref_ PASSED suites/disk
_monitoring/disk_monitor_test.py::test_disk_monitor_init PASSED suites/disk_monitoring/disk_monitor_test.py::test_disk_monitor_ PASSED suites/distrib_plugin/distrib_test.py::test_distrib_init PASSED suites/distrib_plugin/distrib_test.py::test_distrib_ PASSED suites/dna_plugin/dna_test.py::test_dna_init PASSED suites/dna_plugin/dna_test.py::test_dna_ PASSED suites/ds_logs/ds_logs_test.py::test_ds_logs_init PASSED suites/ds_logs/ds_logs_test.py::test_ds_logs_ PASSED suites/dynamic-plugins/test_dynamic_plugins.py::test_dynamic_plugins FAILED suites/filter/filter_test.py::test_filter_init PASSED suites/filter/filter_test.py::test_filter_escaped PASSED suites/filter/filter_test.py::test_filter_search_original_attrs PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_supported_features PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-False-oper_attr_list0] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-False-oper_attr_list0-*] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-False-oper_attr_list0-objectClass] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-True-oper_attr_list1] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-True-oper_attr_list1-*] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[-True-oper_attr_list1-objectClass] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-False-oper_attr_list2] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-False-oper_attr_list2-*] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-False-oper_attr_list2-objectClass] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-True-oper_attr_list3] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-True-oper_attr_list3-*] PASSED suites/filter/rfc3673_all
_oper_attrs_test.py::test_search_basic[ou=people,dc=example,dc=com-True-oper_attr_list3-objectClass] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-False-oper_attr_list4] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-False-oper_attr_list4-*] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-False-oper_attr_list4-objectClass] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-True-oper_attr_list5] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-True-oper_attr_list5-*] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[uid=all_attrs_test,ou=people,dc=example,dc=com-True-oper_attr_list5-objectClass] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[cn=config-False-oper_attr_list6] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[cn=config-False-oper_attr_list6-*] PASSED suites/filter/rfc3673_all_oper_attrs_test.py::test_search_basic[cn=config-False-oper_attr_list6-objectClass] PASSED suites/get_effective_rights/ger_test.py::test_ger_init PASSED suites/get_effective_rights/ger_test.py::test_ger_ PASSED suites/gssapi_repl/gssapi_repl_test.py::test_gssapi_repl PASSED suites/ldapi/ldapi_test.py::test_ldapi_init PASSED suites/ldapi/ldapi_test.py::test_ldapi_ PASSED suites/linkedattrs_plugin/linked_attrs_test.py::test_linked_attrs_init PASSED suites/linkedattrs_plugin/linked_attrs_test.py::test_linked_attrs_ PASSED suites/mapping_tree/mapping_tree_test.py::test_mapping_tree_init PASSED suites/mapping_tree/mapping_tree_test.py::test_mapping_tree_ PASSED suites/memberof_plugin/memberof_test.py::test_memberof_auto_add_oc PASSED suites/memory_leaks/range_search_test.py::test_range_search_init PASSED suites/memory_leaks/ran
ge_search_test.py::test_range_search PASSED suites/monitor/monitor_test.py::test_monitor_init PASSED suites/monitor/monitor_test.py::test_monitor_ PASSED suites/paged_results/paged_results_test.py::test_search_success[6-5] PASSED suites/paged_results/paged_results_test.py::test_search_success[5-5] PASSED suites/paged_results/paged_results_test.py::test_search_success[5-25] PASSED suites/paged_results/paged_results_test.py::test_search_limits_fail[50-200-cn=config,cn=ldbm database,cn=plugins,cn=config-nsslapd-idlistscanlimit-100-UNWILLING_TO_PERFORM] PASSED suites/paged_results/paged_results_test.py::test_search_limits_fail[5-15-cn=config-nsslapd-timelimit-20-UNAVAILABLE_CRITICAL_EXTENSION] PASSED suites/paged_results/paged_results_test.py::test_search_limits_fail[21-50-cn=config-nsslapd-sizelimit-20-SIZELIMIT_EXCEEDED] PASSED suites/paged_results/paged_results_test.py::test_search_limits_fail[21-50-cn=config-nsslapd-pagedsizelimit-5-SIZELIMIT_EXCEEDED] PASSED suites/paged_results/paged_results_test.py::test_search_limits_fail[5-50-cn=config,cn=ldbm database,cn=plugins,cn=config-nsslapd-lookthroughlimit-20-ADMINLIMIT_EXCEEDED] PASSED suites/paged_results/paged_results_test.py::test_search_sort_success PASSED suites/paged_results/paged_results_test.py::test_search_abandon PASSED suites/paged_results/paged_results_test.py::test_search_with_timelimit PASSED suites/paged_results/paged_results_test.py::test_search_dns_ip_aci[dns = "localhost.localdomain"] PASSED suites/paged_results/paged_results_test.py::test_search_dns_ip_aci[ip = "::1" or ip = "127.0.0.1"] PASSED suites/paged_results/paged_results_test.py::test_search_multiple_paging PASSED suites/paged_results/paged_results_test.py::test_search_invalid_cookie[1000] PASSED suites/paged_results/paged_results_test.py::test_search_invalid_cookie[-1] PASSED suites/paged_results/paged_results_test.py::test_search_abandon_with_zero_size PASSED suites/paged_results/paged_results_test.py::test_search_pagedsizelimit_success PASSED suites/paged_results/paged_results_test.py::
test_search_nspagedsizelimit[5-15-PASS] PASSED suites/paged_results/paged_results_test.py::test_search_nspagedsizelimit[15-5-SIZELIMIT_EXCEEDED] PASSED suites/paged_results/paged_results_test.py::test_search_paged_limits[conf_attr_values0-ADMINLIMIT_EXCEEDED] PASSED suites/paged_results/paged_results_test.py::test_search_paged_limits[conf_attr_values1-PASS] PASSED suites/paged_results/paged_results_test.py::test_search_paged_user_limits[conf_attr_values0-ADMINLIMIT_EXCEEDED] PASSED suites/paged_results/paged_results_test.py::test_search_paged_user_limits[conf_attr_values1-PASS] PASSED suites/paged_results/paged_results_test.py::test_ger_basic PASSED suites/paged_results/paged_results_test.py::test_multi_suffix_search PASSED suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_success[None] PASSED suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_success[-1] PASSED suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_success[1000] PASSED suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_failure[0] PASSED suites/paged_results/paged_results_test.py::test_maxsimplepaged_per_conn_failure[1] PASSED suites/pam_passthru_plugin/pam_test.py::test_pam_init PASSED suites/pam_passthru_plugin/pam_test.py::test_pam_ PASSED suites/passthru_plugin/passthru_test.py::test_passthru_init PASSED suites/passthru_plugin/passthru_test.py::test_passthru_ PASSED suites/password/password_test.py::test_password_init PASSED suites/password/password_test.py::test_password_delete_specific_password PASSED suites/password/pwdAdmin_test.py::test_pwdAdmin_init PASSED suites/password/pwdAdmin_test.py::test_pwdAdmin PASSED suites/password/pwdAdmin_test.py::test_pwdAdmin_config_validation PASSED suites/password/pwdPolicy_attribute_test.py::test_change_pwd[on-off-UNWILLING_TO_PERFORM] PASSED suites/password/pwdPolicy_attribute_test.py::test_change_pwd[off-off-UNWILLING_TO_PERFORM] PASSED suites/password/pwdPolicy_attribute_test.py::test_change_pwd[off-on-None] PASSED 
suites/password/pwdPolicy_attribute_test.py::test_change_pwd[on-on-None] PASSED suites/password/pwdPolicy_attribute_test.py::test_pwd_min_age PASSED suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_no_restrictions[off-off] PASSED suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_no_restrictions[on-off] PASSED suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_no_restrictions[off-on] PASSED suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_restrictions[cn=config] PASSED suites/password/pwdPolicy_inherit_global_test.py::test_entry_has_restrictions[cn="cn=nsPwPolicyEntry,ou=People,dc=example,dc=com",cn=nsPwPolicyContainer,ou=People,dc=example,dc=com] PASSED suites/password/pwdPolicy_syntax_test.py::test_pwdPolicy_syntax PASSED suites/password/pwdPolicy_warning_test.py::test_different_values[ ] PASSED suites/password/pwdPolicy_warning_test.py::test_different_values[junk123] PASSED suites/password/pwdPolicy_warning_test.py::test_different_values[on] PASSED suites/password/pwdPolicy_warning_test.py::test_different_values[off] PASSED suites/password/pwdPolicy_warning_test.py::test_expiry_time PASSED suites/password/pwdPolicy_warning_test.py::test_password_warning[passwordSendExpiringTime-off] PASSED suites/password/pwdPolicy_warning_test.py::test_password_warning[passwordWarning-3600] PASSED suites/password/pwdPolicy_warning_test.py::test_with_different_password_states PASSED suites/password/pwdPolicy_warning_test.py::test_default_behavior PASSED suites/password/pwdPolicy_warning_test.py::test_with_local_policy PASSED suites/password/pwd_algo_test.py::test_pwd_algo_test PASSED suites/password/pwp_history_test.py::test_pwp_history_test PASSED suites/posix_winsync_plugin/posix_winsync_test.py::test_posix_winsync_init PASSED suites/posix_winsync_plugin/posix_winsync_test.py::test_posix_winsync_ PASSED suites/psearch/psearch_test.py::test_psearch_init PASSED suites/psearch/psearch_test.py::test_psearch_ PASSED suites/referint_plugin/referint_test.py::test_referint_
init PASSED suites/referint_plugin/referint_test.py::test_referint_ PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_init PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_clean PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_clean_restart PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_clean_force PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_abort PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_abort_restart PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_abort_certify PASSED suites/replication/cleanallruv_test.py::test_cleanallruv_stress_clean PASSED suites/replication/wait_for_async_feature_test.py::test_not_int_value PASSED suites/replication/wait_for_async_feature_test.py::test_multi_value PASSED suites/replication/wait_for_async_feature_test.py::test_value_check[waitfor_async_attr0] PASSED suites/replication/wait_for_async_feature_test.py::test_value_check[waitfor_async_attr1] PASSED suites/replication/wait_for_async_feature_test.py::test_value_check[waitfor_async_attr2] PASSED suites/replication/wait_for_async_feature_test.py::test_value_check[waitfor_async_attr3] PASSED suites/replication/wait_for_async_feature_test.py::test_behavior_with_value[waitfor_async_attr0] PASSED suites/replication/wait_for_async_feature_test.py::test_behavior_with_value[waitfor_async_attr1] PASSED suites/replication/wait_for_async_feature_test.py::test_behavior_with_value[waitfor_async_attr2] PASSED suites/replication/wait_for_async_feature_test.py::test_behavior_with_value[waitfor_async_attr3] PASSED suites/replsync_plugin/repl_sync_test.py::test_repl_sync_init PASSED suites/replsync_plugin/repl_sync_test.py::test_repl_sync_ PASSED suites/resource_limits/res_limits_test.py::test_res_limits_init PASSED suites/resource_limits/res_limits_test.py::test_res_limits_ PASSED suites/retrocl_plugin/retrocl_test.py::test_retrocl_init PASSED suites/retrocl_plugin/retrocl_test.py::test_retrocl_ PASSED suites/reverpwd_plugin/reverp
wd_test.py::test_reverpwd_init PASSED suites/reverpwd_plugin/reverpwd_test.py::test_reverpwd_ PASSED suites/roles_plugin/roles_test.py::test_roles_init PASSED suites/roles_plugin/roles_test.py::test_roles_ PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_init PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_specific_time PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_day_of_week PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_denied_ip PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_denied_host PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_allowed_ip PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_access_allowed_host PASSED suites/rootdn_plugin/rootdn_plugin_test.py::test_rootdn_config_validate PASSED suites/sasl/sasl_test.py::test_sasl_init PASSED suites/sasl/sasl_test.py::test_sasl_ PASSED suites/schema/test_schema.py::test_schema_comparewithfiles PASSED suites/schema_reload_plugin/schema_reload_test.py::test_schema_reload_init PASSED suites/schema_reload_plugin/schema_reload_test.py::test_schema_reload_ PASSED suites/snmp/snmp_test.py::test_snmp_init PASSED suites/snmp/snmp_test.py::test_snmp_ PASSED suites/ssl/ssl_test.py::test_ssl_init PASSED suites/ssl/ssl_test.py::test_ssl_ PASSED suites/syntax_plugin/syntax_test.py::test_syntax_init PASSED suites/syntax_plugin/syntax_test.py::test_syntax_ PASSED suites/usn_plugin/usn_test.py::test_usn_init PASSED suites/usn_plugin/usn_test.py::test_usn_ PASSED suites/views_plugin/views_test.py::test_views_init PASSED suites/views_plugin/views_test.py::test_views_ PASSED suites/vlv/vlv_test.py::test_vlv_init PASSED suites/vlv/vlv_test.py::test_vlv_ PASSED suites/whoami_plugin/whoami_test.py::test_whoami_init PASSED suites/whoami_plugin/whoami_test.py::test_whoami_ PASSED ==================================== ERRORS ==================================== ______________________ ERROR at setup of test_ticket48916 ______________________ requ
est = <SubRequest 'topology' for <Function 'test_ticket48916'>> @pytest.fixture(scope="module") def topology(request): """Create Replication Deployment""" # Creating master 1... if DEBUGGING: master1 = DirSrv(verbose=True) else: master1 = DirSrv(verbose=False) args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_instance[SER_CREATION_SUFFIX] = DEFAULT_SUFFIX args_master = args_instance.copy() master1.allocate(args_master) instance_master1 = master1.exists() if instance_master1: master1.delete() master1.create() master1.open() master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) # Creating master 2... if DEBUGGING: master2 = DirSrv(verbose=True) else: master2 = DirSrv(verbose=False) args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_instance[SER_CREATION_SUFFIX] = DEFAULT_SUFFIX args_master = args_instance.copy() master2.allocate(args_master) instance_master2 = master2.exists() if instance_master2: master2.delete() master2.create() master2.open() master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # # Create all the agreements # # Creating agreement from master 1 to master 2 properties = {RA_NAME: 'meTo_' + master2.host + ':' + str(master2.port), RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} m1_m2_agmt = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not m1_m2_agmt: log.fatal("Fail to create a master -> master replica agreement") sys.exit(1) log.debug("%s created" % m1_m2_agmt) # Creating agreement from master 2 to master 1 properties = {RA_NAME: 'meTo_' + master1.host + ':' + str(master1.port), RA_BINDDN: defaultProper
ties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} m2_m1_agmt = master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) if not m2_m1_agmt: log.fatal("Fail to create a master -> master replica agreement") sys.exit(1) log.debug("%s created" % m2_m1_agmt) # Allow the replicas to get situated with the new agreements... time.sleep(5) # # Initialize all the agreements # master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) master1.waitForReplInit(m1_m2_agmt) # Check replication is working... if master1.testReplication(DEFAULT_SUFFIX, master2): log.info('Replication is working.') else: log.fatal('Replication is not working.') > assert False E assert False <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48916_test.py>:116: AssertionError ---------------------------- Captured stdout setup ----------------------------- OK group dirsrv exists OK user dirsrv exists OK group dirsrv exists OK user dirsrv exists ('Update succeeded: status ', '0 Total update succeeded') ---------------------------- Captured stderr setup ----------------------------- INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Found entry dn: cn=replrepl,cn=config cn: bind dn pseudo user cn: replrepl objectClass: top objectClass: person sn: bind dn pseudo user userPassword: {SSHA512}q7nvw6jOxgHn8newYgDvaqOo8Q2gIlTAWxkQo15OZF1SY6mRD9RMglTwVOF9kTPYfoVztiHR1dBQy1OT3JiSe3UtbMHVeGkI INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Found entry dn: cn=replrepl,cn=config cn: bind dn pseudo user cn: replrepl objectClass: top objectClass: person sn: bind dn pseudo user userPassword: {SSHA512}M95yCD/bO/csc8UcPdQaLbIAKHVv6aUDSOluZNoE7sjM41rHjAtplXEfx7EZ3++wNJPNLEGjLcYvRt03SGaAxiVbkndDQtPb INFO:lib389:Starting total init cn=meTo_localhost.localdomain:38942,cn=replica,cn=dc\3
Dexample\2Cdc\3Dcom,cn=mapping tree,cn=config CRITICAL:lib389:testReplication() failed to modify (dc=example,dc=com), error ({'desc': "Can't contact LDAP server"}) CRITICAL:tickets.ticket48916_test:Replication is not working. _________ ERROR at setup of test_aci_attr_subtype_targetattr[lang-ja] __________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer p
roperties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2
.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contac
t LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ---------------------------- Captured stdout setup ----------------------------- Instance slapd-master_1 removed. Instance slapd-master_2 removed. OK group dirsrv exists OK user dirsrv exists OK group dirsrv exists OK user dirsrv exists ---------------------------- Captured stderr setup ----------------------------- INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Found entry dn: cn=replrepl,cn=config cn: bind dn pseudo user cn: replrepl objectClass: top objectClass: person sn: bind dn pseudo user userPassword: {SSHA512}CpXNtBaOtj2EfWZS8ofFNVaoXZoSm/C9kng/Nnj2odQdbD6DnakmpHkNRye3V8WUJuPOq80FTmoVar7FPXLFG738x/lvSYTA INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Found entry dn: cn=replrepl,cn=config cn: bind dn pseudo user cn: replrepl objectClass: top objectClass: person sn: bind dn pseudo user userPassword: {SSHA512}7+rYVDUDoyNWc6ZB+gOBvvNhwNQtK8ZF0hLxtc3wuo8rQnSKLsyYNamt4ZVmCFJwOytdCh/78zeYSHr3la1MpY9LE6rzs0PJ DEBUG:suites.acl.acl_test:cn=meTo_$host:$port,cn=replica,cn=dc\3Dexample\2Cdc\3Dcom,cn=mapping tree,cn=config created INFO:lib389:Starting total init cn=meTo_$host:$port,cn=replica,cn=dc\3Dexample\2Cdc\3Dcom,cn=mapping tree,cn=config __________ ERROR at setup of test_aci_attr_subtype_targetattr[binary] __________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy(
) master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.a
bc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,a
dd_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _________ ERROR at setup of test_aci_attr_subtype_targetattr[phonetic] _________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_ins
tance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _
 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner retu
rn f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _________________ ERROR at setup of test_mode_default_add_deny _________________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_P
ORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/
lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7
fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_mode_default_delete_deny ________________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) 
# Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com
:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message
_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of test_moddn_staging_prod[0-cn=staged user,dc=example,dc=com-cn=accounts,dc=example,dc=com-False] request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_
master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check
_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**k
wargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of test_moddn_staging_prod[1-cn=staged user,dc=example,dc=com-cn=accounts,dc=example,dc=com-False] request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_ma
ster2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/l
ib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_o
bject_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of test_moddn_staging_prod[2-cn=staged user,dc=example,dc=com-cn=bad*,dc=example,dc=com-True] request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() 
# # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.a
bc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( r
epr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of test_moddn_staging_prod[3-cn=st*,dc=example,dc=com-cn=accounts,dc=example,dc=com-False] request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix
=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry resty
pe, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self.
_trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of test_moddn_staging_prod[4-cn=bad*,dc=example,dc=com-cn=accounts,dc=example,dc=com-True] request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, r
ole=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib
389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_suc
cess = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of test_moddn_staging_prod[5-cn=st*,dc=example,dc=com-cn=ac*,dc=example,dc=com-False] request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo
_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapob
ject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/p
ython2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of test_moddn_staging_prod[6-None-cn=ac*,dc=example,dc=com-False] request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProp
erties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com
:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ERROR at setup of test_moddn_staging_prod[7-cn=st*,dc=example,dc=com-None-False]
 request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suf
fix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ld
ap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _________ ERROR at setup of test_moddn_staging_prod[8-None-None-False] _________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(
request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreeme
nt") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.
lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _________________ ERROR at setup of test_moddn_staging_prod_9 __________________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ glo
bal installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICAT
ION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/pyth
on2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN __________________ ERROR at setup of test_moddn_prod_staging ___________________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installat
ion1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: 
defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.co
m:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN __________________ ERROR at setup of test_check_repl_M2_to_M1 __________________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1
 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properti
es) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/l
dap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_moddn_staging_prod_except _______________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.al
locate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.e
ng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <h
ttp://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_mode_default_ger_no_moddn _______________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DE
PLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, 
**kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ______________ ERROR at setup of test_mode_default_ger_with_moddn ______________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_
MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib38
9/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40>
 func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _____________ ERROR at setup of test_mode_switch_default_to_legacy _____________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the st
atus of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/38
9-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = N
one e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_mode_legacy_ger_no_moddn1 _______________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the i
nstances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.
eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for seriali
zing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN _______________ ERROR at setup of test_mode_legacy_ger_no_moddn2 _______________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open()
 master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)
", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.w
rite('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ______________ ERROR at setup of test_mode_legacy_ger_with_moddn _______________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplicatio
n(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEnt
ry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) 
if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ___________________ ERROR at setup of test_rdn_write_get_ger ___________________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_M
ASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init
__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: 
try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ______________ ERROR at setup of test_rdn_write_modrdn_anonymous _______________ request = <SubRequest 'topology' for <Function 'test_aci_attr_subtype_targetattr[lang-ja]'>> @pytest.fixture(scope="module") def topology(request): """This fixture is used to create a replicated topology for the 'module'. The replicated topology is MASTER1 <-> Master2. """ global installation1_prefix global installation2_prefix # allocate master1 on a given deployement master1 = DirSrv(verbose=False) if installation1_prefix: args_instance[SER_DEPLOYED_DIR] = installation1_prefix # Args for the master1 instance args_instance[SER_HOST] = HOST_MASTER_1 args_instance[SER_PORT] = PORT_MASTER_1 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_1 args_master = args_instance.copy() master1.allocate(args_master) # allocate master1 on a given deployement master2 = DirSrv(verbose=False) if installation2_prefix: args_instance[SER_DEPLOYED_DIR] = installation2_prefix # Args for the consumer instance args_instance[SER_HOST] = HOST_MASTER_2 args_instance[SER_PORT] = PORT_MASTER_2 args_instance[SER_SERVERID_PROP] = SERVERID_MASTER_2 args_master = args_instance.copy() master2.allocate(args_master) # Get the status of the instance and restart it if it exists instance_master1 = master1.exists() instance_master2 = master2.exists() # Remove all the instances if instance_master1: master1.delete() if instance_master2: master2.delete() # Create the instances master1.create() master1.open() master2.create() master2.open() # # Now prepare the Master-Consumer topology # # First Enable replication master1.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_1) master2.replica.enableReplication(suffix=SUFFIX, role=REPLICAROLE_MASTER, replicaId=REPLICAID_MASTER_2) # Initialize the supplier->consumer properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProp
erties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} repl_agreement = master1.agreement.create(suffix=SUFFIX, host=master2.host, port=master2.port, properties=properties) if not repl_agreement: log.fatal("Fail to create a replica agreement") sys.exit(1) log.debug("%s created" % repl_agreement) properties = {RA_NAME: r'meTo_$host:$port', RA_BINDDN: defaultProperties[REPLICATION_BIND_DN], RA_BINDPW: defaultProperties[REPLICATION_BIND_PW], RA_METHOD: defaultProperties[REPLICATION_BIND_METHOD], RA_TRANSPORT_PROT: defaultProperties[REPLICATION_TRANSPORT]} master2.agreement.create(suffix=SUFFIX, host=master1.host, port=master1.port, properties=properties) master1.agreement.init(SUFFIX, HOST_MASTER_2, PORT_MASTER_2) > master1.waitForReplInit(repl_agreement) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/acl/acl_test.py>:162: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:2177: in waitForReplInit return self.replica.wait_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:596: in wait_init done, haserror = self.check_init(agmtdn) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/replica.py>:548: in check_init agmtdn, ldap.SCOPE_BASE, "(objectclass=*)", attrlist) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:1574: in getEntry restype, obj = self.result(res) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, res
p_data, resp_msgid = self.result2(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f164fd40> func = <built-in method result4 of LDAP object at 0x7fa7fa977418> args = (17, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobjec
t.py:106: SERVER_DOWN =================================== FAILURES =================================== ____________________________ test_ticket47669_init _____________________________ topology = <tickets.ticket47669_test.TopologyStandalone object at 0x7fa7fa745990> def test_ticket47669_init(topology): """ Add cn=changelog5,cn=config Enable cn=Retro Changelog Plugin,cn=plugins,cn=config """ log.info('Testing Ticket 47669 - Test duration syntax in the changelogs') # bind as directory manager topology.standalone.log.info("Bind as %s" % DN_DM) topology.standalone.simple_bind_s(DN_DM, PASSWORD) try: changelogdir = "%s/changelog" % topology.standalone.dbdir topology.standalone.add_s(Entry((CHANGELOG, {'objectclass': 'top extensibleObject'.split(), 'nsslapd-changelogdir': changelogdir}))) except ldap.LDAPError as e: log.error('Failed to add ' + CHANGELOG + ': error ' + e.message['desc']) assert False try: topology.standalone.modify_s(RETROCHANGELOG, [(ldap.MOD_REPLACE, 'nsslapd-pluginEnabled', 'on')]) except ldap.LDAPError as e: log.error('Failed to enable ' + RETROCHANGELOG + ': error ' + e.message['desc']) assert False # restart the server > topology.standalone.restart(timeout=10) tickets/ticket47669_test.py:103: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:1215: in restart self.start(timeout) ../../../lib389/lib389/__init__.py:1096: in start "dirsrv@%s" % self.serverid]) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ popenargs = (['/usr/bin/systemctl', 'start', 'dirsrv@standalone'],), kwargs = {} retcode = 1, cmd = ['/usr/bin/systemctl', 'start', 'dirsrv@standalone'] def check_call(*popenargs, **kwargs): """Run command with arguments. Wait for command to complete. If the exit code was zero then return, otherwise raise CalledProcessError. The CalledProcessError object will have the return code in the returncode attribute. The arguments are the same as for the Popen constructor. Example: check_call(["ls", "-l"]) """ r
etcode = call(*popenargs, **kwargs) if retcode: cmd = kwargs.get("args") if cmd is None: cmd = popenargs[0] > raise CalledProcessError(retcode, cmd) E CalledProcessError: Command '['/usr/bin/systemctl', 'start', 'dirsrv@standalone']' returned non-zero exit status 1 /usr/lib64/python2.7/subprocess.py:541: CalledProcessError ---------------------------- Captured stdout setup ----------------------------- OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:Testing Ticket 47669 - Test duration syntax in the changelogs INFO:lib389:Bind as cn=Directory Manager Job for dirsrv@standalone.service failed because the control process exited with error code. See "systemctl status dirsrv@standalone.service" and "journalctl -xe" for details. ______________________ test_ticket47669_changelog_maxage _______________________ topology = <tickets.ticket47669_test.TopologyStandalone object at 0x7fa7fa745990> def test_ticket47669_changelog_maxage(topology): """ Test nsslapd-changelogmaxage in cn=changelog5,cn=config """ log.info('1. Test nsslapd-changelogmaxage in cn=changelog5,cn=config') # bind as directory manager topology.standalone.log.info("Bind as %s" % DN_DM) > topology.standalone.simple_bind_s(DN_DM, PASSWORD) tickets/ticket47669_test.py:159: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:223: in simple_bind_s resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all=1,timeout=self.timeout) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result
4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f9dbb5a8> func = <built-in method result4 of LDAP object at 0x7fa7fa8c2e68> args = (13, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:1. Test nsslapd-changelogmaxage in cn=changelog5,cn=config INFO:lib389:Bind as cn=Directory Manager ___________________ test_ticket47669_changelog_triminterval ____________________ topology = <tickets.ticket47669_test.TopologyStandalone object at 0x7fa7fa745990> def test_ticket47669_changelog_triminterval(topology): """ Test nsslapd-changelogtrim-interval in cn=changelog5,cn=config """ log.info('2. Test nsslapd-changelogtrim-interval in cn=changelog5,cn=config') # bind as directory manager topology.standalone.log.info("Bind as %s" % DN_DM) > topology.standalone.simple_bind_s(DN_DM, PASSWORD) tickets/ticket47669_test.py:179: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packa
ges/ldap/ldapobject.py:222: in simple_bind_s msgid = self.simple_bind(who,cred,serverctrls,clientctrls) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:216: in simple_bind return self._ldap_call(self._l.simple_bind,who,cred,RequestControlTuples(serverctrls),RequestControlTuples(clientctrls)) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f9dbb5a8> func = <built-in method simple_bind of LDAP object at 0x7fa7fa8c2e68> args = ('cn=Directory Manager', 'password', None, None), kwargs = {} diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:2. Test nsslapd-changelogtrim-interval in cn=changelog5,cn=config INFO:lib389:Bind as cn=Directory Manager _________________ test_ticket47669_changelog_compactdbinterval _________________ topology = <tickets.ticket47669_test.TopologyStandalone object at 0x7fa7fa745990> def test_ticket47669_changelog_compactdbinterval(topology): """ Test nsslapd-changelogcompactdb-interval in cn=changelog5,cn=config """ log.info('3. Test nsslapd-changelogcompactdb-interval in cn=change
log5,cn=config') # bind as directory manager topology.standalone.log.info("Bind as %s" % DN_DM) > topology.standalone.simple_bind_s(DN_DM, PASSWORD) tickets/ticket47669_test.py:199: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:222: in simple_bind_s msgid = self.simple_bind(who,cred,serverctrls,clientctrls) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:216: in simple_bind return self._ldap_call(self._l.simple_bind,who,cred,RequestControlTuples(serverctrls),RequestControlTuples(clientctrls)) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f9dbb5a8> func = <built-in method simple_bind of LDAP object at 0x7fa7fa8c2e68> args = ('cn=Directory Manager', 'password', None, None), kwargs = {} diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:3. Test nsslapd-changelogcompactdb-interval in cn=changelog5,cn=config INFO:lib389:Bind as cn=Directory 
Manager ____________________ test_ticket47669_retrochangelog_maxage ____________________ topology = <tickets.ticket47669_test.TopologyStandalone object at 0x7fa7fa745990> def test_ticket47669_retrochangelog_maxage(topology): """ Test nsslapd-changelogmaxage in cn=Retro Changelog Plugin,cn=plugins,cn=config """ log.info('4. Test nsslapd-changelogmaxage in cn=Retro Changelog Plugin,cn=plugins,cn=config') # bind as directory manager topology.standalone.log.info("Bind as %s" % DN_DM) > topology.standalone.simple_bind_s(DN_DM, PASSWORD) tickets/ticket47669_test.py:219: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:222: in simple_bind_s msgid = self.simple_bind(who,cred,serverctrls,clientctrls) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:216: in simple_bind return self._ldap_call(self._l.simple_bind,who,cred,RequestControlTuples(serverctrls),RequestControlTuples(clientctrls)) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f9dbb5a8> func = <built-in method simple_bind of LDAP object at 0x7fa7fa8c2e68> args = ('cn=Directory Manager', 'password', None, None), kwargs = {} diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success
 = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket47669_test:4. Test nsslapd-changelogmaxage in cn=Retro Changelog Plugin,cn=plugins,cn=config INFO:lib389:Bind as cn=Directory Manager ______________________ test_ticket47823_invalid_config_1 _______________________ topology = <tickets.ticket47823_test.TopologyStandalone object at 0x7fa7f9e8f910> def test_ticket47823_invalid_config_1(topology): ''' Check that an invalid config is detected. No uniqueness enforced Using old config: arg0 is missing ''' _header(topology, "Invalid config (old): arg0 is missing") _config_file(topology, action='save') # create an invalid config without arg0 config = _build_config(topology, attr_name='cn', subtree_1=ACTIVE_DN, subtree_2=None, type_config='old', across_subtrees=False) del config.data['nsslapd-pluginarg0'] # replace 'cn' uniqueness entry try: topology.standalone.delete_s(config.dn) except ldap.NO_SUCH_OBJECT: pass topology.standalone.add_s(config) topology.standalone.getEntry(config.dn, ldap.SCOPE_BASE, "(objectclass=nsSlapdPlugin)", ALL_CONFIG_ATTRS) # Check the server did not restart topology.standalone.modify_s(DN_CONFIG, [(ldap.MOD_REPLACE, 'nsslapd-errorlog-level', '65536')]) try: > topology.standalone.restart(timeout=5) tickets/ticket47823_test.py:636: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../../lib389/lib389/__init__.py:1215: in restart self.start(timeout) ../../../lib389/lib389/__init__.py:1096: in start "dirsrv@%s" % self.serverid]) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ popenargs = (['/usr/bin/systemctl', 'start', 'dirsrv@standalone'],), kwargs = {} retcode = 1, cmd = ['/usr/bin/systemctl', 'start', 'dirsrv@standalone'] def check_call(*popenargs, **kwargs): """Run command with arguments. Wait for command to com
plete. If the exit code was zero then return, otherwise raise CalledProcessError. The CalledProcessError object will have the return code in the returncode attribute. The arguments are the same as for the Popen constructor. Example: check_call(["ls", "-l"]) """ retcode = call(*popenargs, **kwargs) if retcode: cmd = kwargs.get("args") if cmd is None: cmd = popenargs[0] > raise CalledProcessError(retcode, cmd) E CalledProcessError: Command '['/usr/bin/systemctl', 'start', 'dirsrv@standalone']' returned non-zero exit status 1 /usr/lib64/python2.7/subprocess.py:541: CalledProcessError ----------------------------- Captured stderr call ----------------------------- INFO:lib389: ############################################### INFO:lib389:####### INFO:lib389:####### Invalid config (old): arg0 is missing INFO:lib389:####### INFO:lib389:############################################### Job for dirsrv@standalone.service failed because the control process exited with error code. See "systemctl status dirsrv@standalone.service" and "journalctl -xe" for details. ______________________ test_ticket47823_invalid_config_2 _______________________ topology = <tickets.ticket47823_test.TopologyStandalone object at 0x7fa7f9e8f910> def test_ticket47823_invalid_config_2(topology): ''' Check that an invalid config is detected. No uniqueness enforced Using old config: arg1 is missing ''' _header(topology, "Invalid config (old): arg1 is missing") _config_file(topology, action='save') # create an invalid config without arg0 > config = _build_config(topology, attr_name='cn', subtree_1=ACTIVE_DN, subtree_2=None, type_config='old', across_subtrees=False) tickets/ticket47823_test.py:672: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:124: in _build_config config = _uniqueness_config_entry(topology, attr_name) tickets/ticket47823_test.py:112: in _uniqueness_config_entry 'nsslapd-pluginDescription']) ../../../lib389/lib389/__init__.py:1574: in getEntry restype, obj = self.result(res) ../.
./../lib389/lib389/__init__.py:127: in inner objtype, data = f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:503: in result resp_type, resp_data, resp_msgid = self.result2(msgid,all,timeout) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:507: in result2 resp_type, resp_data, resp_msgid, resp_ctrls = self.result3(msgid,all,timeout) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:514: in result3 resp_ctrl_classes=resp_ctrl_classes ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:521: in result4 ldap_result = self._ldap_call(self._l.result4,msgid,all,timeout,add_ctrls,add_intermediates,add_extop) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f969eea8> func = <built-in method result4 of LDAP object at 0x7fa7fa8b4800> args = (15, 1, -1, 0, 0, 0), kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: #########################
###################### INFO:lib389:####### INFO:lib389:####### Invalid config (old): arg1 is missing INFO:lib389:####### INFO:lib389:############################################### ______________________ test_ticket47823_invalid_config_3 _______________________ topology = <tickets.ticket47823_test.TopologyStandalone object at 0x7fa7f9e8f910> def test_ticket47823_invalid_config_3(topology): ''' Check that an invalid config is detected. No uniqueness enforced Using old config: arg0 is missing ''' _header(topology, "Invalid config (old): arg0 is missing but new config attrname exists") _config_file(topology, action='save') # create an invalid config without arg0 > config = _build_config(topology, attr_name='cn', subtree_1=ACTIVE_DN, subtree_2=None, type_config='old', across_subtrees=False) tickets/ticket47823_test.py:723: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:124: in _build_config config = _uniqueness_config_entry(topology, attr_name) tickets/ticket47823_test.py:112: in _uniqueness_config_entry 'nsslapd-pluginDescription']) ../../../lib389/lib389/__init__.py:1573: in getEntry res = self.search(*args, **kwargs) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:594: in search return self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f969eea8> func = <built-in method search_ext of LDAP object at 0x7fa7fa8b4800> args = ('cn=attribute uniqueness,cn=plugins,cn=config', 0, '(objectclass=nsSlapdPlugin)', ['objectClass', 'cn', 'nsslapd-pluginPath', 'nsslapd-pluginInitfunc', 'nsslapd-pluginType', 'nsslapd-plu
ginEnabled', ...], 0, None, ...) kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: ############################################### INFO:lib389:####### INFO:lib389:####### Invalid config (old): arg0 is missing but new config attrname exists INFO:lib389:####### INFO:lib389:############################################### ______________________ test_ticket47823_invalid_config_4 _______________________ topology = <tickets.ticket47823_test.TopologyStandalone object at 0x7fa7f9e8f910> def test_ticket47823_invalid_config_4(topology): ''' Check that an invalid config is detected. No uniqueness enforced Using old config: arg1 is missing ''' _header(topology, "Invalid config (old): arg1 is missing but new config exist") _config_file(topology, action='save') # create an invalid config without arg0 > config = _build_config(topology, attr_name='cn', subtree_1=ACTIVE_DN, subtree_2=None, type_config='old', across_subtrees=False) tickets/ticket47823_test.py:776: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:124: in _build_config config = _uniqueness_config_entry(topology, attr_name) tickets/ticket47823_test.py:112: in _uniqueness_config_entry 'nsslapd-pluginDescription']) ../../..
/lib389/lib389/__init__.py:1573: in getEntry res = self.search(*args, **kwargs) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:594: in search return self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f969eea8> func = <built-in method search_ext of LDAP object at 0x7fa7fa8b4800> args = ('cn=attribute uniqueness,cn=plugins,cn=config', 0, '(objectclass=nsSlapdPlugin)', ['objectClass', 'cn', 'nsslapd-pluginPath', 'nsslapd-pluginInitfunc', 'nsslapd-pluginType', 'nsslapd-pluginEnabled', ...], 0, None, ...) kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: ############################################### INFO:lib389:####### INFO:lib389:####### Invalid config (old): arg1 is missing but new config exist INFO:lib389:####### INFO:lib389:############################################### _
_____________________ test_ticket47823_invalid_config_5 _______________________ topology = <tickets.ticket47823_test.TopologyStandalone object at 0x7fa7f9e8f910> def test_ticket47823_invalid_config_5(topology): ''' Check that an invalid config is detected. No uniqueness enforced Using new config: uniqueness-attribute-name is missing ''' _header(topology, "Invalid config (new): uniqueness-attribute-name is missing") _config_file(topology, action='save') # create an invalid config without arg0 > config = _build_config(topology, attr_name='cn', subtree_1=ACTIVE_DN, subtree_2=None, type_config='new', across_subtrees=False) tickets/ticket47823_test.py:828: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:131: in _build_config config = _uniqueness_config_entry(topology, attr_name) tickets/ticket47823_test.py:112: in _uniqueness_config_entry 'nsslapd-pluginDescription']) ../../../lib389/lib389/__init__.py:1573: in getEntry res = self.search(*args, **kwargs) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:594: in search return self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f969eea8> func = <built-in method search_ext of LDAP object at 0x7fa7fa8b4800> args = ('cn=attribute uniqueness,cn=plugins,cn=config', 0, '(objectclass=nsSlapdPlugin)', ['objectClass', 'cn', 'nsslapd-pluginPath', 'nsslapd-pluginInitfunc', 'nsslapd-pluginType', 'nsslapd-pluginEnabled', ...], 0, None, ...) kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kw
args): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: ############################################### INFO:lib389:####### INFO:lib389:####### Invalid config (new): uniqueness-attribute-name is missing INFO:lib389:####### INFO:lib389:############################################### ______________________ test_ticket47823_invalid_config_6 _______________________ topology = <tickets.ticket47823_test.TopologyStandalone object at 0x7fa7f9e8f910> def test_ticket47823_invalid_config_6(topology): ''' Check that an invalid config is detected. No uniqueness enforced Using new config: uniqueness-subtrees is missing ''' _header(topology, "Invalid config (new): uniqueness-subtrees is missing") _config_file(topology, action='save') # create an invalid config without arg0 > config = _build_config(topology, attr_name='cn', subtree_1=ACTIVE_DN, subtree_2=None, type_config='new', across_subtrees=False) tickets/ticket47823_test.py:879: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:131: in _build_config config = _uniqueness_config_entry(topology, attr_name) tickets/ticket47823_test.py:112: in _uniqueness_config_entry 'nsslapd-pluginDescription']) ../../../lib389/lib389/__init__.py:1573: in getEntry res = self.search(*args, **kwargs) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python
2.7/site-packages/ldap/ldapobject.py:594: in search return self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f969eea8> func = <built-in method search_ext of LDAP object at 0x7fa7fa8b4800> args = ('cn=attribute uniqueness,cn=plugins,cn=config', 0, '(objectclass=nsSlapdPlugin)', ['objectClass', 'cn', 'nsslapd-pluginPath', 'nsslapd-pluginInitfunc', 'nsslapd-pluginType', 'nsslapd-pluginEnabled', ...], 0, None, ...) kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: ############################################### INFO:lib389:####### INFO:lib389:####### Invalid config (new): uniqueness-subtrees is missing INFO:lib389:####### INFO:lib389:############################################### ______________________ test_ticket47823_invalid_config_7 _______________________ topology = <tickets.ticket47823_test.TopologyStandalone object at 0x7fa7f9e8f910> def test_ticket4
7823_invalid_config_7(topology): ''' Check that an invalid config is detected. No uniqueness enforced Using new config: uniqueness-subtrees is missing ''' _header(topology, "Invalid config (new): uniqueness-subtrees are invalid") _config_file(topology, action='save') # create an invalid config without arg0 > config = _build_config(topology, attr_name='cn', subtree_1="this_is dummy DN", subtree_2="an other=dummy DN", type_config='new', across_subtrees=False) tickets/ticket47823_test.py:930: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tickets/ticket47823_test.py:131: in _build_config config = _uniqueness_config_entry(topology, attr_name) tickets/ticket47823_test.py:112: in _uniqueness_config_entry 'nsslapd-pluginDescription']) ../../../lib389/lib389/__init__.py:1573: in getEntry res = self.search(*args, **kwargs) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:594: in search return self.search_ext(base,scope,filterstr,attrlist,attrsonly,None,None) ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:586: in search_ext timeout,sizelimit, ../../../lib389/lib389/__init__.py:159: in inner return f(*args, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f969eea8> func = <built-in method search_ext of LDAP object at 0x7fa7fa8b4800> args = ('cn=attribute uniqueness,cn=plugins,cn=config', 0, '(objectclass=nsSlapdPlugin)', ['objectClass', 'cn', 'nsslapd-pluginPath', 'nsslapd-pluginInitfunc', 'nsslapd-pluginType', 'nsslapd-pluginEnabled', ...], 0, None, ...) kwargs = {}, diagnostic_message_success = None e = SERVER_DOWN({'desc': "Can't contact LDAP server"},) def _ldap_call(self,func,*args,**kwargs): """ Wrapper method mainly for serializing calls into OpenLDAP libs and trace logs """ self._ldap_object_lock.acquire() if __debug__: if self._trace_level>=1: 
self._trace_file.write('*** %s %s - %s\n%s\n' % ( repr(self), self._uri, '.'.join((self.__class__.__name__,func.__name__)), pprint.pformat((args,kwargs)) )) if self._trace_level>=9: traceback.print_stack(limit=self._trace_stack_limit,file=self._trace_file) diagnostic_message_success = None try: try: > result = func(*args,**kwargs) E SERVER_DOWN: {'desc': "Can't contact LDAP server"} /usr/lib64/python2.7/site-packages/ldap/ldapobject.py:106: SERVER_DOWN ----------------------------- Captured stderr call ----------------------------- INFO:lib389: ############################################### INFO:lib389:####### INFO:lib389:####### Invalid config (new): uniqueness-subtrees are invalid INFO:lib389:####### INFO:lib389:############################################### ______________________________ test_ticket48226_1 ______________________________ topology = <tickets.ticket48226_test.TopologyReplication object at 0x7fa7f9e2eb50> def test_ticket48226_1(topology): name = 'test_entry' dn = "cn=%s,%s" % (name, SUFFIX) topology.master1.add_s(Entry((dn, {'objectclass': "top person".split(), 'sn': name, 'cn': name}))) # First do an update that is replicated mods = [(ldap.MOD_ADD, 'description', '5')] topology.master1.modify_s(dn, mods) nbtry = 0 while (nbtry <= 10): try: ent = topology.master2.getEntry(dn, ldap.SCOPE_BASE, "(objectclass=*)", ['description']) if ent.hasAttr('description') and ent.getValue('description') == '5': break except ldap.NO_SUCH_OBJECT: pass nbtry = nbtry + 1 time.sleep(1) assert nbtry <= 10 # Stop M2 so that it will not receive the next update topology.master2.stop(10) # ADD a new value that is not replicated mods = [(ldap.MOD_DELETE, 'description', '5')] topology.master1.modify_s(dn, mods) # Stop M1 so that it will keep del '5' that is unknown from master2 topology.master1.stop(10) # Get the sbin directory so we know where to replace 'ns-slapd' sbin_dir = topology.master2.get_sbin_dir() # Enable valgrind if not topology.master2.has_asan(): valgrind_enable(sbin_dir) # start M2 to do the next updates t
opology.master2.start() # ADD 'description' by '5' mods = [(ldap.MOD_DELETE, 'description', '5')] topology.master2.modify_s(dn, mods) # DEL 'description' by '5' mods = [(ldap.MOD_ADD, 'description', '5')] topology.master2.modify_s(dn, mods) # sleep of purge delay so that the next update will purge the CSN_7 time.sleep(6) # ADD 'description' by '6' that purge the state info mods = [(ldap.MOD_ADD, 'description', '6')] topology.master2.modify_s(dn, mods) # Restart master1 #topology.master1.start(30) if not topology.master2.has_asan(): results_file = valgrind_get_results_file(topology.master2) # Stop master2 topology.master2.stop(30) # Check for leak if not topology.master2.has_asan(): > if valgrind_check_file(results_file, VALGRIND_LEAK_STR, 'csnset_dup'): <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48226_test.py>:220: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ results_file = 'valgrind' patterns = (' blocks are definitely lost in loss record ', 'csnset_dup') found = False, pattern_count = 2, matched_count = 0 def valgrind_check_file(results_file, *patterns): ''' Check the valgrind results file for the all the patterns @param result_file - valgrind results file (must be read after server is stopped) @param patterns - A plain text or regex pattern string args that should be searched for @return True/False - Return true if one if the patterns match a stack trace @raise IOError ''' # Verify results file if not results_file: assert False # Check the result file fo the leak text results_file = results_file.replace('\n', '') found = False pattern_count = len(patterns) matched_count = 0 > vlog = open(results_file) E IOError: [Errno 2] No such file or directory: 'valgrind' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/utils.py>:354: IOError ----------------------------- Captured stderr call ----------------------------- INFO:lib389.utils:Valgrind is already enable
d. ____________________ test_ticket48266_count_csn_evaluation _____________________ topology = <tickets.ticket48266_test.TopologyReplication object at 0x7fa7f9e086d0> entries = None def test_ticket48266_count_csn_evaluation(topology, entries): ents = topology.master1.agreement.list(suffix=SUFFIX) assert len(ents) == 1 > first_csn = _get_first_not_replicated_csn(topology) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48266_test.py>:328: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ topology = <tickets.ticket48266_test.TopologyReplication object at 0x7fa7f9e086d0> def _get_first_not_replicated_csn(topology): name = "cn=%s2,%s" % (NEW_ACCOUNT, SUFFIX) # read the first CSN that will not be replicated mod = [(ldap.MOD_REPLACE, 'telephonenumber', str(123456))] topology.master1.modify_s(name, mod) msgid = topology.master1.search_ext(name, ldap.SCOPE_SUBTREE, 'objectclass=*', ['nscpentrywsi']) rtype, rdata, rmsgid = topology.master1.result2(msgid) attrs = None for dn, raw_attrs in rdata: topology.master1.log.info("dn: %s" % dn) if 'nscpentrywsi' in raw_attrs: attrs = raw_attrs['nscpentrywsi'] assert attrs for attr in attrs: if attr.lower().startswith('telephonenumber'): break assert attr # now retrieve the CSN of the operation we are looking for csn = None topology.master1.stop(timeout=10) file_path = os.path.join(topology.master1.prefix, "var/log/dirsrv/slapd-%s/access" % topology.master1.serverid) > file_obj = open(file_path, "r") E IOError: [Errno 2] No such file or directory: '/usr/var/log/dirsrv/slapd-master_1/access' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48266_test.py>:276: IOError ----------------------------- Captured stderr call ----------------------------- INFO:lib389:dn: cn=new_account2,dc=example,dc=com __________________ test_ticket48270_homeDirectory_indexed_cis __________________ topology = <tickets.ticket48270_test.Topo
logyStandalone object at 0x7fa7f0f75f90> def test_ticket48270_homeDirectory_indexed_cis(topology): log.info("\n\nindex homeDirectory in caseIgnoreIA5Match and caseExactIA5Match") try: ent = topology.standalone.getEntry(HOMEDIRECTORY_INDEX, ldap.SCOPE_BASE) except ldap.NO_SUCH_OBJECT: topology.standalone.add_s(Entry((HOMEDIRECTORY_INDEX, { 'objectclass': "top nsIndex".split(), 'cn': HOMEDIRECTORY_CN, 'nsSystemIndex': 'false', 'nsIndexType': 'eq'}))) #log.info("attach debugger") #time.sleep(60) IGNORE_MR_NAME='caseIgnoreIA5Match' EXACT_MR_NAME='caseExactIA5Match' mod = [(ldap.MOD_REPLACE, MATCHINGRULE, (IGNORE_MR_NAME, EXACT_MR_NAME))] topology.standalone.modify_s(HOMEDIRECTORY_INDEX, mod) #topology.standalone.stop(timeout=10) log.info("successfully checked that filter with exact mr , a filter with lowercase eq is failing") #assert topology.standalone.db2index(bename=DEFAULT_BENAME, suffixes=None, attrs=['homeDirectory']) #topology.standalone.start(timeout=10) args = {TASK_WAIT: True} topology.standalone.tasks.reindex(suffix=SUFFIX, attrname='homeDirectory', args=args) log.info("Check indexing succeeded with a specified matching rule") file_path = os.path.join(topology.standalone.prefix, "var/log/dirsrv/slapd-%s/errors" % topology.standalone.serverid) > file_obj = open(file_path, "r") E IOError: [Errno 2] No such file or directory: '/usr/var/log/dirsrv/slapd-standalone/errors' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48270_test.py>:100: IOError ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket48270_test: index homeDirectory in caseIgnoreIA5Match and caseExactIA5Match INFO:tickets.ticket48270_test:successfully checked that filter with exact mr , a filter with lowercase eq is failing INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_013221 completed successfully INFO:tickets.ticket48270_test:Check indexing succeeded with a specified m
atching rule _______________________________ test_ticket48383 _______________________________ topology = <tickets.ticket48383_test.TopologyStandalone object at 0x7fa7f0f9d450> def test_ticket48383(topology): """ This test case will check that we re-alloc buffer sizes on import.c We achieve this by setting the servers dbcachesize to a stupid small value and adding huge objects to ds. Then when we run db2index, either: data stress suites tickets tmp If we are not using the re-alloc code, it will FAIL (Bad) data stress suites tickets tmp If we re-alloc properly, it all works regardless. """ topology.standalone.config.set('nsslapd-maxbersize', '200000000') topology.standalone.restart() # Create some stupid huge objects / attributes in DS. # seeAlso is indexed by default. Lets do that! # This will take a while ... data = [random.choice(string.letters) for x in xrange(10000000)] s = "".join(data) # This was here for an iteration test. i = 1 USER_DN = 'uid=user%s,ou=people,%s' % (i, DEFAULT_SUFFIX) padding = ['%s' % n for n in range(400)] user = Entry((USER_DN, { 'objectclass': 'top posixAccount person extensibleObject'.split(), 'uid': 'user%s' % (i), 'cn': 'user%s' % (i), 'uidNumber': '%s' % (i), 'gidNumber': '%s' % (i), 'homeDirectory': '/home/user%s' % (i), 'description': 'user description', 'sn' : s , 'padding' : padding , })) try: topology.standalone.add_s(user) except ldap.LDAPError as e: log.fatal('test 48383: Failed to user%s: error %s ' % (i, e.message['desc'])) assert False # Set the dbsize really low. try: topology.standalone.modify_s(DEFAULT_BENAME, [(ldap.MOD_REPLACE, 'nsslapd-cachememsize', '1')]) except ldap.LDAPError as e: log.fatal('Failed to change nsslapd-cachememsize ' + e.message['desc']) ## Does ds try and set a minimum possible value for this? ## Yes: [16/Feb/2016:16:39:18 +1000] - WARNING: cache too small, increasing to 500K bytes # Given the formula, by default, this means DS will make the buffsize 400k # So an object with a 1MB attribute should break indexing # stop the server topology.standalo
ne.stop(timeout=30) # Now export and import the DB. It's easier than db2index ... topology.standalone.db2ldif(bename=DEFAULT_BENAME, suffixes=[DEFAULT_SUFFIX], excludeSuffixes=[], encrypt=False, \ repl_data=True, outputfile='%s/ldif/%s.ldif' % (topology.standalone.dbdir,SERVERID_STANDALONE )) result = topology.standalone.ldif2db(DEFAULT_BENAME, None, None, False, '%s/ldif/%s.ldif' % (topology.standalone.dbdir,SERVERID_STANDALONE )) > assert(result) E assert False <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48383_test.py>:123: AssertionError ---------------------------- Captured stdout setup ----------------------------- OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stdout call ----------------------------- OK group dirsrv exists OK user dirsrv exists Exported ldif file: /var/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stderr call ----------------------------- CRITICAL:tickets.ticket48383_test:Failed to change nsslapd-cachememsize No such object INFO:lib389:Running script: /usr/sbin/db2ldif -Z standalone -n userRoot -s dc=example,dc=com -a /var/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif -r [06/Nov/2016:01:45:01.114302980 +0100] - DEBUG - ldbm_back_start - userRoot: entry cache size: 10485760 B; db size: 10321920 B [06/Nov/2016:01:45:01.122790760 +0100] - DEBUG - ldbm_back_start - total cache size: 20971520 B; [06/Nov/2016:01:45:01.127497484 +0100] - DEBUG - ldbm_back_start - Total entry cache size: 20971520 B; dbcache size: 10000000 B; available memory size: 2061860864 B; ldiffile: /var/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif [06/Nov/2016:01:45:01.176685068 +0100] - ERR - ldbm_back_ldbm2ldif - db2ldif: can't open /var/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif: 2 (No such file or directory) [06/Nov/2016:01:45:01.186710413 +0100] - INFO - dblayer_pre_close - Waiting for 4 database thr
eads to stop [06/Nov/2016:01:45:02.138980064 +0100] - INFO - dblayer_pre_close - All database threads now stopped ERROR:lib389:ldif2db: Can't find file: /var/lib/dirsrv/slapd-standalone/db/ldif/standalone.ldif ___________________ test_ticket48497_homeDirectory_index_run ___________________ topology = <tickets.ticket48497_test.TopologyStandalone object at 0x7fa7f9e9c050> def test_ticket48497_homeDirectory_index_run(topology): args = {TASK_WAIT: True} topology.standalone.tasks.reindex(suffix=SUFFIX, attrname='homeDirectory', args=args) log.info("Check indexing succeeded with a specified matching rule") file_path = os.path.join(topology.standalone.prefix, "var/log/dirsrv/slapd-%s/errors" % topology.standalone.serverid) > file_obj = open(file_path, "r") E IOError: [Errno 2] No such file or directory: '/usr/var/log/dirsrv/slapd-standalone/errors' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48497_test.py>:139: IOError ----------------------------- Captured stderr call ----------------------------- INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_014532 completed successfully INFO:tickets.ticket48497_test:Check indexing succeeded with a specified matching rule __________________ test_ticket48745_homeDirectory_indexed_cis __________________ topology = <tickets.ticket48745_test.TopologyStandalone object at 0x7fa7f0faae10> def test_ticket48745_homeDirectory_indexed_cis(topology): log.info("\n\nindex homeDirectory in caseIgnoreIA5Match and caseExactIA5Match") try: ent = topology.standalone.getEntry(HOMEDIRECTORY_INDEX, ldap.SCOPE_BASE) except ldap.NO_SUCH_OBJECT: topology.standalone.add_s(Entry((HOMEDIRECTORY_INDEX, { 'objectclass': "top nsIndex".split(), 'cn': HOMEDIRECTORY_CN, 'nsSystemIndex': 'false', 'nsIndexType': 'eq'}))) #log.info("attach debugger") #time.sleep(60) IGNORE_MR_NAME='caseIgnoreIA5Match' EXACT_MR_NAME='caseExactIA5Match' mod = [(ldap.MOD_REPLACE, MATCHINGRULE, (IGNORE_MR_NA
ME, EXACT_MR_NAME))] topology.standalone.modify_s(HOMEDIRECTORY_INDEX, mod) #topology.standalone.stop(timeout=10) log.info("successfully checked that filter with exact mr , a filter with lowercase eq is failing") #assert topology.standalone.db2index(bename=DEFAULT_BENAME, suffixes=None, attrs=['homeDirectory']) #topology.standalone.start(timeout=10) args = {TASK_WAIT: True} topology.standalone.tasks.reindex(suffix=SUFFIX, attrname='homeDirectory', args=args) log.info("Check indexing succeeded with a specified matching rule") file_path = os.path.join(topology.standalone.prefix, "var/log/dirsrv/slapd-%s/errors" % topology.standalone.serverid) > file_obj = open(file_path, "r") E IOError: [Errno 2] No such file or directory: '/usr/var/log/dirsrv/slapd-standalone/errors' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48745_test.py>:110: IOError ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket48745_test: index homeDirectory in caseIgnoreIA5Match and caseExactIA5Match INFO:tickets.ticket48745_test:successfully checked that filter with exact mr , a filter with lowercase eq is failing INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_014702 completed successfully INFO:tickets.ticket48745_test:Check indexing succeeded with a specified matching rule __________________ test_ticket48746_homeDirectory_indexed_cis __________________ topology = <tickets.ticket48746_test.TopologyStandalone object at 0x7fa7f9e08050> def test_ticket48746_homeDirectory_indexed_cis(topology): log.info("\n\nindex homeDirectory in caseIgnoreIA5Match and caseExactIA5Match") try: ent = topology.standalone.getEntry(HOMEDIRECTORY_INDEX, ldap.SCOPE_BASE) except ldap.NO_SUCH_OBJECT: topology.standalone.add_s(Entry((HOMEDIRECTORY_INDEX, { 'objectclass': "top nsIndex".split(), 'cn': HOMEDIRECTORY_CN, 'nsSystemIndex': 'false', 'nsIndexType': 'eq'}))) #log.info("attach debugger") #tim
e.sleep(60) IGNORE_MR_NAME='caseIgnoreIA5Match' EXACT_MR_NAME='caseExactIA5Match' mod = [(ldap.MOD_REPLACE, MATCHINGRULE, (IGNORE_MR_NAME, EXACT_MR_NAME))] topology.standalone.modify_s(HOMEDIRECTORY_INDEX, mod) #topology.standalone.stop(timeout=10) log.info("successfully checked that filter with exact mr , a filter with lowercase eq is failing") #assert topology.standalone.db2index(bename=DEFAULT_BENAME, suffixes=None, attrs=['homeDirectory']) #topology.standalone.start(timeout=10) args = {TASK_WAIT: True} topology.standalone.tasks.reindex(suffix=SUFFIX, attrname='homeDirectory', args=args) log.info("Check indexing succeeded with a specified matching rule") file_path = os.path.join(topology.standalone.prefix, "var/log/dirsrv/slapd-%s/errors" % topology.standalone.serverid) > file_obj = open(file_path, "r") E IOError: [Errno 2] No such file or directory: '/usr/var/log/dirsrv/slapd-standalone/errors' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48746_test.py>:108: IOError ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket48746_test: index homeDirectory in caseIgnoreIA5Match and caseExactIA5Match INFO:tickets.ticket48746_test:successfully checked that filter with exact mr , a filter with lowercase eq is failing INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_014731 completed successfully INFO:tickets.ticket48746_test:Check indexing succeeded with a specified matching rule __________________ test_ticket48746_homeDirectory_indexed_ces __________________ topology = <tickets.ticket48746_test.TopologyStandalone object at 0x7fa7f9e08050> def test_ticket48746_homeDirectory_indexed_ces(topology): log.info("\n\nindex homeDirectory in caseExactIA5Match, this would trigger the crash") try: ent = topology.standalone.getEntry(HOMEDIRECTORY_INDEX, ldap.SCOPE_BASE) except ldap.NO_SUCH_OBJECT: topology.standalone.add_s(Entry((HOMEDIRECTORY_INDEX, { 'ob
jectclass': "top nsIndex".split(), 'cn': HOMEDIRECTORY_CN, 'nsSystemIndex': 'false', 'nsIndexType': 'eq'}))) # log.info("attach debugger") # time.sleep(60) EXACT_MR_NAME='caseExactIA5Match' mod = [(ldap.MOD_REPLACE, MATCHINGRULE, (EXACT_MR_NAME))] topology.standalone.modify_s(HOMEDIRECTORY_INDEX, mod) #topology.standalone.stop(timeout=10) log.info("successfully checked that filter with exact mr , a filter with lowercase eq is failing") #assert topology.standalone.db2index(bename=DEFAULT_BENAME, suffixes=None, attrs=['homeDirectory']) #topology.standalone.start(timeout=10) args = {TASK_WAIT: True} topology.standalone.tasks.reindex(suffix=SUFFIX, attrname='homeDirectory', args=args) log.info("Check indexing succeeded with a specified matching rule") file_path = os.path.join(topology.standalone.prefix, "var/log/dirsrv/slapd-%s/errors" % topology.standalone.serverid) > file_obj = open(file_path, "r") E IOError: [Errno 2] No such file or directory: '/usr/var/log/dirsrv/slapd-standalone/errors' <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48746_test.py>:172: IOError ----------------------------- Captured stderr call ----------------------------- INFO:tickets.ticket48746_test: index homeDirectory in caseExactIA5Match, this would trigger the crash INFO:tickets.ticket48746_test:successfully checked that filter with exact mr , a filter with lowercase eq is failing INFO:lib389:List backend with suffix=dc=example,dc=com INFO:lib389:Index task index_homeDirectory_11062016_014733 completed successfully INFO:tickets.ticket48746_test:Check indexing succeeded with a specified matching rule _____________________ test_ticket48906_dblock_ldap_update ______________________ topology = <tickets.ticket48906_test.TopologyStandalone object at 0x7fa7ef30aa90> def test_ticket48906_dblock_ldap_update(topology): topology.standalone.log.info('###################################') topology.standalone.log.info('###') topology.standalone.log.info('### Check that after ldap 
update') topology.standalone.log.info('### - monitor contains DEFAULT') topology.standalone.log.info('### - configured contains DBLOCK_LDAP_UPDATE') topology.standalone.log.info('### - After stop dse.ldif contains DBLOCK_LDAP_UPDATE') topology.standalone.log.info('### - After stop guardian contains DEFAULT') topology.standalone.log.info('### In fact guardian should differ from config to recreate the env') topology.standalone.log.info('### Check that after restart (DBenv recreated)') topology.standalone.log.info('### - monitor contains DBLOCK_LDAP_UPDATE ') topology.standalone.log.info('### - configured contains DBLOCK_LDAP_UPDATE') topology.standalone.log.info('### - dse.ldif contains DBLOCK_LDAP_UPDATE') topology.standalone.log.info('###') topology.standalone.log.info('###################################') topology.standalone.modify_s(ldbm_config, [(ldap.MOD_REPLACE, DBLOCK_ATTR_CONFIG, DBLOCK_LDAP_UPDATE)]) _check_monitored_value(topology, DBLOCK_DEFAULT) _check_configured_value(topology, attr=DBLOCK_ATTR_CONFIG, expected_value=DBLOCK_LDAP_UPDATE, required=True) topology.standalone.stop(timeout=10) _check_dse_ldif_value(topology, attr=DBLOCK_ATTR_CONFIG, expected_value=DBLOCK_LDAP_UPDATE) > _check_guardian_value(topology, attr=DBLOCK_ATTR_GUARDIAN, expected_value=DBLOCK_DEFAULT) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:218: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ topology = <tickets.ticket48906_test.TopologyStandalone object at 0x7fa7ef30aa90> attr = 'locks', expected_value = '10000' def _check_guardian_value(topology, attr=DBLOCK_ATTR_CONFIG, expected_value=None): guardian_file = topology.standalone.dbdir + '/db/guardian' > assert(os.path.exists(guardian_file)) E assert <function exists at 0x7fa80a67a050>('/var/lib/dirsrv/slapd-standalone/db/db/guardian') E + where <function exists at 0x7fa80a67a050> = <module 'posixpath' from '/usr/lib64/python2.7/posixpath.pyc'>.exists E + wher
e <module 'posixpath' from '/usr/lib64/python2.7/posixpath.pyc'> = os.path <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:164: AssertionError ----------------------------- Captured stderr call ----------------------------- INFO:lib389:################################### INFO:lib389:### INFO:lib389:### Check that after ldap update INFO:lib389:### - monitor contains DEFAULT INFO:lib389:### - configured contains DBLOCK_LDAP_UPDATE INFO:lib389:### - After stop dse.ldif contains DBLOCK_LDAP_UPDATE INFO:lib389:### - After stop guardian contains DEFAULT INFO:lib389:### In fact guardian should differ from config to recreate the env INFO:lib389:### Check that after restart (DBenv recreated) INFO:lib389:### - monitor contains DBLOCK_LDAP_UPDATE INFO:lib389:### - configured contains DBLOCK_LDAP_UPDATE INFO:lib389:### - dse.ldif contains DBLOCK_LDAP_UPDATE INFO:lib389:### INFO:lib389:################################### _____________________ test_ticket48906_dblock_edit_update ______________________ topology = <tickets.ticket48906_test.TopologyStandalone object at 0x7fa7ef30aa90> def test_ticket48906_dblock_edit_update(topology): topology.standalone.log.info('###################################') topology.standalone.log.info('###') topology.standalone.log.info('### Check that after stop') topology.standalone.log.info('### - dse.ldif contains DBLOCK_LDAP_UPDATE') topology.standalone.log.info('### - guardian contains DBLOCK_LDAP_UPDATE') topology.standalone.log.info('### Check that edit dse+restart') topology.standalone.log.info('### - monitor contains DBLOCK_EDIT_UPDATE') topology.standalone.log.info('### - configured contains DBLOCK_EDIT_UPDATE') topology.standalone.log.info('### Check that after stop') topology.standalone.log.info('### - dse.ldif contains DBLOCK_EDIT_UPDATE') topology.standalone.log.info('### - guardian contains DBLOCK_EDIT_UPDATE') topology.standalone.log.info('###') topology.standalone.log.info('#########################
##########') topology.standalone.stop(timeout=10) _check_dse_ldif_value(topology, attr=DBLOCK_ATTR_CONFIG, expected_value=DBLOCK_LDAP_UPDATE) > _check_guardian_value(topology, attr=DBLOCK_ATTR_GUARDIAN, expected_value=DBLOCK_LDAP_UPDATE) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:243: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ topology = <tickets.ticket48906_test.TopologyStandalone object at 0x7fa7ef30aa90> attr = 'locks', expected_value = '20000' def _check_guardian_value(topology, attr=DBLOCK_ATTR_CONFIG, expected_value=None): guardian_file = topology.standalone.dbdir + '/db/guardian' > assert(os.path.exists(guardian_file)) E assert <function exists at 0x7fa80a67a050>('/var/lib/dirsrv/slapd-standalone/db/db/guardian') E + where <function exists at 0x7fa80a67a050> = <module 'posixpath' from '/usr/lib64/python2.7/posixpath.pyc'>.exists E + where <module 'posixpath' from '/usr/lib64/python2.7/posixpath.pyc'> = os.path <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:164: AssertionError ----------------------------- Captured stderr call ----------------------------- INFO:lib389:################################### INFO:lib389:### INFO:lib389:### Check that after stop INFO:lib389:### - dse.ldif contains DBLOCK_LDAP_UPDATE INFO:lib389:### - guardian contains DBLOCK_LDAP_UPDATE INFO:lib389:### Check that edit dse+restart INFO:lib389:### - monitor contains DBLOCK_EDIT_UPDATE INFO:lib389:### - configured contains DBLOCK_EDIT_UPDATE INFO:lib389:### Check that after stop INFO:lib389:### - dse.ldif contains DBLOCK_EDIT_UPDATE INFO:lib389:### - guardian contains DBLOCK_EDIT_UPDATE INFO:lib389:### INFO:lib389:################################### ________________________ test_ticket48906_dblock_robust ________________________ topology = <tickets.ticket48906_test.TopologyStandalone object at 0x7fa7ef30aa90> def test_tic
ket48906_dblock_robust(topology): topology.standalone.log.info('###################################') topology.standalone.log.info('###') topology.standalone.log.info('### Check that the following values are rejected') topology.standalone.log.info('### - negative value') topology.standalone.log.info('### - insuffisant value') topology.standalone.log.info('### - invalid value') topology.standalone.log.info('### Check that minimum value is accepted') topology.standalone.log.info('###') topology.standalone.log.info('###################################') topology.standalone.start(timeout=10) > _check_monitored_value(topology, DBLOCK_EDIT_UPDATE) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:291: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ topology = <tickets.ticket48906_test.TopologyStandalone object at 0x7fa7ef30aa90> expected_value = '40000' def _check_monitored_value(topology, expected_value): entries = topology.standalone.search_s(ldbm_monitor, ldap.SCOPE_BASE, '(objectclass=*)') > assert(entries[0].hasValue(DBLOCK_ATTR_MONITOR) and entries[0].getValue(DBLOCK_ATTR_MONITOR) == expected_value) E assert (True and '20000' == '40000' E + where True = <bound method Entry.hasValue of dn: cn=database,cn=monitor,cn=ldbm database,cn...pd-db-txn-region-wait-rate: 0\nobjectClass: top\nobjectClass: extensibleObject\n\n>('nsslapd-db-configured-locks') E + where <bound method Entry.hasValue of dn: cn=database,cn=monitor,cn=ldbm database,cn...pd-db-txn-region-wait-rate: 0\nobjectClass: top\nobjectClass: extensibleObject\n\n> = dn: cn=database,cn=monitor,cn=ldbm database,cn=plugins,cn=config\ncn: database\n...apd-db-txn-region-wait-rate: 0\nobjectClass: top\nobjectClass: extensibleObject\n\n.hasValue E - 20000 E ? ^ E + 40000 E ? ^) <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/tickets/ticket48906_test.py>:144: AssertionError ------------------------
----- Captured stderr call ----------------------------- INFO:lib389:################################### INFO:lib389:### INFO:lib389:### Check that the following values are rejected INFO:lib389:### - negative value INFO:lib389:### - insuffisant value INFO:lib389:### - invalid value INFO:lib389:### Check that minimum value is accepted INFO:lib389:### INFO:lib389:################################### INFO:lib389:open(): Connecting to uri ldap://localhost.localdomain:38931/ INFO:lib389:open(): bound as cn=Directory Manager _____________________________ test_dynamic_plugins _____________________________ topology = <test_dynamic_plugins.TopologyStandalone object at 0x7fa7f157d510> def test_dynamic_plugins(topology): """ Test Dynamic Plugins - exercise each plugin and its main features, while changing the configuration without restarting the server. Need to test: functionality, stability, and stress. These tests need to run with replication disabled, and with replication setup with a second instance. Then test if replication is working, and we have same entries on each side. Functionality - Make sure that as configuration changes are made they take effect immediately. Cross plugin interaction (e.g. automember/memberOf) needs to tested, as well as plugin tasks. Need to test plugin config validation(dependencies, etc). Memory Corruption - Restart the plugins many times, and in different orders and test functionality, and stability. This will excerise the internal plugin linked lists, dse callbacks, and task handlers. Stress - Put the server under load that will trigger multiple plugins(MO, RI, DNA, etc) Restart various plugins while these operations are going on. Perform this test 5 times(stress_max_run). """ REPLICA_PORT = 33334 RUV_FILTER = '(&(nsuniqueid=ffffffff-ffffffff-ffffffff-ffffffff)(objectclass=nstombstone))' master_maxcsn = 0 replica_maxcsn = 0 msg = ' (no replication)' replication_run = False stress_max_runs = 5 # First enable dynamic plugins try: topology.standalone.modify_s(DN_CONFIG, [(ldap.MOD_REPLACE, 'ns
slapd-dynamic-plugins', 'on')]) except ldap.LDAPError as e: ldap.fatal('Failed to enable dynamic plugin!' + e.message['desc']) assert False # Test that critical plugins can be updated even though the change might not be applied try: topology.standalone.modify_s(DN_LDBM, [(ldap.MOD_REPLACE, 'description', 'test')]) except ldap.LDAPError as e: ldap.fatal('Failed to apply change to critical plugin' + e.message['desc']) assert False while 1: # # First run the tests with replication disabled, then rerun them with replication set up # ############################################################################ # Test plugin functionality ############################################################################ log.info('####################################################################') log.info('Testing Dynamic Plugins Functionality' + msg + '...') log.info('####################################################################\n') plugin_tests.test_all_plugins(topology.standalone) log.info('####################################################################') log.info('Successfully Tested Dynamic Plugins Functionality' + msg + '.') log.info('####################################################################\n') ############################################################################ # Test the stability by exercising the internal lists, callabcks, and task handlers ############################################################################ log.info('####################################################################') log.info('Testing Dynamic Plugins for Memory Corruption' + msg + '...') log.info('####################################################################\n') prev_plugin_test = None prev_prev_plugin_test = None for plugin_test in plugin_tests.func_tests: # # Restart the plugin several times (and prev plugins) - work that linked list # plugin_test(topology.standalone, "restart") if prev_prev_plugin_test: prev_prev_plugin_test(topology.standalone, "restart") plugin_test(topology.standalone, "res
tart") if prev_plugin_test: prev_plugin_test(topology.standalone, "restart") plugin_test(topology.standalone, "restart") # Now run the functional test plugin_test(topology.standalone) # Set the previous tests if prev_plugin_test: prev_prev_plugin_test = prev_plugin_test prev_plugin_test = plugin_test log.info('####################################################################') log.info('Successfully Tested Dynamic Plugins for Memory Corruption' + msg + '.') log.info('####################################################################\n') ############################################################################ # Stress two plugins while restarting it, and while restarting other plugins. # The goal is to not crash, and have the plugins work after stressing them. ############################################################################ log.info('####################################################################') log.info('Stressing Dynamic Plugins' + msg + '...') log.info('####################################################################\n') stress_tests.configureMO(topology.standalone) stress_tests.configureRI(topology.standalone) stress_count = 0 while stress_count < stress_max_runs: log.info('####################################################################') log.info('Running stress test' + msg + '. Run (%d/%d)...' % (stress_count + 1, stress_max_runs)) log.info('####################################################################\n') try: # Launch three new threads to add a bunch of users add_users = stress_tests.AddUsers(topology.standalone, 'employee', True) add_users.start() add_users2 = stress_tests.AddUsers(topology.standalone, 'entry', True) add_users2.start() add_users3 = stress_tests.AddUsers(topology.standalone, 'person', True) add_users3.start() time.sleep(1) # While we are adding users restart the MO plugin and an idle plugin topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) time.sleep(1) topology.standalone.plugin
s.disable(name=PLUGIN_MEMBER_OF) time.sleep(1) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.disable(name=PLUGIN_LINKED_ATTRS) topology.standalone.plugins.enable(name=PLUGIN_LINKED_ATTRS) time.sleep(1) topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) time.sleep(2) topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF) time.sleep(1) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.disable(name=PLUGIN_LINKED_ATTRS) topology.standalone.plugins.enable(name=PLUGIN_LINKED_ATTRS) topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF) time.sleep(1) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) # Wait for the 'adding' threads to complete add_users.join() add_users2.join() add_users3.join() # Now launch three threads to delete the users del_users = stress_tests.DelUsers(topology.standalone, 'employee') del_users.start() del_users2 = stress_tests.DelUsers(topology.standalone, 'entry') del_users2.start() del_users3 = stress_tests.DelUsers(topology.standalone, 'person') del_users3.start() time.sleep(1) # Restart both the MO, RI plugins during these deletes, and an idle plugin topology.standalone.plugins.disable(name=PLUGIN_REFER_INTEGRITY) topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.enable(name=PLUGIN_REFER_INTEGRITY) time.sleep(1) topology.standalone.plugins.disable(name=PLUGIN_REFER_INTEGRITY) time.sleep(1) topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF) time.sleep(1) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) time.sleep(1) topology.standalone.plugins.enable(name=PLUGIN_REFER_INTEGRITY) topology.standalone.plugins.disable(name=PLUGIN_LINKED_ATTRS) topology.standalone.plugins.enable(name=PLUGIN_LINKED_ATTRS) topology.standa
lone.plugins.disable(name=PLUGIN_REFER_INTEGRITY) topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) topology.standalone.plugins.enable(name=PLUGIN_REFER_INTEGRITY) time.sleep(2) topology.standalone.plugins.disable(name=PLUGIN_REFER_INTEGRITY) time.sleep(1) topology.standalone.plugins.disable(name=PLUGIN_MEMBER_OF) time.sleep(1) topology.standalone.plugins.enable(name=PLUGIN_MEMBER_OF) time.sleep(1) topology.standalone.plugins.enable(name=PLUGIN_REFER_INTEGRITY) topology.standalone.plugins.disable(name=PLUGIN_LINKED_ATTRS) topology.standalone.plugins.enable(name=PLUGIN_LINKED_ATTRS) # Wait for the 'deleting' threads to complete del_users.join() del_users2.join() del_users3.join() # Now make sure both the MO and RI plugins still work correctly plugin_tests.func_tests[8](topology.standalone) # RI plugin plugin_tests.func_tests[5](topology.standalone) # MO plugin # Cleanup the stress tests stress_tests.cleanup(topology.standalone) except: log.info('Stress test failed!') repl_fail(replica_inst) stress_count += 1 log.info('####################################################################') log.info('Successfully Stressed Dynamic Plugins' + msg + '. Completed (%d/%d)' % (stress_count, stress_max_runs)) log.info('####################################################################\n') if replication_run: # We're done. break else: # # Enable replication and run everything one more time # log.info('Setting up replication, and rerunning the tests...\n') # Create replica instance replica_inst = DirSrv(verbose=False) args_instance[SER_HOST] = LOCALHOST args_instance[SER_PORT] = REPLICA_PORT args_instance[SER_SERVERID_PROP] = 'replica' args_instance[SER_CREATION_SUFFIX] = DEFAULT_SUFFIX args_replica_inst = args_instance.copy() replica_inst.allocate(args_replica_inst) > replica_inst.create() <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/ds/dirsrvtests/tests/suites/dynamic-plugins/test_dynamic_plugins.py>:318: _ _ _ _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:905: in create self._createDirsrv() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <lib389.DirSrv instance at 0x7fa7f131dd88> def _createDirsrv(self): """Create a new instance of directory server @param self - containing the set properties SER_HOST (host) SER_PORT (port) SER_SECURE_PORT (sslport) SER_ROOT_DN (binddn) SER_ROOT_PW (bindpw) SER_CREATION_SUFFIX (creation_suffix) SER_USER_ID (userid) SER_SERVERID_PROP (serverid) SER_GROUP_ID (groupid) SER_DEPLOYED_DIR (prefix) SER_BACKUP_INST_DIR (backupdir) SER_REALM (krb5_realm) @return None @raise None } """ DirSrvTools.lib389User(user=DEFAULT_USER) prog = os.path.join(self.ds_paths.sbin_dir, CMD_PATH_SETUP_DS) if not os.path.isfile(prog): log.error("Can't find file: %r, removing extension" % prog) prog = prog[:-3] # Create and extract a service keytab args = {SER_HOST: self.host, SER_PORT: self.port, SER_SECURE_PORT: self.sslport, SER_ROOT_DN: self.binddn, SER_ROOT_PW: self.bindpw, SER_CREATION_SUFFIX: self.creation_suffix, SER_USER_ID: self.userid, SER_SERVERID_PROP: self.serverid, SER_GROUP_ID: self.groupid, SER_DEPLOYED_DIR: self.prefix, SER_BACKUP_INST_DIR: self.backupdir, SER_STRICT_HOSTNAME_CHECKING: self.strict_hostname} content = formatInfData(args) result = DirSrvTools.runInfProg(prog, content, self.verbose, prefix=self.prefix) if result != 0: > raise Exception('Failed to run setup-ds.pl') E Exception: Failed to run setup-ds.pl <http://vm-058-081.abc.idm.lab.eng.brq.redhat.com:8080/job/389-DS-NIGHTLY/ws/source/lib389/lib389/__init__.py>:861: Exception ---------------------------- Captured stdout setup ----------------------------- OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stdout call ----------------------------- OK group dirsrv exists OK user dirsrv exists Instance slapd-passthru removed. OK group 
dirsrv exists OK user dirsrv exists Instance slapd-passthru removed. OK group dirsrv exists OK user dirsrv exists ----------------------------- Captured stderr call ----------------------------- INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Testing Dynamic Plugins Functionality (no replication)... INFO:test_dynamic_plugins:#################################################################### INFO:plugin_tests:Testing Account Policy Plugin... INFO:plugin_tests:test_acctpolicy: PASS INFO:plugin_tests:Testing attribute uniqueness... INFO:plugin_tests:test_attruniq: PASS INFO:plugin_tests:Testing Auto Membership Plugin... INFO:plugin_tests:test_automember: PASS INFO:plugin_tests:Testing Distributed Numeric Assignment Plugin... INFO:plugin_tests:test_dna: PASS INFO:plugin_tests:Testing Linked Attributes... INFO:plugin_tests:test_linkedattrs: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:plugin_tests:Testing Managed Entries... INFO:plugin_tests:test_mep: PASS INFO:plugin_tests:Testing Pass Through Authentication... INFO:lib389:List backend with suffix=dc=pass2,dc=thru INFO:lib389:Creating a local backend INFO:lib389:List backend cn=PASS2,cn=ldbm database,cn=plugins,cn=config INFO:lib389:Found entry dn: cn=PASS2,cn=ldbm database,cn=plugins,cn=config cn: PASS2 nsslapd-cachememsize: 10485760 nsslapd-cachesize: -1 nsslapd-directory: /var/lib/dirsrv/slapd-passthru/db/PASS2 nsslapd-dncachememsize: 10485760 nsslapd-readonly: off nsslapd-require-index: off nsslapd-suffix: dc=pass2,dc=thru objectClass: top objectClass: extensibleObject objectClass: nsBackendInstance INFO:lib389:Entry dn: cn="dc=pass2,dc=thru",cn=mapping tree,cn=config cn: dc=pass2,dc=thru nsslapd-backend: PASS2 nsslapd-state: backend objectclass: top objectclass: extensibleObject objectclass: nsMappingTree INFO:lib389:Found entry dn: cn=dc\3Dpass2\2Cdc\3Dthru,cn=mapping tree,cn=config cn: dc=pass2,dc=thru nsslapd-backend: PASS2 nsslapd-state:
 backend objectClass: top objectClass: extensibleObject objectClass: nsMappingTree INFO:plugin_tests:test_passthru: PASS INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing Retro Changelog Plugin... INFO:plugin_tests:test_retrocl: PASS INFO:plugin_tests:Testing RootDN Access Control... INFO:plugin_tests:test_rootdn: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Tested Dynamic Plugins Functionality (no replication). INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Testing Dynamic Plugins for Memory Corruption (no replication)... INFO:test_dynamic_plugins:#################################################################### INFO:plugin_tests:Testing Account Policy Plugin... INFO:plugin_tests:test_acctpolicy: PASS INFO:plugin_tests:Testing attribute uniqueness... INFO:plugin_tests:test_attruniq: PASS INFO:plugin_tests:Testing Auto Membership Plugin... INFO:plugin_tests:test_automember: PASS INFO:plugin_tests:Testing Distributed Numeric Assignment Plugin... INFO:plugin_tests:test_dna: PASS INFO:plugin_tests:Testing Linked Attributes... INFO:plugin_tests:test_linkedattrs: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:plugin_tests:Testing Managed Entries... INFO:plugin_tests:test_mep: PASS INFO:plugin_tests:Testing Pass Through Authentication... INFO:lib389:List backend with suffix=dc=pass2,dc=thru INFO:lib389:Creating a local backend INFO:lib389:List backend cn=PASS2,cn=ldbm database,cn=plugins,cn=config INFO:lib389:Found entry dn: cn=PASS2,cn=ldbm database,cn=plugins,cn=config cn: PASS2 nsslapd-cachememsize: 10485760 nsslapd-cachesize: -1 nsslapd-directory: /var/lib/dirsrv/slapd-passthru/db/PASS2 nsslapd-dncachememsize: 10485760 nsslapd-read
only: off nsslapd-require-index: off nsslapd-suffix: dc=pass2,dc=thru objectClass: top objectClass: extensibleObject objectClass: nsBackendInstance INFO:lib389:Entry dn: cn="dc=pass2,dc=thru",cn=mapping tree,cn=config cn: dc=pass2,dc=thru nsslapd-backend: PASS2 nsslapd-state: backend objectclass: top objectclass: extensibleObject objectclass: nsMappingTree INFO:lib389:Found entry dn: cn=dc\3Dpass2\2Cdc\3Dthru,cn=mapping tree,cn=config cn: dc=pass2,dc=thru nsslapd-backend: PASS2 nsslapd-state: backend objectClass: top objectClass: extensibleObject objectClass: nsMappingTree INFO:plugin_tests:test_passthru: PASS INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing Retro Changelog Plugin... INFO:plugin_tests:test_retrocl: PASS INFO:plugin_tests:Testing RootDN Access Control... INFO:plugin_tests:test_rootdn: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Tested Dynamic Plugins for Memory Corruption (no replication). INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Stressing Dynamic Plugins (no replication)... INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress test (no replication). Run (1/5)... INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries (employee)... INFO:stress_tests:AddUsers - Adding 250 entries (entry)... INFO:stress_tests:AddUsers - Adding 250 entries (person)... INFO:stress_tests:AddUsers - Finished adding 250 entries (employee). INFO:stress_tests:AddUsers - Finished adding 250 entries (entry). INFO:stress_te
sts:AddUsers - Finished adding 250 entries (person). INFO:stress_tests:DelUsers - Deleting 250 entries (employee)... INFO:stress_tests:DelUsers - Deleting 250 entries (person)... INFO:stress_tests:DelUsers - Deleting 250 entries (entry)... INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee). INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry). INFO:stress_tests:DelUsers - Finished deleting 250 entries (person). INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication). Completed (1/5) INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress test (no replication). Run (2/5)... INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries (employee)... INFO:stress_tests:AddUsers - Adding 250 entries (person)... INFO:stress_tests:AddUsers - Adding 250 entries (entry)... INFO:stress_tests:AddUsers - Finished adding 250 entries (employee). INFO:stress_tests:AddUsers - Finished adding 250 entries (entry). INFO:stress_tests:AddUsers - Finished adding 250 entries (person). INFO:stress_tests:DelUsers - Deleting 250 entries (employee)... INFO:stress_tests:DelUsers - Deleting 250 entries (entry)... INFO:stress_tests:DelUsers - Deleting 250 entries (person)... INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee). INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry). INFO:stress_tests:DelUsers - Finished deleting 250 entries (person). INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_test
s:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication). Completed (2/5) INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress test (no replication). Run (3/5)... INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries (entry)... INFO:stress_tests:AddUsers - Adding 250 entries (employee)... INFO:stress_tests:AddUsers - Adding 250 entries (person)... INFO:stress_tests:AddUsers - Finished adding 250 entries (employee). INFO:stress_tests:AddUsers - Finished adding 250 entries (entry). INFO:stress_tests:AddUsers - Finished adding 250 entries (person). INFO:stress_tests:DelUsers - Deleting 250 entries (employee)... INFO:stress_tests:DelUsers - Deleting 250 entries (person)... INFO:stress_tests:DelUsers - Deleting 250 entries (entry)... INFO:stress_tests:DelUsers - Finished deleting 250 entries (person). INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry). INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee). INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication). Completed (3/5) INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress 
test (no replication). Run (4/5)... INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries (employee)... INFO:stress_tests:AddUsers - Adding 250 entries (entry)... INFO:stress_tests:AddUsers - Adding 250 entries (person)... INFO:stress_tests:AddUsers - Finished adding 250 entries (person). INFO:stress_tests:AddUsers - Finished adding 250 entries (entry). INFO:stress_tests:AddUsers - Finished adding 250 entries (employee). INFO:stress_tests:DelUsers - Deleting 250 entries (employee)... INFO:stress_tests:DelUsers - Deleting 250 entries (entry)... INFO:stress_tests:DelUsers - Deleting 250 entries (person)... INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry). INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee). INFO:stress_tests:DelUsers - Finished deleting 250 entries (person). INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication). Completed (4/5) INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Running stress test (no replication). Run (5/5)... INFO:test_dynamic_plugins:#################################################################### INFO:stress_tests:AddUsers - Adding 250 entries (employee)... INFO:stress_tests:AddUsers - Adding 250 entries (entry)... INFO:stress_tests:AddUsers - Adding 250 entries (person)... INFO:stress_tests:AddUsers - Finished adding 250 entries (entry). INFO:stress_tests:AddUsers - Finished adding 250 entries (person). INFO:stress_tests:AddUsers - Finished adding 250 entries (employee). INFO:stress_tes
ts:DelUsers - Deleting 250 entries (entry)... INFO:stress_tests:DelUsers - Deleting 250 entries (person)... INFO:stress_tests:DelUsers - Deleting 250 entries (employee)... INFO:stress_tests:DelUsers - Finished deleting 250 entries (person). INFO:stress_tests:DelUsers - Finished deleting 250 entries (employee). INFO:stress_tests:DelUsers - Finished deleting 250 entries (entry). INFO:plugin_tests:Testing referential integrity postoperation... INFO:plugin_tests:test_referint: PASS INFO:plugin_tests:Testing MemberOf Plugin... INFO:plugin_tests:test_memberof: PASS INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Successfully Stressed Dynamic Plugins (no replication). Completed (5/5) INFO:test_dynamic_plugins:#################################################################### INFO:test_dynamic_plugins:Setting up replication, and rerunning the tests... ============== 24 failed, 471 passed, 27 error in 8991.82 seconds ==============
+ MSG=FAILED
+ RC=1
+ sudo /usr/sbin/sendmail mreynolds@xxxxxxxxxx
+ exit 1
Build step 'Execute shell' marked build as failure
_______________________________________________
389-devel mailing list -- 389-devel@xxxxxxxxxxxxxxxxxxxxxxx
To unsubscribe send an email to 389-devel-leave@xxxxxxxxxxxxxxxxxxxxxxx





[Index of Archives]     [Fedora Directory Announce]     [Fedora Users]     [Older Fedora Users Mail]     [Fedora Advisory Board]     [Fedora Security]     [Fedora Devel Java]     [Fedora Desktop]     [ATA RAID]     [Fedora Marketing]     [Fedora Mentors]     [Fedora Package Review]     [Fedora Art]     [Fedora Music]     [Fedora Packaging]     [CentOS]     [Fedora SELinux]     [Big List of Linux Books]     [KDE Users]     [Fedora Art]     [Fedora Docs]

  Powered by Linux