Note: This is a beta release of Red Hat Bugzilla 5.0. The data contained within is a snapshot of the live data so any changes you make will not be reflected in the production Bugzilla. Also email is disabled so feel free to test any aspect of the site that you want. File any problems you find or give feedback here.

Bug 1054140

Summary: traceback when calling hosted-engine --vm-status
Product: Red Hat Enterprise Virtualization Manager Reporter: movciari
Component: ovirt-hosted-engine-haAssignee: Martin Sivák <msivak>
Status: CLOSED NOTABUG QA Contact: Artyom <alukiano>
Severity: unspecified Docs Contact:
Priority: unspecified    
Version: 3.3.0CC: dfediuck, gklein, iheim, mavital, movciari, sherold
Target Milestone: ---   
Target Release: 3.4.0   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard: sla
Fixed In Version: Doc Type: Bug Fix
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2014-03-04 10:27:43 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: SLA RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Bug Depends On:    
Bug Blocks: 1078909, 1142926    
Description Flags
setup log
agent log
broker log none

Description movciari 2014-01-16 10:25:15 UTC
Created attachment 850985 [details]
setup log

Description of problem:
traceback when calling hosted-engine --vm-status:

Traceback (most recent call last):
  File "/usr/lib64/python2.6/", line 122, in _run_module_as_main
    "__main__", fname, loader, pkg_name)
  File "/usr/lib64/python2.6/", line 34, in _run_code
    exec code in run_globals
  File "/usr/lib/python2.6/site-packages/ovirt_hosted_engine_setup/", line 111, in <module>
    if not status_checker.print_status():
  File "/usr/lib/python2.6/site-packages/ovirt_hosted_engine_setup/", line 58, in print_status
    all_host_stats = ha_cli.get_all_host_stats()
  File "/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/client/", line 137, in get_all_host_stats
    return self.get_all_stats(self.StatModes.HOST)
  File "/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/client/", line 86, in get_all_stats
  File "/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/lib/", line 171, in get_stats_from_storage
    result = self._checked_communicate(request)
  File "/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/lib/", line 198, in _checked_communicate
    raise RequestError("Request failed: {0}".format(msg))
ovirt_hosted_engine_ha.lib.exceptions.RequestError: Request failed: success

Version-Release number of selected component (if applicable):

How reproducible:

Steps to Reproduce:

Actual results:

Expected results:

Additional info:

Comment 1 movciari 2014-01-16 10:26:03 UTC
Created attachment 850996 [details]
agent log

Comment 2 movciari 2014-01-16 10:26:25 UTC
Created attachment 850997 [details]
broker log

Comment 3 Martin Sivák 2014-01-16 15:07:25 UTC
1) Agent is not able to get sanlock lease

Exception: Failed to initialize sanlock: cannot get lock on host id 1: host already holds lock on a different host id

2) There is only this single host connected to the cluster and because this host does not have an ID and thus haven't published it's state yet

The metadata file contains only zeros:

[root@slot-6 ha_agent]# hexdump hosted-engine.metadata 
0000000 0000 0000 0000 0000 0000 0000 0000 0000

3) Because of this the get_all_stats_for_service_type returns an empty string
4) Because of this the RPC brokerlink.get_stats_from_storage returns only "success "
5) And the brokerlink._checked_communicate expects success to contain at least one additional part
6) Because it does not, it falls back to the else clause that reports Request failed: success

I think the sanlock still holds an old lease from previous test run and does not want to give it up and take a new one from a fresh set of IDs (following an environment cleanup)

Comment 4 movciari 2014-01-22 14:36:38 UTC
sanlock cleanup and new setup solved problem

Comment 5 Doron Fediuck 2014-02-10 13:59:42 UTC
(In reply to movciari from comment #4)
> sanlock cleanup and new setup solved problem

So is this still relevant? If not, we'll close it.

Comment 6 movciari 2015-06-10 11:29:16 UTC
removing old needinfo