rados.py ‘error calling connect’ during nova.openstack.common.periodic_task caused nova-compute down
Dear folks,
There is a problem on my MOS 6.0 environment that nova-compute service was state down sometimes(about 2 times/day), it seems that compute node can't connected to ceph cluster.
Env:
Mirantis Openstack 6.0(Juno 2014.2)
3 controller, 2 compute, 3 ceph
oslo.messaging was updated by using the whole folder from fuel6.1 oslo.messaging 1.4.1.
nova service-list... and ceph -w worked very well on my environment.
Could you please check the following error messages below,
2015-07-25 02:37:02.251 13298 AUDIT nova.compute.
2015-07-25 02:37:02.252 13298 AUDIT nova.compute.
2015-07-25 02:37:02.252 13298 INFO nova.compute.
2015-07-25 02:38:02.397 13298 AUDIT nova.compute.
2015-07-25 02:38:02.418 13298 ERROR nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
2015-07-25 02:38:02.418 13298 TRACE nova.openstack.
Configuration:
root@node-
# glance.
stores = glance.
#rbd_store_
rbd_store_ceph_conf = /etc/ceph/ceph.conf
# in rbd_store_ceph_conf
#rbd_store_user = <None>
rbd_store_user = images
#rbd_store_pool = images
rbd_store_pool = images
#rbd_store_
rbd_store_
default_store=rbd
root@node-
ceph.client.
root@node-
[global]
fsid = xxxxxxxxxxxxxxx
mon_initial_members = node-2 node-3 node-4
mon_host = 10.14.xx.2 10.14.xx.3 10.14.xx.4
auth_cluster_
auth_service_
auth_client_
filestore_
log_to_syslog_level = info
log_to_syslog = True
osd_pool_
osd_pool_
log_file = /var/log/
osd_pool_
public_network = 10.14.xx.2/22
log_to_
osd_journal_size = 2048
auth_supported = cephx
osd_pool_
osd_mkfs_type = xfs
cluster_network = 10.14.xx.3/22
osd_max_backfills = 2
osd_recovery_
[client]
rbd cache writethrough until flush = True
rbd cache = True
admin socket = /var/run/
[client.
rgw_keystone_
keyring = /etc/ceph/
rgw_socket_path = /tmp/radosgw.sock
rgw_keystone_
rgw_keystone_url = 10.14.xx.1:5000
rgw_keystone_
host = node-2
rgw_dns_name = *.xxx.xxx
rgw_print_continue = True
rgw_keystone_
rgw_data = /var/lib/
Question information
- Language:
- English Edit question
- Status:
- Expired
- For:
- Ubuntu ceph Edit question
- Assignee:
- No assignee Edit question
- Last query:
- Last reply: