Ceph作为OpenStack后端存储
Total PGs = (Total_number_of_OSD * 100) /max_replication_count)1、Create pool for OpenStack
pg=256
ceph osd pool create volumes $pg #disk_pool
ceph osd pool create images $pg #image_pool
ceph osd pool create vms $pg #host_pool
ceph osd pool create backups $pg #backup_pool 2、setup ceph client authentication
ceph auth get-or-create client.cinder mon'allow r' osd 'allow class-read object_prefix rbd_children, allow rwxpool=volumes, allow rwx pool=vms, allow rx pool=images'
ceph auth get-or-create client.glance mon'allow r' osd 'allow class-read object_prefix rbd_children, allow rwxpool=images'
ceph auth get-or-createclient.cinder-backup mon 'allow r' osd 'allow class-read object_prefixrbd_children, allow rwx pool=backups' 3、Add the keyringsfor client.cinder, client.glance, and client.cinder-backup to the appropriatenodes and change their ownership
ceph auth get-or-create client.glance | ssh{your-glance-api-server} sudo tee /etc/ceph/ceph.client.glance.keyring
ssh {your-glance-api-server} sudo chownglance:glance /etc/ceph/ceph.client.glance.keyring
ceph auth get-or-create client.cinder | ssh{your-volume-server} sudo tee /etc/ceph/ceph.client.cinder.keyring
ssh {your-cinder-volume-server} sudo chowncinder:cinder /etc/ceph/ceph.client.cinder.keyring 4、Nodes running nova-compute need the keyring file for thenova-compute process
ceph auth get-or-create client.cinder | ssh{your-nova-compute-server} sudo tee /etc/ceph/ceph.client.cinder.keyring 5、Create a temporary copy of the secret key on the nodes runningnova-compute:
ceph auth get-key client.cinder > client.cinder.key
ceph auth get-key client.cinder | ssh{your-compute-node} tee client.cinder.key
uuidgen
4f859b95-406e-49f6-9ff8-d6e04f7ba1ef
cat > secret.xml <<EOF
<secret ephemeral='no' private='no'>
<uuid>4f859b95-406e-49f6-9ff8-d6e04f7ba1ef</uuid>
<usage type='ceph'>
<name>client.cinder secret</name>
</usage>
</secret>
EOF
sudo virsh secret-define --file secret.xml
sudo virsh secret-set-value --secret e05983be-9251-44cb-9738-198bf9ec2d7e --base64 $(cat client.cinder.key)&& rm client.cinder.key secret.xml 6、Configuring Glance
vi /etc/glance/glance-api.conf
...
default_store=rbd
rbd_store_user=glance
rbd_store_pool=images
show_image_direct_url=True
... 7、Configuring Cinder
vi /etc/cinder/cinder.conf
...
volume_driver=cinder.volume.drivers.rbd.RBDDriver
rbd_pool=volumes
rbd_ceph_conf=/etc/ceph/ceph.conf
rbd_flatten_volume_from_snapshot=false
rbd_max_clone_depth=5
glance_api_version=2
rbd_user=cinder
rbd_secret_uuid=e05983be-9251-44cb-9738-198bf9ec2d7e 8、Configuring Cinder Backup
vi /etc/cinder/cinder.conf
...
backup_driver=cinder.backup.drivers.ceph
backup_ceph_conf=/etc/ceph/ceph.conf
backup_ceph_user=cinder-backup
backup_ceph_chunk_size=134217728
backup_ceph_pool=backups
backup_ceph_stripe_unit=0
backup_ceph_stripe_count=0
restore_discard_excess_bytes=true 9、Configuring Nova
vi /etc/nova/nova.conf
...
libvirt_images_type=rbd
libvirt_images_rbd_pool=vms
libvirt_images_rbd_ceph_conf=/etc/ceph/ceph.conf
rbd_user=cinder
rbd_secret_uuid=e05983be-9251-44cb-9738-198bf9ec2d7e
libvirt_inject_password=false
libvirt_inject_key=false
libvirt_inject_partition=-2
libvirt_live_migration_flag="VIR_MIGRATE_UNDEFINE_SOURCE,VIR_MIGRATE_PEER2PEER,VIR_MIGRATE_LIVE,VIR_MIGRATE_PERSIST_DEST" 10、Restart OpenStack
service openstack-glance-api restart
service openstack-nova-compute restart
service openstack-cinder-volume restart
10、Test
页:
[1]