CentOS7 desktop openstack queens

2018-12-11  本文已影响0人  余带盐
  1. 网卡配置基本格式
TYPE=Ethernet
BOOTPROTO=static
ONBOOT=yes
NAME=eth0
DEVICE=eth0
IPADDR=192.168.0.51
NETMASK=255.255.255.0
GATEWAY=192.168.0.1
DNS1=192.168.0.1

192.168.0.51       controller
192.168.0.52       compute1

reboot

为了方便,所有节点都作为chrony客户端
vi /etc/chrony.conf

#server 0.centos.pool.ntp.org iburst
#server 1.centos.pool.ntp.org iburst
#server 2.centos.pool.ntp.org iburst
#server 3.centos.pool.ntp.org iburst

server ntp1.aliyun.com iburst

yum install centos-release-openstack-queens -y
yum install python-openstackclient -y

CentOS7 minimal 关闭 firewall NetworkManager selinux
安装 openstack-selinux软件包以自动管理OpenStack服务的安全策略
yum install openstack-selinux -y

  1. 下载
    yum install mariadb mariadb-server python2-PyMySQL -y
  2. 备份配置
    cd /etc/my.cnf.d/
    tar czvf my.cnf.d.tar.gz *
  3. 创建配置openstack.cnf
    vi openstack.cnf
[mysqld]
bind-address = 192.168.0.51

default-storage-engine = innodb
innodb_file_per_table = on
max_connections = 4096
collation-server = utf8_general_ci
character-set-server = utf8
  1. 启动
    systemctl enable mariadb.service && systemctl start mariadb.service
  2. 配置数据库安装密码(可选)
    mysql_secure_installation
  3. 修改密码
    登录:mysql -u root mysql
    改密码:UPDATE user SET PASSWORD=PASSWORD('123456') where USER='root';
    FLUSH PRIVILEGES;
    退出:quit
    重启服务:systemctl restart mariadb.service
    注:发现如果不重启服务,那么密码不生效
  1. 下载
    yum install rabbitmq-server -y
  2. 启动
    systemctl enable rabbitmq-server.service && systemctl start rabbitmq-server.service
  3. 添加openstack用户
    rabbitmqctl add_user openstack 123456
    rabbitmqctl set_permissions openstack ".*" ".*" ".*"

注:用于身份的tokens的缓存

  1. 下载
    yum install memcached python-memcached -y
  2. 配置
    vi /etc/sysconfig/memcached
OPTIONS="-l 127.0.0.1,::1,controller"
  1. 启动
    systemctl enable memcached.service && systemctl start memcached.service
  1. 下载
    yum install etcd -y
  2. 配置
    vi /etc/etcd/etcd.conf
#[Member]
ETCD_DATA_DIR="/var/lib/etcd/default.etcd"
ETCD_LISTEN_PEER_URLS="http://192.168.0.51:2380"
ETCD_LISTEN_CLIENT_URLS="http://192.168.0.51:2379"
ETCD_NAME="controller"
#[Clustering]
ETCD_INITIAL_ADVERTISE_PEER_URLS="http://192.168.0.51:2380"
ETCD_ADVERTISE_CLIENT_URLS="http://192.168.0.51:2379"
ETCD_INITIAL_CLUSTER="controller=http://192.168.0.51:2380"
ETCD_INITIAL_CLUSTER_TOKEN="etcd-cluster-01"
ETCD_INITIAL_CLUSTER_STATE="new"
  1. 启动
    systemctl enable etcd && systemctl start etcd

  1. 创建数据库
    mysql -uroot -p123456
    CREATE DATABASE keystone;
    GRANT ALL PRIVILEGES ON keystone.* TO 'keystone'@'localhost' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON keystone.* TO 'keystone'@'controller' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON keystone.* TO 'keystone'@'%' IDENTIFIED BY '123456';
  2. 安装配置
    yum install openstack-keystone httpd mod_wsgi -y
    vi /etc/keystone/keystone.conf
[database]
connection = mysql+pymysql://keystone:123456@controller/keystone

[token]
provider = fernet
  1. 填充数据库
    su -s /bin/sh -c "keystone-manage db_sync" keystone
  2. 初始化Fernet密钥存储库
    keystone-manage fernet_setup --keystone-user keystone --keystone-group keystone
    keystone-manage credential_setup --keystone-user keystone --keystone-group keystone
  3. 引导身份服务
    keystone-manage bootstrap --bootstrap-password 123456 --bootstrap-admin-url http://controller:5000/v3/ --bootstrap-internal-url http://controller:5000/v3/ --bootstrap-public-url http://controller:5000/v3/ --bootstrap-region-id RegionOne
    注:keystone-manage bootstrap命令实际上是创建了default domain
  4. 配置Apache HTTP服务器
    vi /etc/httpd/conf/httpd.conf
ServerName controller

ln -s /usr/share/keystone/wsgi-keystone.conf /etc/httpd/conf.d/

  1. 启动httpd
    systemctl enable httpd.service && systemctl start httpd.service
  2. 登录admin
export OS_USERNAME=admin
export OS_PASSWORD=123456
export OS_PROJECT_NAME=admin
export OS_USER_DOMAIN_NAME=Default
export OS_PROJECT_DOMAIN_NAME=Default
export OS_AUTH_URL=http://controller:5000/v3
export OS_IDENTITY_API_VERSION=3
  1. 创建exampledomain(这个可用可不用,只是一个创建domain的示例)
    openstack domain create --description "An Example Domain" example
  2. 创建service project
    openstack project create --domain default --description "Service Project" service
    注:service project是给服务用的,每个服务会使用唯一的user添加到serviceproject中
  3. 创建非特权用户和项目(这就是创建用户、角色、项目的一般步骤,可用可不用,也是一个示例)
    openstack project create --domain default --description "Demo Project" demo
    openstack user create --domain default --password-prompt demo
    openstack role create user
    openstack role add --project demo --user demo user
    注:user角色是必须存在的,否则在管理端创建项目的时候会失败
  4. demo用户登录
export OS_PROJECT_DOMAIN_NAME=Default
export OS_USER_DOMAIN_NAME=Default
export OS_PROJECT_NAME=demo
export OS_USERNAME=demo
export OS_PASSWORD=123456
export OS_AUTH_URL=http://controller:5000/v3
export OS_IDENTITY_API_VERSION=3
export OS_IMAGE_API_VERSION=2
  1. 请求admin用户身份认证令牌
    openstack --os-auth-url http://controller:5000/v3 --os-project-domain-name Default --os-user-domain-name Default --os-project-name admin --os-username admin token issue
    注:参数跟8.登录admin是一样的,如果存在相应环境变量,那么对应参数可以忽略,比如执行了8.登录admin,那么获取token只需要使用
    openstack token issue
  2. 请求demo用户身份认证令牌
    openstack --os-auth-url http://controller:5000/v3 --os-project-domain-name Default --os-user-domain-name Default --os-project-name demo --os-username demo token issue
  1. 创建数据库
    mysql -uroot -p123456
    CREATE DATABASE glance;
    GRANT ALL PRIVILEGES ON glance.* TO 'glance'@'localhost' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON glance.* TO 'glance'@'controller' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON glance.* TO 'glance'@'%' IDENTIFIED BY '123456';
  2. 创建glance用户(其domain、project、角色)
    . admin-openrc
    openstack user create --domain default --password-prompt glance
    openstack role add --project service --user glance admin
    注:前面说过serviceproject用于服务,现在就把glance用户添加到serviceproject中,并且添加管理员角色
  3. 创建glance服务
    openstack service create --name glance --description "OpenStack Image" image
    openstack endpoint create --region RegionOne image public http://controller:9292
    openstack endpoint create --region RegionOne image internal http://controller:9292
    openstack endpoint create --region RegionOne image admin http://controller:9292
    注:openstack service createglance是服务名,image是服务type,这个应该是随意的
  4. 安装glance包
    yum install openstack-glance -y
  5. 配置
    vi /etc/glance/glance-api.conf
[database]
connection = mysql+pymysql://glance:123456@controller/glance

[keystone_authtoken]
auth_uri = http://controller:5000
auth_url = http://controller:5000
memcached_servers = controller:11211
auth_type = password
project_domain_name = Default
user_domain_name = Default
project_name = service
username = glance
password = 123456

[paste_deploy]
flavor = keystone

[glance_store]
stores = file,http
default_store = file
filesystem_store_datadir = /var/lib/glance/images/
  1. 配置/etc/glance/glance-registry.conf
    注:Glance注册服务及其API已在皇后区版本中弃用,即systemctl start openstack-glance-registry.service没用了
  2. 填充数据库
    su -s /bin/sh -c "glance-manage db_sync" glance
  3. 启动
    systemctl enable openstack-glance-api.service && systemctl start openstack-glance-api.service
  1. 创建数据库
    mysql -uroot -p123456
    CREATE DATABASE nova_api;
    CREATE DATABASE nova;
    CREATE DATABASE nova_cell0;
    GRANT ALL PRIVILEGES ON nova_api.* TO 'nova'@'localhost' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON nova_api.* TO 'nova'@'controller' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON nova_api.* TO 'nova'@'%' IDENTIFIED BY '123456';

GRANT ALL PRIVILEGES ON nova.* TO 'nova'@'localhost' IDENTIFIED BY '123456';
GRANT ALL PRIVILEGES ON nova.* TO 'nova'@'controller' IDENTIFIED BY '123456';
GRANT ALL PRIVILEGES ON nova.* TO 'nova'@'%' IDENTIFIED BY '123456';

GRANT ALL PRIVILEGES ON nova_cell0.* TO 'nova'@'localhost' IDENTIFIED BY '123456';
GRANT ALL PRIVILEGES ON nova_cell0.* TO 'nova'@'controller' IDENTIFIED BY '123456';
GRANT ALL PRIVILEGES ON nova_cell0.* TO 'nova'@'%' IDENTIFIED BY '123456';

  1. 创建nova用户
    . admin-openrc
    openstack user create --domain default --password-prompt nova
    openstack role add --project service --user nova admin
  2. 创建nova service
    openstack service create --name nova --description "OpenStack Compute" compute
    openstack endpoint create --region RegionOne compute public http://controller:8774/v2.1
    openstack endpoint create --region RegionOne compute internal http://controller:8774/v2.1
    openstack endpoint create --region RegionOne compute admin http://controller:8774/v2.1
  3. 创建placement用户
    . admin-openrc
    openstack user create --domain default --password-prompt placement
    openstack role add --project service --user placement admin
  4. 创建placement service
    openstack service create --name placement --description "Placement API" placement
    openstack endpoint create --region RegionOne placement public http://controller:8778
    openstack endpoint create --region RegionOne placement internal http://controller:8778
    openstack endpoint create --region RegionOne placement admin http://controller:8778
  5. 安装
    yum install openstack-nova-api openstack-nova-conductor openstack-nova-console openstack-nova-novncproxy openstack-nova-scheduler openstack-nova-placement-api -y
  6. 配置
    vi /etc/nova/nova.conf
[DEFAULT]
enabled_apis = osapi_compute,metadata
my_ip = 192.168.0.51
use_neutron = True
firewall_driver = nova.virt.firewall.NoopFirewallDriver
transport_url = rabbit://openstack:123456@controller

[api_database]
connection = mysql+pymysql://nova:123456@controller/nova_api

[database]
connection = mysql+pymysql://nova:123456@controller/nova

[api]
auth_strategy = keystone

[keystone_authtoken]
auth_url = http://controller:5000/v3
memcached_servers = controller:11211
auth_type = password
project_domain_name = default
user_domain_name = default
project_name = service
username = nova
password = 123456

[vnc]
enabled = true
server_listen = $my_ip
server_proxyclient_address = $my_ip

[glance]
api_servers = http://controller:9292

[oslo_concurrency]
lock_path = /var/lib/nova/tmp

[placement]
os_region_name = RegionOne
project_domain_name = Default
project_name = service
auth_type = password
user_domain_name = Default
auth_url = http://controller:5000/v3
username = placement
password = 123456

注:关于nova.virt.firewall.NoopFirewallDriver,默认情况下,Compute使用内部防火墙驱动程序。由于Networking服务包含防火墙驱动程序,因此必须使用nova.virt.firewall.NoopFirewallDriver防火墙驱动程序禁用Compute防火墙驱动 程序。nova.virt.firewall.NoopFirewallDriver firewall driver

  1. 配置httpd,启用对Placement API的访问
    vi /etc/httpd/conf.d/00-nova-placement-api.conf
<Directory /usr/bin>
  <IfVersion >= 2.4>
     Require all granted
  </IfVersion>
  <IfVersion < 2.4>
     Order allow,deny
     Allow from all
  </IfVersion>
</Directory>
  1. 重启httpd
    systemctl restart httpd
  2. 填充数据库
    su -s /bin/sh -c "nova-manage api_db sync" nova
    su -s /bin/sh -c "nova-manage cell_v2 map_cell0" nova
    su -s /bin/sh -c "nova-manage cell_v2 create_cell --name=cell1 --verbose" nova
    su -s /bin/sh -c "nova-manage db sync" nova
  3. 验证是否正确注册
    nova-manage cell_v2 list_cells
  4. 启动服务
    systemctl enable openstack-nova-api.service openstack-nova-consoleauth.service openstack-nova-scheduler.service openstack-nova-conductor.service openstack-nova-novncproxy.service && systemctl start openstack-nova-api.service openstack-nova-consoleauth.service openstack-nova-scheduler.service openstack-nova-conductor.service openstack-nova-novncproxy.service
  1. 下载
    yum install openstack-nova-compute -y
  2. 配置
    vi /etc/nova/nova.conf
[DEFAULT]
enabled_apis = osapi_compute,metadata
transport_url = rabbit://openstack:123456@controller
my_ip = 192.168.0.52
use_neutron = True
firewall_driver = nova.virt.firewall.NoopFirewallDriver

[api]
auth_strategy = keystone

[keystone_authtoken]
auth_url = http://controller:5000/v3
memcached_servers = controller:11211
auth_type = password
project_domain_name = default
user_domain_name = default
project_name = service
username = nova
password = 123456

[vnc]
enabled = True
server_listen = 0.0.0.0
server_proxyclient_address = $my_ip
novncproxy_base_url = http://controller:6080/vnc_auto.html

[glance]
api_servers = http://controller:9292

[oslo_concurrency]
lock_path = /var/lib/nova/tmp

[placement]
os_region_name = RegionOne
project_domain_name = Default
project_name = service
auth_type = password
user_domain_name = Default
auth_url = http://controller:5000/v3
username = placement
password = 123456
  1. 启动
    systemctl enable libvirtd.service openstack-nova-compute.service && systemctl restart libvirtd.service openstack-nova-compute.service
  2. 将本计算节点添加到cell数据库中(在控制节点上执行)
    . admin-openrc
    重启计算节点,否则下面的命令获取的Hostlocalhost.localdomain,更会影响下下条命令
    openstack compute service list --service nova-compute
    su -s /bin/sh -c "nova-manage cell_v2 discover_hosts --verbose" nova
  3. 验证(在控制节点上)
    openstack compute service list
  1. 创建数据库
    mysql -uroot -p123456
    CREATE DATABASE neutron;
    GRANT ALL PRIVILEGES ON neutron.* TO 'neutron'@'localhost' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON neutron.* TO 'neutron'@'controller' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON neutron.* TO 'neutron'@'%' IDENTIFIED BY '123456';
  2. 创建neutron用户
    . admin-openrc
    openstack user create --domain default --password-prompt neutron
    openstack role add --project service --user neutron admin
  3. 创建neutron service
    openstack service create --name neutron --description "OpenStack Networking" network
    openstack endpoint create --region RegionOne network public http://controller:9696
    openstack endpoint create --region RegionOne network internal http://controller:9696
    openstack endpoint create --region RegionOne network admin http://controller:9696
  4. 下载(self-service networks
    yum install openstack-neutron openstack-neutron-ml2 openstack-neutron-linuxbridge ebtables -y
  5. 配置(self-service networks
    vi /etc/neutron/neutron.conf
[DEFAULT]
core_plugin = ml2
service_plugins = router
allow_overlapping_ips = true
transport_url = rabbit://openstack:123456@controller
auth_strategy = keystone
notify_nova_on_port_status_changes = true
notify_nova_on_port_data_changes = true

[database]
connection = mysql+pymysql://neutron:123456@controller/neutron

[keystone_authtoken]
auth_uri = http://controller:5000
auth_url = http://controller:35357
memcached_servers = controller:11211
auth_type = password
project_domain_name = default
user_domain_name = default
project_name = service
username = neutron
password = 123456

[nova]
auth_url = http://controller:35357
auth_type = password
project_domain_name = default
user_domain_name = default
region_name = RegionOne
project_name = service
username = nova
password = 123456

[oslo_concurrency]
lock_path = /var/lib/neutron/tmp
  1. 配置(self-service networks
    vi /etc/neutron/plugins/ml2/ml2_conf.ini
[ml2]
type_drivers = flat,vlan,vxlan
tenant_network_types = vxlan
mechanism_drivers = linuxbridge,l2population
extension_drivers = port_security

[ml2_type_flat]
flat_networks = provider

[ml2_type_vxlan]
vni_ranges = 1:1000

[securitygroup]
enable_ipset = true
  1. 配置(self-service networks
    vi /etc/neutron/plugins/ml2/linuxbridge_agent.ini
[linux_bridge]
physical_interface_mappings = provider:ens33

[vxlan]
enable_vxlan = true
local_ip = 192.168.0.51
l2_population = true

[securitygroup]
enable_security_group = true
firewall_driver = neutron.agent.linux.iptables_firewall.IptablesFirewallDriver
  1. 配置(self-service networks
    通过验证以下所有sysctl值设置为1:确保您的Linux操作系统内核支持网桥过滤器
    vi /etc/sysctl.conf
net.bridge.bridge-nf-call-iptables = 1
net.bridge.bridge-nf-call-ip6tables = 1

重启生效:reboot
立即生效:
modprobe br_netfilter
sysctl -p
引用:br_netfilter 模块开机自动方法
CentOS 7 开机加载内核模块
注:因为openstack已经配置了br_netfilter开机自启动,所以不需要进行开机启动重复配置了

  1. 配置(self-service networks
    vi /etc/neutron/l3_agent.ini
[DEFAULT]
interface_driver = linuxbridge
  1. 配置(self-service networks
    vi /etc/neutron/dhcp_agent.ini
[DEFAULT]
interface_driver = linuxbridge
dhcp_driver = neutron.agent.linux.dhcp.Dnsmasq
enable_isolated_metadata = true
  1. 配置
    vi /etc/neutron/metadata_agent.ini
[DEFAULT]
nova_metadata_host = controller
metadata_proxy_shared_secret = 123456
  1. 配置
    vi /etc/nova/nova.conf
[neutron]
url = http://controller:9696
auth_url = http://controller:35357
auth_type = password
project_domain_name = default
user_domain_name = default
region_name = RegionOne
project_name = service
username = neutron
password = 123456
service_metadata_proxy = true
metadata_proxy_shared_secret = 123456
  1. 数据库
    ln -s /etc/neutron/plugins/ml2/ml2_conf.ini /etc/neutron/plugin.ini
    su -s /bin/sh -c "neutron-db-manage --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini upgrade head" neutron
  2. 启动
    systemctl restart openstack-nova-api.service
    systemctl enable neutron-server.service neutron-linuxbridge-agent.service neutron-dhcp-agent.service neutron-metadata-agent.service && systemctl restart neutron-server.service neutron-linuxbridge-agent.service neutron-dhcp-agent.service neutron-metadata-agent.service
  3. 启动(self-service networks
    systemctl enable neutron-l3-agent.service && systemctl restart neutron-l3-agent.service
  1. 下载
    yum install openstack-neutron-linuxbridge ebtables ipset -y
  2. 配置
    vi /etc/neutron/neutron.conf
[DEFAULT]
transport_url = rabbit://openstack:123456@controller
auth_strategy = keystone

[keystone_authtoken]
auth_uri = http://controller:5000
auth_url = http://controller:35357
memcached_servers = controller:11211
auth_type = password
project_domain_name = default
user_domain_name = default
project_name = service
username = neutron
password = 123456

[oslo_concurrency]
lock_path = /var/lib/neutron/tmp
  1. 配置(self-service networks
    vi /etc/neutron/plugins/ml2/linuxbridge_agent.ini
[linux_bridge]
physical_interface_mappings = provider:ens33

[vxlan]
enable_vxlan = true
local_ip = 192.168.0.52
l2_population = true

[securitygroup]
enable_security_group = true
firewall_driver = neutron.agent.linux.iptables_firewall.IptablesFirewallDriver
  1. 配置(self-service networks
    通过验证以下所有sysctl值设置为1:确保您的Linux操作系统内核支持网桥过滤器
    vi /etc/sysctl.conf
net.bridge.bridge-nf-call-iptables = 1
net.bridge.bridge-nf-call-ip6tables = 1

重启生效:reboot
立即生效:
modprobe br_netfilter
sysctl -p

  1. 配置
    vi /etc/nova/nova.conf
[neutron]
url = http://controller:9696
auth_url = http://controller:35357
auth_type = password
project_domain_name = default
user_domain_name = default
region_name = RegionOne
project_name = service
username = neutron
password = 123456
  1. 启动
    systemctl restart openstack-nova-compute.service
  2. 启动
    systemctl enable neutron-linuxbridge-agent.service && systemctl start neutron-linuxbridge-agent.service
  3. 验证(在控制节点)
    openstack network agent list

* cinder(在控制节点)

  1. 创建数据库
    mysql -uroot -p123456
    CREATE DATABASE cinder;
    GRANT ALL PRIVILEGES ON cinder.* TO 'cinder'@'localhost' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON cinder.* TO 'cinder'@'controller' IDENTIFIED BY '123456';
    GRANT ALL PRIVILEGES ON cinder.* TO 'cinder'@'%' IDENTIFIED BY '123456';
  2. 创建cinder用户
    . admin-openrc
    openstack user create --domain default --password-prompt cinder
    openstack role add --project service --user cinder admin
  3. 创建cinder2、cinder3服务
    openstack service create --name cinderv2 --description "OpenStack Block Storage" volumev2
    openstack service create --name cinderv3 --description "OpenStack Block Storage" volumev3
    openstack endpoint create --region RegionOne volumev2 public http://controller:8776/v2/%\(project_id\)s
    openstack endpoint create --region RegionOne volumev2 internal http://controller:8776/v2/%\(project_id\)s
    openstack endpoint create --region RegionOne volumev2 admin http://controller:8776/v2/%\(project_id\)s
    openstack endpoint create --region RegionOne volumev3 public http://controller:8776/v3/%\(project_id\)s
    openstack endpoint create --region RegionOne volumev3 internal http://controller:8776/v3/%\(project_id\)s
    openstack endpoint create --region RegionOne volumev3 admin http://controller:8776/v3/%\(project_id\)s
    注:cinder需要2个服务
  4. 下载
    yum install openstack-cinder -y
  5. 配置
    vi /etc/cinder/cinder.conf
[DEFAULT]
transport_url = rabbit://openstack:123456@controller
auth_strategy = keystone
my_ip = 192.168.0.51

[database]
connection = mysql+pymysql://cinder:123456@controller/cinder

[keystone_authtoken]
auth_uri = http://controller:5000
auth_url = http://controller:5000
memcached_servers = controller:11211
auth_type = password
project_domain_id = default
user_domain_id = default
project_name = service
username = cinder
password = 123456

[oslo_concurrency]
lock_path = /var/lib/cinder/tmp
  1. 填充数据库
    su -s /bin/sh -c "cinder-manage db sync" cinder
  2. 配置
    vi /etc/nova/nova.conf
[cinder]
os_region_name = RegionOne
  1. 重启
    systemctl restart openstack-nova-api.service
  2. 启动
    systemctl enable openstack-cinder-api.service openstack-cinder-scheduler.service && systemctl start openstack-cinder-api.service openstack-cinder-scheduler.service

* cinder-lvm(在块存储节点)

  1. 下载
    yum install lvm2 device-mapper-persistent-data -y
  2. 启动
    systemctl enable lvm2-lvmetad.service && systemctl start lvm2-lvmetad.service
  3. 创建lvm物理卷
    pvcreate /dev/vda
    注:这里的/dev/vda是一个新挂载的磁盘
  4. 创建lvm卷组cinder-volumes
    vgcreate cinder-volumes /dev/vda
    注:cinder-volumes是一个卷组,是把多个分区或磁盘合并成的一个磁盘,就是把这个提供出去,感觉块存储就是提供了一个磁盘,然后里面由openstack自己分区
  5. 配置lvm仅扫描/dev/vda
    vi /etc/lvm/lvm.conf
devices {

filter = [ "a/vda/", "r/.*/"]

配置/etc/lvm/lvm.conf的原因
注:上面aacceptrreject

* cinder(在块存储节点)

  1. 安装
    yum install openstack-cinder targetcli python-keystone -y
  2. 配置
    vi /etc/cinder/cinder.conf
[DEFAULT]
transport_url = rabbit://openstack:123456@controller
auth_strategy = keystone
my_ip = 192.168.0.53
enabled_backends = lvm
glance_api_servers = http://controller:9292

[database]
connection = mysql+pymysql://cinder:123456@controller/cinder

[keystone_authtoken]
auth_uri = http://controller:5000
auth_url = http://controller:5000
memcached_servers = controller:11211
auth_type = password
project_domain_id = default
user_domain_id = default
project_name = service
username = cinder
password = 123456

[lvm]
volume_driver = cinder.volume.drivers.lvm.LVMVolumeDriver
volume_group = cinder-volumes
iscsi_protocol = iscsi
iscsi_helper = lioadm

[oslo_concurrency]
lock_path = /var/lib/cinder/tmp

enabled_backends = lvmlvm[lvm]是关联的,enabled_backends是任意的,比如enabled_backends = lvm1,就有[lvm1]

  1. 启动
    systemctl enable openstack-cinder-volume.service target.service && systemctl start openstack-cinder-volume.service target.service
  2. 验证(在控制节点)
    openstack volume service list
  1. 下载
    yum install openstack-dashboard -y
  2. 配置
    vi /etc/openstack-dashboard/local_settings
OPENSTACK_HOST = "controller"
OPENSTACK_KEYSTONE_URL = "http://%s:5000/v3" % OPENSTACK_HOST
OPENSTACK_KEYSTONE_DEFAULT_ROLE = "user"

ALLOWED_HOSTS = ['*']

SESSION_ENGINE = 'django.contrib.sessions.backends.cache'
CACHES = {
   'default': {
        'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',
        'LOCATION': 'controller:11211',
   }
}
OPENSTACK_KEYSTONE_MULTIDOMAIN_SUPPORT = True
OPENSTACK_API_VERSIONS = {
   "identity": 3,
   "image": 2,
   "volume": 2,
}
OPENSTACK_KEYSTONE_DEFAULT_DOMAIN = "Default"
OPENSTACK_NEUTRON_NETWORK = {
   ...
   'enable_router': False,
   'enable_quotas': False,
   'enable_distributed_router': False,
   'enable_ha_router': False,
   'enable_lb': False,
   'enable_firewall': False,
   'enable_vpn': False,
   'enable_fip_topology_check': False,
}
TIME_ZONE = "UTC"

注:'enable_router': True,可以使用路由器(router),但是这个必须有self-service network

  1. 配置
    vi /etc/httpd/conf.d/openstack-dashboard.conf
WSGIApplicationGroup %{GLOBAL}
  1. 启动服务
    systemctl restart httpd.service memcached.service
  2. 访问
    http://controller/dashboard
上一篇下一篇

猜你喜欢

热点阅读