OpenStack私有云部署
Controller Node: em1(10.6.17.11),em2()
Computer Node: em1(10.6.17.12),em2()
Block Storage Node: em1(10.6.17.13)
Object Storage Node: em1(10.6.17.14)
1. Controller 修改网络配置
em1 =
IPADDR=10.6.17.11
NETMASK=255.255.255.0
GATEWAY=10.6.17.1
em2 = (HWADDR 与 UUID 不变)
BOOTPROTO=none
ONBOOT=yes
Computer 修改网络配置
em1 =
IPADDR=10.6.17.12
NETMASK=255.255.255.0
GATEWAY=10.6.17.1
em2 = (HWADDR 与 UUID 不变)
BOOTPROTO=none
ONBOOT=yes
2. 修改节点 主机名 方便后续查看
centos 7 使用如下命令修改:
hostnamectl --static set-hostname <主机名>
-----------------------------------------------------
3. 修改 host 文件 四台主机分别修改
vi /etc/hosts
10.6.17.11 controller
10.6.17.12 computer1
10.6.17.13 block1
10.6.17.14 object1
----------------------------------------------------
修改完毕,使用 ping -c 4 openstack.org 以及 controller computer1 block1 object1 是否可以通
4. controller 安装ntp服务!
centos 7 使用 chrony 来做时间服务器.
[root@controller ~]# yum install chrony
##修改配置文件
[root@controller ~]# vi /etc/chrony.conf
server cn.pool.ntp.org iburst
allow 10.6.17.0/24
[root@controller ~]# systemctl enable chronyd.service
[root@controller ~]# systemctl start chronyd.service
[root@controller ~]# systemctl stop firewalld.service
[root@controller ~]# systemctl disable firewalld.service
##其他节点 也分别安装 chrony
yum install chrony
##修改 配置文件
vi /etc/chrony.conf
server controller iburst
systemctl enable chronyd.service
systemctl start chronyd.service
systemctl stop firewalld.service
systemctl disable firewalld.service
##验证时间
[root@controller ~]# chronyc sources
210 Number of sources = 1
MS Name/IP address Stratum Poll Reach LastRx Last sample
===============================================================================
^* news.neu.edu.cn 2 6 17 44 -795us[ -812us] +/- 31ms
[root@computer1 ~]# chronyc sources
210 Number of sources = 1
MS Name/IP address Stratum Poll Reach LastRx Last sample
===============================================================================
^* controller 3 6 17 35 -1007ns[ -300us] +/- 33ms
[root@block1 ~]# chronyc sources
210 Number of sources = 1
MS Name/IP address Stratum Poll Reach LastRx Last sample
===============================================================================
^* controller 3 6 37 5 -8ns[ -385us] +/- 30ms
[root@object1 ~]# chronyc sources
210 Number of sources = 1
MS Name/IP address Stratum Poll Reach LastRx Last sample
===============================================================================
^* controller 3 6 37 6 -707ns[ -548us] +/- 31ms
5. 添加 yum ELPL 源 以及 openstack-liberty 源
# yum install http://dl.fedoraproject.org/pub/epel/7/x86_64/e/epel-release-7-5.noarch.rpm
# yum install centos-release-openstack-liberty
# yum upgrade
# reboot
# yum install python-openstackclient
# yum install openstack-selinux
6. controller 安装 mariadb 数据库
使用 yum 安装,或者使用 源码安装
[root@controller ~]# yum install mariadb mariadb-server MySQL-python
my.conf 中添加如下配置 以支持 utf8
[mysqld]
...
bind-address = 10.6.17.11
[mysqld]
...
default-storage-engine = innodb
innodb_file_per_table
collation-server = utf8_general_ci
init-connect = 'SET NAMES utf8'
character-set-server = utf8
# systemctl enable mariadb.service
# systemctl start mariadb.service
7. controller 安装 rabbitMQ 消息队列
[root@controller ~]# yum install rabbitmq-server
[root@controller ~]# systemctl enable rabbitmq-server.service
[root@controller ~]# systemctl start rabbitmq-server.service
## 添加 openstack 用户 ( RABBIT_PASS 替换为 自己的密码 )
[root@controller ~]# rabbitmqctl add_user openstack RABBIT_PASS
Creating user "openstack" ...
...done.
## 授权帐号
[root@controller ~]# rabbitmqctl set_permissions openstack ".*" ".*" ".*"
Setting permissions for user "openstack" in vhost "/" ...
...done.
二、 配置 Identity service 服务
## 创建 OpenStack Identity service 的 管理员 帐号
[root@controller ~]# mysql -u root -p
MariaDB [(none)]> CREATE DATABASE keystone;
MariaDB [(none)]> GRANT ALL PRIVILEGES ON keystone.* TO 'keystone'@'localhost' \
-> IDENTIFIED BY 'keystone';
MariaDB [(none)]> GRANT ALL PRIVILEGES ON keystone.* TO 'keystone'@'%' \
-> IDENTIFIED BY 'keystone';
## 创建 管理员 token
[root@controller ~]# openssl rand -hex 10
d22ea88344b5d4fa864f
## 安装openstack-keystone 组件
[root@controller ~]# yum install openstack-keystone httpd mod_wsgi memcached python-memcached
[root@controller ~]# systemctl enable memcached.service
[root@controller ~]# systemctl start memcached.service
## 编辑 /etc/keystone/keystone.conf 文件 修改 token
[root@controller ~]# vim /etc/keystone/keystone.conf
#admin_token = ADMIN
修改为 刚才生成的 token
admin_token = d22ea88344b5d4fa864f
[database]
...
connection = mysql://keystone:keystone@controller/keystone
[memcache]
...
servers = localhost:11211
[token]
...
provider = uuid
driver = memcache
[revoke]
...
driver = sql
[DEFAULT]
...
verbose = True
## 同步Identity service 数据库
[root@controller ~]# su -s /bin/sh -c "keystone-manage db_sync" keystone
提示 No handlers could be found for logger "oslo_config.cfg" 但是查看keystone.log 日志,发现已经同步完成。
## 配置 httpd 文件
[root@controller ~]# vim /etc/httpd/conf/httpd.conf 修改如下:
ServerName controller
## 创建一个 wsgi-keystone.conf 文件
[root@controller ~]# vim /etc/httpd/conf.d/wsgi-keystone.conf 内容如下:
-------------------------------------------------------------------------------------------------------------------------------------
Listen 5000
Listen 35357
<VirtualHost *:5000>
WSGIDaemonProcess keystone-public processes=5 threads=1 user=keystone group=keystone display-name=%{GROUP}
WSGIProcessGroup keystone-public
WSGIScriptAlias / /usr/bin/keystone-wsgi-public
WSGIApplicationGroup %{GLOBAL}
WSGIPassAuthorization On
<IfVersion >= 2.4>
ErrorLogFormat "%{cu}t %M"
</IfVersion>
ErrorLog /var/log/httpd/keystone-error.log
CustomLog /var/log/httpd/keystone-access.log combined
<Directory /usr/bin>
<IfVersion >= 2.4>
Require all granted
</IfVersion>
<IfVersion < 2.4>
Order allow,deny
Allow from all
</IfVersion>
</Directory>
</VirtualHost>
<VirtualHost *:35357>
WSGIDaemonProcess keystone-admin processes=5 threads=1 user=keystone group=keystone display-name=%{GROUP}
WSGIProcessGroup keystone-admin
WSGIScriptAlias / /usr/bin/keystone-wsgi-admin
WSGIApplicationGroup %{GLOBAL}
WSGIPassAuthorization On
<IfVersion >= 2.4>
ErrorLogFormat "%{cu}t %M"
</IfVersion>
ErrorLog /var/log/httpd/keystone-error.log
CustomLog /var/log/httpd/keystone-access.log combined
<Directory /usr/bin>
<IfVersion >= 2.4>
Require all granted
</IfVersion>
<IfVersion < 2.4>
Order allow,deny
Allow from all
</IfVersion>
</Directory>
</VirtualHost>
------------------------------------------------------------------------------------------------------------------------------------
## 启动 httpd
[root@controller ~]# systemctl enable httpd.service
[root@controller ~]# systemctl start httpd.service
## 配置 身份验证 token:
[root@controller ~]# export OS_TOKEN=d22ea88344b5d4fa864f
[root@controller ~]# export OS_URL=http://controller:35357/v3
[root@controller ~]# export OS_IDENTITY_API_VERSION=3
## 创建API用户
[root@controller ~]# openstack service create \
> --name keystone --description "OpenStack Identity" identity
+-------------+----------------------------------+
| Field | Value |
+-------------+----------------------------------+
| description | OpenStack Identity |
| enabled | True |
| id | 528702ea888749d8b91cdf303cbec285 |
| name | keystone |
| type | identity |
+-------------+----------------------------------+
## 创建 API访问节点
[root@controller ~]# openstack endpoint create --region RegionOne \
> identity public http://controller:5000/v2.0
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | 6d1d13e2a74940338337e2faeb404291 |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 528702ea888749d8b91cdf303cbec285 |
| service_name | keystone |
| service_type | identity |
| url | http://controller:5000/v2.0 |
+--------------+----------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> identity internal http://controller:5000/v2.0
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | 1a6e4df9f7fb444f98da5274c4b45916 |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 528702ea888749d8b91cdf303cbec285 |
| service_name | keystone |
| service_type | identity |
| url | http://controller:5000/v2.0 |
+--------------+----------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> identity admin http://controller:35357/v2.0
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | 08e839319e5143229abc893e413e0dcc |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 528702ea888749d8b91cdf303cbec285 |
| service_name | keystone |
| service_type | identity |
| url | http://controller:35357/v2.0 |
+--------------+----------------------------------+
## 创建 admin 项目
[root@controller ~]# openstack project create --domain default \
> --description "Admin Project" admin
+-------------+----------------------------------+
| Field | Value |
+-------------+----------------------------------+
| description | Admin Project |
| domain_id | default |
| enabled | True |
| id | 0d1a69b17e444cb3b4884529b6bdb372 |
| is_domain | False |
| name | admin |
| parent_id | None |
+-------------+----------------------------------+
## 创建 管理员 用户
[root@controller ~]# openstack user create --domain default \
> --password-prompt admin
User Password:
Repeat User Password:
+-----------+----------------------------------+
| Field | Value |
+-----------+----------------------------------+
| domain_id | default |
| enabled | True |
| id | 28c24be804a649cb825a8349ca3e6ce3 |
| name | admin |
+-----------+----------------------------------+
## 创建 管理员 role
[root@controller ~]# openstack role create admin
+-------+----------------------------------+
| Field | Value |
+-------+----------------------------------+
| id | c378be38e9574421a54d90e62ee2d0aa |
| name | admin |
+-------+----------------------------------+
## 关联 管理员用户 到 项目与role 中
[root@controller ~]# openstack role add --project admin --user admin admin
## 创建 service 项目
[root@controller ~]# openstack project create --domain default \
> --description "Service Project" service
+-------------+----------------------------------+
| Field | Value |
+-------------+----------------------------------+
| description | Service Project |
| domain_id | default |
| enabled | True |
| id | 90b44e1c63104b92a1fadccc2258bbd1 |
| is_domain | False |
| name | service |
| parent_id | None |
+-------------+----------------------------------+
## 验证账户
编辑 /usr/share/keystone/keystone-dist-paste.ini 文件
[root@controller ~]# vim /usr/share/keystone/keystone-dist-paste.ini
将
[pipeline:public_api]
[pipeline:admin_api]
[pipeline:api_v3]
中的
admin_token_auth 删除
## 清除环境变量
[root@controller ~]# unset OS_TOKEN OS_URL
## 以管理员身份 创建一个 token
[root@controller ~]# openstack --os-auth-url http://controller:35357/v3 \
> --os-project-domain-id default --os-user-domain-id default \
> --os-project-name admin --os-username admin --os-auth-type password \
> token issue
Password:
+------------+----------------------------------+
| Field | Value |
+------------+----------------------------------+
| expires | 2015-12-03T04:03:34.754196Z |
| id | cba3fce8d9dc4bc7956f3b4aa566051c |
| project_id | 0d1a69b17e444cb3b4884529b6bdb372 |
| user_id | 28c24be804a649cb825a8349ca3e6ce3 |
+------------+----------------------------------+
## 创建完成,表示验证通过
## 创建一个 客户端 脚本,方便操作
[root@controller ~]# vim admin-openrc.sh
添加如下环境变量:
export OS_PROJECT_DOMAIN_ID=default
export OS_USER_DOMAIN_ID=default
export OS_PROJECT_NAME=admin
export OS_TENANT_NAME=admin
export OS_USERNAME=admin
export OS_PASSWORD=admin
export OS_AUTH_URL=http://controller:35357/v3
export OS_IDENTITY_API_VERSION=3
## 导入 脚本,并测试
[root@controller ~]# source admin-openrc.sh
[root@controller ~]# openstack token issue
+------------+----------------------------------+
| Field | Value |
+------------+----------------------------------+
| expires | 2015-12-03T04:07:26.391998Z |
| id | 24a890cc24b443dc9658a34aba8462d8 |
| project_id | 0d1a69b17e444cb3b4884529b6bdb372 |
| user_id | 28c24be804a649cb825a8349ca3e6ce3 |
+------------+----------------------------------+
## 验证通过
三、安装配置 image server
## 创建并导入数据库
[root@controller ~]# mysql -u root -p
Enter password:
MariaDB [(none)]> CREATE DATABASE glance;
MariaDB [(none)]> GRANT ALL PRIVILEGES ON glance.* TO 'glance'@'localhost' \
-> IDENTIFIED BY 'glance';
MariaDB [(none)]> GRANT ALL PRIVILEGES ON glance.* TO 'glance'@'%' \
-> IDENTIFIED BY 'glance';
## 导入 脚本 创建用户
[root@controller ~]# source admin-openrc.sh
## 创建 glance 用户
[root@controller ~]# openstack user create --domain default --password-prompt glance
User Password:
Repeat User Password:
+-----------+----------------------------------+
| Field | Value |
+-----------+----------------------------------+
| domain_id | default |
| enabled | True |
| id | 19121c440d77403e8b66f197381c68e9 |
| name | glance |
+-----------+----------------------------------+
## 关联到 项目 与 admin 用户
[root@controller ~]# openstack role add --project service --user glance admin
## 创建 glance 服务
[root@controller ~]# openstack service create --name glance \
> --description "OpenStack Image service" image
+-------------+----------------------------------+
| Field | Value |
+-------------+----------------------------------+
| description | OpenStack Image service |
| enabled | True |
| id | bb8f3d6386cb423eb682147cf2b5ab92 |
| name | glance |
| type | image |
+-------------+----------------------------------+
## 创建 image API
[root@controller ~]# openstack endpoint create --region RegionOne \
> image public http://controller:9292
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | ac858877305542679883df3afb974a95 |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| service_id | bb8f3d6386cb423eb682147cf2b5ab92 |
| service_name | glance |
| service_type | image |
| url | http://controller:9292 |
+--------------+----------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> image internal http://controller:9292
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | 290dc30efb6b45c9bb654252aaf8b878 |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| service_id | bb8f3d6386cb423eb682147cf2b5ab92 |
| service_name | glance |
| service_type | image |
| url | http://controller:9292 |
+--------------+----------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> image admin http://controller:9292
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | 04fac23ce557445d92d77ca53cf4c620 |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| service_id | bb8f3d6386cb423eb682147cf2b5ab92 |
| service_name | glance |
| service_type | image |
| url | http://controller:9292 |
+--------------+----------------------------------+
## 安装配置 glance 相关软件
[root@controller ~]# yum install openstack-glance python-glance python-glanceclient -y
## 编辑 /etc/glance/glance-api.conf 配置文件
[root@controller ~]# vim /etc/glance/glance-api.conf
[database]
...
connection = mysql://glance:glance@controller/glance
[keystone_authtoken]
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = glance
password = glance
[paste_deploy]
...
flavor = keystone
[glance_store]
...
default_store = file
filesystem_store_datadir = /opt/glance/images/
[DEFAULT]
...
notification_driver = noop
[DEFAULT]
...
verbose = True
## 编辑 /etc/glance/glance-registry.conf 配置文件
[root@controller ~]# vim /etc/glance/glance-registry.conf
[database]
...
connection = mysql://glance:glance@controller/glance
[keystone_authtoken]
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = glance
password = glance
[paste_deploy]
...
flavor = keystone
[DEFAULT]
...
notification_driver = noop
[DEFAULT]
...
verbose = True
## 同步 glance 数据库
[root@controller ~]# su -s /bin/sh -c "glance-manage db_sync" glance
## 启动服务
[root@controller ~]# systemctl enable openstack-glance-api.service \
> openstack-glance-registry.service
[root@controller ~]# systemctl start openstack-glance-api.service \
> openstack-glance-registry.service
## 验证 glance 服务
添加一个环境变量到 脚本中
[root@controller ~]# echo "export OS_IMAGE_API_VERSION=2" \
> | tee -a admin-openrc.sh
[root@controller ~]# source admin-openrc.sh
## 下载一个测试镜像
wget http://download.cirros-cloud.net/0.3.4/cirros-0.3.4-x86_64-disk.img
## 上传镜像到 服务中
[root@controller ~]# glance image-create --name "cirros" \
> --file cirros-0.3.4-x86_64-disk.img \
> --disk-format qcow2 --container-format bare \
> --visibility public --progress
[=============================>] 100%
+------------------+--------------------------------------+
| Property | Value |
+------------------+--------------------------------------+
| checksum | ee1eca47dc88f4879d8a229cc70a07c6 |
| container_format | bare |
| created_at | 2015-12-03T05:05:33Z |
| disk_format | qcow2 |
| id | 3846ec1a-ad85-4b4d-88d9-17f2374ed41d |
| min_disk | 0 |
| min_ram | 0 |
| name | cirros |
| owner | 0d1a69b17e444cb3b4884529b6bdb372 |
| protected | False |
| size | 13287936 |
| status | active |
| tags | [] |
| updated_at | 2015-12-03T05:05:33Z |
| virtual_size | None |
| visibility | public |
+------------------+--------------------------------------+
[root@controller ~]# glance image-list
+--------------------------------------+--------+
| ID | Name |
+--------------------------------------+--------+
| 3846ec1a-ad85-4b4d-88d9-17f2374ed41d | cirros |
+--------------------------------------+--------+
四、安装配置 Computer server
## 添加数据库 与 用户
[root@controller ~]# mysql -u root -p
Enter password:
MariaDB [(none)]> CREATE DATABASE nova;
MariaDB [(none)]> GRANT ALL PRIVILEGES ON nova.* TO 'nova'@'localhost' \
-> IDENTIFIED BY 'nova';
MariaDB [(none)]> GRANT ALL PRIVILEGES ON nova.* TO 'nova'@'%' \
-> IDENTIFIED BY 'nova';
## 创建 nova 用户
[root@controller ~]# source admin-openrc.sh
[root@controller ~]# openstack user create --domain default --password-prompt nova
User Password:
Repeat User Password:
+-----------+----------------------------------+
| Field | Value |
+-----------+----------------------------------+
| domain_id | default |
| enabled | True |
| id | 05dfb9b8091b4d708455281540a030d1 |
| name | nova |
+-----------+----------------------------------+
## 关联 到 role
[root@controller ~]# openstack role add --project service --user nova admin
## 创建 nova service
[root@controller ~]# openstack service create --name nova \
> --description "OpenStack Compute" compute
+-------------+----------------------------------+
| Field | Value |
+-------------+----------------------------------+
| description | OpenStack Compute |
| enabled | True |
| id | 1f73377bd0bc4f208ca6c904a71e6279 |
| name | nova |
| type | compute |
+-------------+----------------------------------+
## 创建 Compute service API
[root@controller ~]# openstack endpoint create --region RegionOne \
> compute public http://controller:8774/v2/%\(tenant_id\)s
+--------------+-----------------------------------------+
| Field | Value |
+--------------+-----------------------------------------+
| enabled | True |
| id | b545b7bd81ed4b64b559b905dd3b4532 |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 1f73377bd0bc4f208ca6c904a71e6279 |
| service_name | nova |
| service_type | compute |
| url | http://controller:8774/v2/%(tenant_id)s |
+--------------+-----------------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> compute internal http://controller:8774/v2/%\(tenant_id\)s
+--------------+-----------------------------------------+
| Field | Value |
+--------------+-----------------------------------------+
| enabled | True |
| id | 5b23e9f26e094f8f99a4416a085b6fc8 |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 1f73377bd0bc4f208ca6c904a71e6279 |
| service_name | nova |
| service_type | compute |
| url | http://controller:8774/v2/%(tenant_id)s |
+--------------+-----------------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> compute admin http://controller:8774/v2/%\(tenant_id\)s
+--------------+-----------------------------------------+
| Field | Value |
+--------------+-----------------------------------------+
| enabled | True |
| id | 77638cfde23049cd9bf1c06347385562 |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 1f73377bd0bc4f208ca6c904a71e6279 |
| service_name | nova |
| service_type | compute |
| url | http://controller:8774/v2/%(tenant_id)s |
+--------------+-----------------------------------------+
## controller 安装软件
[root@computer1 ~]# yum install openstack-nova-api openstack-nova-cert \
> openstack-nova-conductor openstack-nova-console \
> openstack-nova-novncproxy openstack-nova-scheduler \
> python-novaclient
## 编辑 /etc/nova/nova.conf 配置文件
[database]
...
connection = mysql://nova:nova@controller/nova
[DEFAULT]
...
rpc_backend = rabbit
my_ip = 10.6.17.11
auth_strategy = keystone
network_api_class = nova.network.neutronv2.api.API
security_group_api = neutron
linuxnet_interface_driver = nova.network.linux_net.NeutronLinuxBridgeInterfaceDriver
firewall_driver = nova.virt.firewall.NoopFirewallDriver
enabled_apis=osapi_compute,metadata
verbose = True
[oslo_messaging_rabbit]
...
rabbit_host = controller
rabbit_userid = openstack
rabbit_password = openstack
[keystone_authtoken]
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = nova
password = nova
[glance]
...
host = controller
[oslo_concurrency]
...
lock_path = /var/lib/nova/tmp
## 同步数据库
su -s /bin/sh -c "nova-manage db sync" nova
## 启动服务
[root@controller ~]# systemctl enable openstack-nova-api.service \
> openstack-nova-cert.service openstack-nova-consoleauth.service \
> openstack-nova-scheduler.service openstack-nova-conductor.service \
> openstack-nova-novncproxy.service
[root@controller ~]# systemctl start openstack-nova-api.service \
> openstack-nova-cert.service openstack-nova-consoleauth.service \
> openstack-nova-scheduler.service openstack-nova-conductor.service \
> openstack-nova-novncproxy.service
## 安装配置 计算机节点 Compute 节点
yum install openstack-nova-compute sysfsutils
## 编辑 /etc/nova/nova.conf 配置文件
[DEFAULT]
...
rpc_backend = rabbit
auth_strategy = keystone
my_ip = 10.6.17.12
network_api_class = nova.network.neutronv2.api.API
security_group_api = neutron
linuxnet_interface_driver = nova.network.linux_net.NeutronLinuxBridgeInterfaceDriver
firewall_driver = nova.virt.firewall.NoopFirewallDriver
verbose = True
[oslo_messaging_rabbit]
...
rabbit_host = controller
rabbit_userid = openstack
rabbit_password = openstack
[keystone_authtoken]
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = nova
password = nova
[glance]
...
host = controller
[oslo_concurrency]
...
lock_path = /var/lib/nova/tmp
## 查看否支持 虚拟化 大于0 表示支持, 大于0 不需要添加配置支持
[root@computer1 ~]# egrep -c '(vmx|svm)' /proc/cpuinfo
8
## 如果 硬件不支持虚拟化 则添加虚拟化支持
[root@computer1 ~]# vim /etc/nova/nova.conf
[libvirt]
...
virt_type = qemu
## 启动服务
[root@computer1 ~]# systemctl enable libvirtd.service openstack-nova-compute.service
[root@computer1 ~]# systemctl start libvirtd.service openstack-nova-compute.service
## 验证 服务
[root@controller ~]# nova service-list
+----+------------------+------------+----------+---------+-------+----------------------------+-----------------+
| Id | Binary | Host | Zone | Status | State | Updated_at | Disabled Reason |
+----+------------------+------------+----------+---------+-------+----------------------------+-----------------+
| 1 | nova-consoleauth | controller | internal | enabled | up | 2015-12-03T07:06:27.000000 | - |
| 2 | nova-conductor | controller | internal | enabled | up | 2015-12-03T07:06:27.000000 | - |
| 5 | nova-cert | controller | internal | enabled | up | 2015-12-03T07:06:27.000000 | - |
| 6 | nova-scheduler | controller | internal | enabled | up | 2015-12-03T07:06:27.000000 | - |
| 7 | nova-compute | computer1 | nova | enabled | up | 2015-12-03T07:06:31.000000 | - |
+----+------------------+------------+----------+---------+-------+----------------------------+-----------------+
[root@controller ~]# nova endpoints
+-----------+------------------------------------------------------------+
| nova | Value |
+-----------+------------------------------------------------------------+
| id | 5b23e9f26e094f8f99a4416a085b6fc8 |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| url | http://controller:8774/v2/0d1a69b17e444cb3b4884529b6bdb372 |
+-----------+------------------------------------------------------------+
+-----------+------------------------------------------------------------+
| nova | Value |
+-----------+------------------------------------------------------------+
| id | 77638cfde23049cd9bf1c06347385562 |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| url | http://controller:8774/v2/0d1a69b17e444cb3b4884529b6bdb372 |
+-----------+------------------------------------------------------------+
+-----------+------------------------------------------------------------+
| nova | Value |
+-----------+------------------------------------------------------------+
| id | b545b7bd81ed4b64b559b905dd3b4532 |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| url | http://controller:8774/v2/0d1a69b17e444cb3b4884529b6bdb372 |
+-----------+------------------------------------------------------------+
+-----------+----------------------------------+
| keystone | Value |
+-----------+----------------------------------+
| id | 08e839319e5143229abc893e413e0dcc |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| url | http://controller:35357/v2.0 |
+-----------+----------------------------------+
+-----------+----------------------------------+
| keystone | Value |
+-----------+----------------------------------+
| id | 1a6e4df9f7fb444f98da5274c4b45916 |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| url | http://controller:5000/v2.0 |
+-----------+----------------------------------+
+-----------+----------------------------------+
| keystone | Value |
+-----------+----------------------------------+
| id | 6d1d13e2a74940338337e2faeb404291 |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| url | http://controller:5000/v2.0 |
+-----------+----------------------------------+
+-----------+----------------------------------+
| glance | Value |
+-----------+----------------------------------+
| id | 04fac23ce557445d92d77ca53cf4c620 |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| url | http://controller:9292 |
+-----------+----------------------------------+
+-----------+----------------------------------+
| glance | Value |
+-----------+----------------------------------+
| id | 290dc30efb6b45c9bb654252aaf8b878 |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| url | http://controller:9292 |
+-----------+----------------------------------+
+-----------+----------------------------------+
| glance | Value |
+-----------+----------------------------------+
| id | ac858877305542679883df3afb974a95 |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| url | http://controller:9292 |
+-----------+----------------------------------+
[root@controller ~]# nova image-list
+--------------------------------------+--------+--------+--------+
| ID | Name | Status | Server |
+--------------------------------------+--------+--------+--------+
| 3846ec1a-ad85-4b4d-88d9-17f2374ed41d | cirros | ACTIVE | |
+--------------------------------------+--------+--------+--------+
五、安装配置 Network server
## 添加数据库、用户
[root@controller ~]# mysql -u root -p
Enter password:
MariaDB [(none)]> CREATE DATABASE neutron;
MariaDB [(none)]> GRANT ALL PRIVILEGES ON neutron.* TO 'neutron'@'localhost' \
-> IDENTIFIED BY 'neutron';
MariaDB [(none)]> GRANT ALL PRIVILEGES ON neutron.* TO 'neutron'@'%' \
-> IDENTIFIED BY 'neutron';
## 添加 openstack 用户
[root@controller ~]# openstack user create --domain default --password-prompt neutron
User Password:
Repeat User Password:
+-----------+----------------------------------+
| Field | Value |
+-----------+----------------------------------+
| domain_id | default |
| enabled | True |
| id | 04b2950e0b96451494a4391c5c7bcd2e |
| name | neutron |
+-----------+----------------------------------+
[root@controller ~]# openstack role add --project service --user neutron admin
## 添加 neutron service
[root@controller ~]# openstack service create --name neutron \
> --description "OpenStack Networking" network
+-------------+----------------------------------+
| Field | Value |
+-------------+----------------------------------+
| description | OpenStack Networking |
| enabled | True |
| id | b9ca3cbea4df4966bbfff153e676d404 |
| name | neutron |
| type | network |
+-------------+----------------------------------+
## 添加 network server API
[root@controller ~]# openstack endpoint create --region RegionOne \
> network public http://controller:9696
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | e741a15b8f4f424d8ac605083d079911 |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| service_id | b9ca3cbea4df4966bbfff153e676d404 |
| service_name | neutron |
| service_type | network |
| url | http://controller:9696 |
+--------------+----------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> network internal http://controller:9696
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | ba9df6bd296f4f42bd24588b292ed54b |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| service_id | b9ca3cbea4df4966bbfff153e676d404 |
| service_name | neutron |
| service_type | network |
| url | http://controller:9696 |
+--------------+----------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> network admin http://controller:9696
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | ba86430058a243da8c50ab9d42178eab |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| service_id | b9ca3cbea4df4966bbfff153e676d404 |
| service_name | neutron |
| service_type | network |
| url | http://controller:9696 |
+--------------+----------------------------------+
## 安装配置 network 软件 ( 我这里选择的是 Networking Option 1: Provider networks 这种网络方式 )
[root@controller ~]# yum install openstack-neutron openstack-neutron-ml2 \
> openstack-neutron-linuxbridge python-neutronclient ebtables ipset
## 修改配置文件 /etc/neutron/neutron.conf
[database]
...
connection = mysql://neutron:neutron@controller/neutron
[DEFAULT]
...
core_plugin = ml2
service_plugins =
rpc_backend = rabbit
auth_strategy = keystone
notify_nova_on_port_status_changes = True
notify_nova_on_port_data_changes = True
nova_url = http://controller:8774/v2
verbose = True
[oslo_messaging_rabbit]
...
rabbit_host = controller
rabbit_userid = openstack
rabbit_password = openstack
## 检查 [keystone_authtoken] 下面配置,注释掉其他所有的选项只使用下面
[keystone_authtoken]
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = neutron
password = neutron
[nova]
...
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
region_name = RegionOne
project_name = service
username = nova
password = nova
[oslo_concurrency]
...
lock_path = /var/lib/neutron/tmp
## 配置 Layer 2 (ML2) 插件
编辑 /etc/neutron/plugins/ml2/ml2_conf.ini 配置文件
[root@controller ~]# vim /etc/neutron/plugins/ml2/ml2_conf.ini
[ml2]
...
type_drivers = flat,vlan
tenant_network_types =
mechanism_drivers = linuxbridge
extension_drivers = port_security
[ml2_type_flat]
...
flat_networks = public
[securitygroup]
...
enable_ipset = True
## 配置Linux bridge
编辑 /etc/neutron/plugins/ml2/linuxbridge_agent.ini 配置文件
[linux_bridge]
physical_interface_mappings = public:em2
[vxlan]
enable_vxlan = False
[agent]
...
prevent_arp_spoofing = True
[securitygroup]
...
enable_security_group = True
firewall_driver = neutron.agent.linux.iptables_firewall.IptablesFirewallDriver
## 配置 DHCP 服务
编辑 /etc/neutron/dhcp_agent.ini 配置文件
[DEFAULT]
...
interface_driver = neutron.agent.linux.interface.BridgeInterfaceDriver
dhcp_driver = neutron.agent.linux.dhcp.Dnsmasq
enable_isolated_metadata = True
verbose = True
## 配置 metadata 服务
编辑 /etc/neutron/metadata_agent.ini 配置文件 注释掉 [DEFAULT] 下面的选项,添加如下选项
[metadata_proxy_shared_secret = METADATA_SECRET # METADATA_SECRET 替换为自己的密码 ]
[DEFAULT]
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_region = RegionOne
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = neutron
password = neutron
nova_metadata_ip = controller
metadata_proxy_shared_secret = metadata
verbose = True
## 配置 Networking 服务
编辑 /etc/nova/nova.conf 配置文件
[metadata_proxy_shared_secret = METADATA_SECRET # METADATA_SECRET 替换为自己的密码 ]
[neutron]
...
url = http://controller:9696
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
region_name = RegionOne
project_name = service
username = neutron
password = neutron
service_metadata_proxy = True
metadata_proxy_shared_secret = metadata
## 创建一些链接
[root@controller ~]# ln -s /etc/neutron/plugins/ml2/ml2_conf.ini /etc/neutron/plugin.ini
## 同步数据库
[root@controller ~]# su -s /bin/sh -c "neutron-db-manage --config-file /etc/neutron/neutron.conf \
> --config-file /etc/neutron/plugins/ml2/ml2_conf.ini upgrade head" neutron
INFO [alembic.runtime.migration] Context impl MySQLImpl.
........
OK
## 重启 api 服务
[root@controller ~]# systemctl restart openstack-nova-api.service
## 启动 network 服务
[root@controller ~]# systemctl enable neutron-server.service \
> neutron-linuxbridge-agent.service neutron-dhcp-agent.service \
> neutron-metadata-agent.service
[root@controller ~]# systemctl start neutron-server.service \
> neutron-linuxbridge-agent.service neutron-dhcp-agent.service \
> neutron-metadata-agent.service
## 配置 Compute 节点的 network 服务
## 安装 network 软件
[root@computer1 ~]# yum install openstack-neutron openstack-neutron-linuxbridge ebtables ipset
## 编辑 /etc/neutron/neutron.conf 配置文件
注释掉 [database] 部分的所有选项,计算节点 不连接 数据库
[DEFAULT]
...
rpc_backend = rabbit
auth_strategy = keystone
verbose = True
[oslo_messaging_rabbit]
...
rabbit_host = controller
rabbit_userid = openstack
rabbit_password = openstack
[keystone_authtoken]
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = neutron
password = neutron
注释掉 [keystone_authtoken] 其他部分的选项。
[oslo_concurrency]
...
lock_path = /var/lib/neutron/tmp
## 安装 network 服务
## 配置 Linux bridge 服务
编辑 /etc/neutron/plugins/ml2/linuxbridge_agent.ini 配置文件
[linux_bridge]
physical_interface_mappings = public:em2
[vxlan]
enable_vxlan = False
[agent]
...
prevent_arp_spoofing = True
[securitygroup]
...
enable_security_group = True
firewall_driver = neutron.agent.linux.iptables_firewall.IptablesFirewallDriver
## 配置 Networking 服务
## 编辑 /etc/nova/nova.conf 配置文件
[neutron]
...
url = http://controller:9696
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
region_name = RegionOne
project_name = service
username = neutron
password = neutron
## 配置链接, 并启动服务
[root@computer1 ~]# ln -s /etc/neutron/plugins/ml2/ml2_conf.ini /etc/neutron/plugin.ini
[root@computer1 ~]# systemctl restart openstack-nova-compute.service
[root@computer1 ~]# systemctl enable neutron-linuxbridge-agent.service
[root@computer1 ~]# systemctl start neutron-linuxbridge-agent.service
## 最后 验证服务
[root@controller ~]# source admin-openrc.sh
[root@controller ~]# neutron ext-list
+-----------------------+--------------------------+
| alias | name |
+-----------------------+--------------------------+
| flavors | Neutron Service Flavors |
| security-group | security-group |
| dns-integration | DNS Integration |
| net-mtu | Network MTU |
| port-security | Port Security |
| binding | Port Binding |
| provider | Provider Network |
| agent | agent |
| quotas | Quota management support |
| subnet_allocation | Subnet Allocation |
| dhcp_agent_scheduler | DHCP Agent Scheduler |
| rbac-policies | RBAC Policies |
| external-net | Neutron external network |
| multi-provider | Multi Provider Network |
| allowed-address-pairs | Allowed Address Pairs |
| extra_dhcp_opt | Neutron Extra DHCP opts |
+-----------------------+--------------------------+
[root@controller ~]# neutron agent-list
+--------------------------------------+--------------------+------------+-------+----------------+---------------------------+
| id | agent_type | host | alive | admin_state_up | binary |
+--------------------------------------+--------------------+------------+-------+----------------+---------------------------+
| 3083cb45-f248-4de5-828e-38d9bc961814 | Linux bridge agent | controller | :-) | True | neutron-linuxbridge-agent |
| 3b2564c7-e797-4e20-a257-166ade425655 | Linux bridge agent | computer1 | :-) | True | neutron-linuxbridge-agent |
| b61a8667-416d-4ed5-b14d-87d568bab59f | Metadata agent | controller | :-) | True | neutron-metadata-agent |
| f87d6cc0-4277-4259-96ed-638294503236 | DHCP agent | controller | :-) | True | neutron-dhcp-agent |
+--------------------------------------+--------------------+------------+-------+----------------+---------------------------+
六、安装配置 dashboard
[root@controller ~]# yum install openstack-dashboard -y
## 编辑/etc/openstack-dashboard/local_settings 配置文件
[root@controller ~]# vim /etc/openstack-dashboard/local_settings
OPENSTACK_HOST = "controller"
ALLOWED_HOSTS = ['*', ]
CACHES = {
'default': {
'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',
'LOCATION': '127.0.0.1:11211',
}
}
OPENSTACK_KEYSTONE_DEFAULT_ROLE = "user"
TIME_ZONE = "Asia/Shanghai"
## 启动服务
[root@controller ~]# systemctl enable httpd.service memcached.service
[root@controller ~]# systemctl restart httpd.service memcached.service
## 验证服务
浏览器 访问 http://controller/dashboard
帐号密码为 admin
七、 安装配置 Block Storage service
控制端
## 添加数据库
[root@controller ~]# mysql -u root -p
Enter password:
MariaDB [(none)]> CREATE DATABASE cinder;
MariaDB [(none)]> GRANT ALL PRIVILEGES ON cinder.* TO 'cinder'@'localhost' \
-> IDENTIFIED BY 'cinder';
MariaDB [(none)]> GRANT ALL PRIVILEGES ON cinder.* TO 'cinder'@'%' \
-> IDENTIFIED BY 'cinder';
## 添加 openstack cinder用户
[root@controller ~]# source admin-openrc.sh
[root@controller ~]# openstack user create --domain default --password-prompt cinder
User Password:
Repeat User Password:
+-----------+----------------------------------+
| Field | Value |
+-----------+----------------------------------+
| domain_id | default |
| enabled | True |
| id | 212ced82b0a844f099a80a712786740c |
| name | cinder |
+-----------+----------------------------------+
[root@controller ~]# openstack role add --project service --user cinder admin
## 添加 cinder v2 service 服务
[root@controller ~]# openstack service create --name cinder \
> --description "OpenStack Block Storage" volume
+-------------+----------------------------------+
| Field | Value |
+-------------+----------------------------------+
| description | OpenStack Block Storage |
| enabled | True |
| id | 6af1b2a1843f42058b03368960ae6b09 |
| name | cinder |
| type | volume |
+-------------+----------------------------------+
[root@controller ~]# openstack service create --name cinderv2 \
> --description "OpenStack Block Storage" volumev2
+-------------+----------------------------------+
| Field | Value |
+-------------+----------------------------------+
| description | OpenStack Block Storage |
| enabled | True |
| id | 3e42f0d7868745579ce658d88eef4c67 |
| name | cinderv2 |
| type | volumev2 |
+-------------+----------------------------------+
## 添加 Block Storage service API
[root@controller ~]# openstack endpoint create --region RegionOne \
> volume public http://controller:8776/v1/%\(tenant_id\)s
+--------------+-----------------------------------------+
| Field | Value |
+--------------+-----------------------------------------+
| enabled | True |
| id | 14ee6bd40f08489bb314e5cc5fa39a80 |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 6af1b2a1843f42058b03368960ae6b09 |
| service_name | cinder |
| service_type | volume |
| url | http://controller:8776/v1/%(tenant_id)s |
+--------------+-----------------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> volume internal http://controller:8776/v1/%\(tenant_id\)s
+--------------+-----------------------------------------+
| Field | Value |
+--------------+-----------------------------------------+
| enabled | True |
| id | 5895f13f971a4ee390ef104ddbe7422b |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 6af1b2a1843f42058b03368960ae6b09 |
| service_name | cinder |
| service_type | volume |
| url | http://controller:8776/v1/%(tenant_id)s |
+--------------+-----------------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> volume admin http://controller:8776/v1/%\(tenant_id\)s
+--------------+-----------------------------------------+
| Field | Value |
+--------------+-----------------------------------------+
| enabled | True |
| id | e4993535c6504f95862cab4f7ab7cccb |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 6af1b2a1843f42058b03368960ae6b09 |
| service_name | cinder |
| service_type | volume |
| url | http://controller:8776/v1/%(tenant_id)s |
+--------------+-----------------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> volumev2 public http://controller:8776/v2/%\(tenant_id\)s
+--------------+-----------------------------------------+
| Field | Value |
+--------------+-----------------------------------------+
| enabled | True |
| id | 8696e7142c17488e973c653713c14379 |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 3e42f0d7868745579ce658d88eef4c67 |
| service_name | cinderv2 |
| service_type | volumev2 |
| url | http://controller:8776/v2/%(tenant_id)s |
+--------------+-----------------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> volumev2 internal http://controller:8776/v2/%\(tenant_id\)s
+--------------+-----------------------------------------+
| Field | Value |
+--------------+-----------------------------------------+
| enabled | True |
| id | da6b9e902a4d413dadd951ec751b1082 |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 3e42f0d7868745579ce658d88eef4c67 |
| service_name | cinderv2 |
| service_type | volumev2 |
| url | http://controller:8776/v2/%(tenant_id)s |
+--------------+-----------------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> volumev2 admin http://controller:8776/v2/%\(tenant_id\)s
+--------------+-----------------------------------------+
| Field | Value |
+--------------+-----------------------------------------+
| enabled | True |
| id | 0cfa27d0a5b946e8b35901d242ac165f |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| service_id | 3e42f0d7868745579ce658d88eef4c67 |
| service_name | cinderv2 |
| service_type | volumev2 |
| url | http://controller:8776/v2/%(tenant_id)s |
+--------------+-----------------------------------------+
## 安装以及配置
[root@controller ~]# yum install openstack-cinder python-cinderclient -y
## 编辑 /etc/cinder/cinder.conf 配置文件
[root@controller ~]# vim /etc/cinder/cinder.conf
[database]
...
connection = mysql://cinder:cinder@controller/cinder
[DEFAULT]
...
rpc_backend = rabbit
auth_strategy = keystone
my_ip = 10.6.17.11
verbose = True
[oslo_messaging_rabbit]
...
rabbit_host = controller
rabbit_userid = openstack
rabbit_password = openstack
注释掉 [keystone_authtoken] 其他部分的选项。添加如下内容:
[keystone_authtoken]
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = cinder
password = cinder
[oslo_concurrency]
...
lock_path = /var/lib/cinder/tmp
## 同步数据库
[root@controller ~]# su -s /bin/sh -c "cinder-manage db sync" cinder
## 编辑 /etc/nova/nova.conf 配置文件
[cinder]
os_region_name = RegionOne
## 启动服务
[root@controller ~]# systemctl restart openstack-nova-api.service
[root@controller ~]# systemctl enable openstack-cinder-api.service openstack-cinder-scheduler.service
[root@controller ~]# systemctl start openstack-cinder-api.service openstack-cinder-scheduler.service
## Storage 节点操作
## 安装软件
[root@block1 ~]# yum install lvm2
## 启动 LVM2 服务
[root@block1 ~]# systemctl enable lvm2-lvmetad.service
[root@block1 ~]# systemctl start lvm2-lvmetad.service
## 创建分区
[root@block1 ~]# fdisk -l
磁盘 /dev/sda:599.6 GB, 599550590976 字节,1170997248 个扇区
Units = 扇区 of 1 * 512 = 512 bytes
扇区大小(逻辑/物理):512 字节 / 512 字节
I/O 大小(最小/最佳):512 字节 / 512 字节
磁盘标签类型:dos
磁盘标识符:0x000e0185
设备 Boot Start End Blocks Id System
/dev/sda1 * 2048 2099199 1048576 83 Linux
/dev/sda2 2099200 211814399 104857600 83 Linux
/dev/sda3 211814400 211830783 8192 82 Linux swap / Solaris
## 使用 parted 软件,查看分区
[root@block1 ~]# parted /dev/sda
GNU Parted 3.1
使用 /dev/sda
Welcome to GNU Parted! Type 'help' to view a list of commands.
(parted) p
Model: Dell Virtual Disk (scsi)
Disk /dev/sda: 600GB
Sector size (logical/physical): 512B/512B
Partition Table: msdos
Disk Flags:
Number Start End Size Type File system 标志
1 1049kB 1075MB 1074MB primary xfs 启动
2 1075MB 108GB 107GB primary xfs
3 108GB 108GB 8389kB primary linux-swap(v1)
## 使用 parted 软件, 创建分区
(parted) mkpart primary 108GB 208GB
(parted) p
Model: Dell Virtual Disk (scsi)
Disk /dev/sda: 600GB
Sector size (logical/physical): 512B/512B
Partition Table: msdos
Disk Flags:
Number Start End Size Type File system 标志
1 1049kB 1075MB 1074MB primary xfs 启动
2 1075MB 108GB 107GB primary xfs
3 108GB 108GB 8389kB primary linux-swap(v1)
4 108GB 208GB 99.5GB primary
[root@block1 ~]# fdisk -l
磁盘 /dev/sda:599.6 GB, 599550590976 字节,1170997248 个扇区
Units = 扇区 of 1 * 512 = 512 bytes
扇区大小(逻辑/物理):512 字节 / 512 字节
I/O 大小(最小/最佳):512 字节 / 512 字节
磁盘标签类型:dos
磁盘标识符:0x000e0185
设备 Boot Start End Blocks Id System
/dev/sda1 * 2048 2099199 1048576 83 Linux
/dev/sda2 2099200 211814399 104857600 83 Linux
/dev/sda3 211814400 211830783 8192 82 Linux swap / Solaris
/dev/sda4 211830784 406249471 97209344 83 Linux
[root@block1 ~]# pvcreate /dev/sda4
Physical volume "/dev/sda4" successfully created
[root@block1 ~]# vgcreate cinder-volumes /dev/sda4
Volume group "cinder-volumes" successfully created
## 编辑 /etc/lvm/lvm.conf 配置文件,添加分区
devices {
...
filter = [ "a/sda4/", "r/.*/"]
## 安装 openstack 软件
[root@block1 ~]# yum install openstack-cinder targetcli python-oslo-policy -y
## 编辑 /etc/cinder/cinder.conf 配置文件
[database]
...
connection = mysql://cinder:cinder@controller/cinder
[DEFAULT]
...
rpc_backend = rabbit
auth_strategy = keystone
my_ip = 10.6.17.13
enabled_backends = lvm
glance_host = controller
verbose = True
volume_driver = cinder.volume.drivers.lvm.LVMVolumeDriver
volume_group = cinder-volumes
iscsi_protocol = iscsi
iscsi_helper = lioadm
[oslo_messaging_rabbit]
...
rabbit_host = controller
rabbit_userid = openstack
rabbit_password = openstack
# 注释掉[keystone_authtoken]下其他所有的选项 只使用下面
[keystone_authtoken]
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = cinder
password = cinder
[oslo_concurrency]
...
lock_path = /var/lib/cinder/tmp
## 启动 服务
[root@block1 ~]# systemctl enable openstack-cinder-volume.service target.service
[root@block1 ~]# systemctl start openstack-cinder-volume.service target.service
## 验证服务
[root@controller ~]# source admin-openrc.sh
[root@controller ~]# cinder service-list
+------------------+------------+------+---------+-------+----------------------------+-----------------+
| Binary | Host | Zone | Status | State | Updated_at | Disabled Reason |
+------------------+------------+------+---------+-------+----------------------------+-----------------+
| cinder-scheduler | controller | nova | enabled | up | 2015-12-04T07:35:57.000000 | - |
| cinder-volume | block1@lvm | nova | enabled | up | 2015-12-04T07:36:03.000000 | - |
+------------------+------------+------+---------+-------+----------------------------+-----------------+
八、 安装配置 Object Storage 服务
## 创建 openstack 帐号
[root@controller ~]# openstack user create --domain default --password-prompt swift
User Password:
Repeat User Password:
+-----------+----------------------------------+
| Field | Value |
+-----------+----------------------------------+
| domain_id | default |
| enabled | True |
| id | 7a2915af03464268ab80ef732aa5ba93 |
| name | swift |
+-----------+----------------------------------+
[root@controller ~]# openstack role add --project service --user swift admin
## 创建 swift service
[root@controller ~]# openstack service create --name swift \
> --description "OpenStack Object Storage" object-store
+-------------+----------------------------------+
| Field | Value |
+-------------+----------------------------------+
| description | OpenStack Object Storage |
| enabled | True |
| id | f7b97bc04bd24d67b12694bb3004843b |
| name | swift |
| type | object-store |
+-------------+----------------------------------+
## 创建 Object Storage service API
[root@controller ~]# openstack endpoint create --region RegionOne \
> object-store public http://controller:8080/v1/AUTH_%\(tenant_id\)s
+--------------+----------------------------------------------+
| Field | Value |
+--------------+----------------------------------------------+
| enabled | True |
| id | f8d7a017753947d981e0f42aa5fe4bde |
| interface | public |
| region | RegionOne |
| region_id | RegionOne |
| service_id | f7b97bc04bd24d67b12694bb3004843b |
| service_name | swift |
| service_type | object-store |
| url | http://controller:8080/v1/AUTH_%(tenant_id)s |
+--------------+----------------------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> object-store internal http://controller:8080/v1/AUTH_%\(tenant_id\)s
+--------------+----------------------------------------------+
| Field | Value |
+--------------+----------------------------------------------+
| enabled | True |
| id | 6d125e48860b4f948300b412f744b23f |
| interface | internal |
| region | RegionOne |
| region_id | RegionOne |
| service_id | f7b97bc04bd24d67b12694bb3004843b |
| service_name | swift |
| service_type | object-store |
| url | http://controller:8080/v1/AUTH_%(tenant_id)s |
+--------------+----------------------------------------------+
[root@controller ~]# openstack endpoint create --region RegionOne \
> object-store admin http://controller:8080/v1
+--------------+----------------------------------+
| Field | Value |
+--------------+----------------------------------+
| enabled | True |
| id | cdbea58743c946cca41c98682ea98793 |
| interface | admin |
| region | RegionOne |
| region_id | RegionOne |
| service_id | f7b97bc04bd24d67b12694bb3004843b |
| service_name | swift |
| service_type | object-store |
| url | http://controller:8080/v1 |
+--------------+----------------------------------+
## 安装配置软件
[root@controller ~]# yum install openstack-swift-proxy python-swiftclient \
> python-keystoneclient python-keystonemiddleware \
> memcached
## 下载swift配置文件
[root@controller ~]# curl -o /etc/swift/proxy-server.conf https://git.openstack.org/cgit/openstack/swift/plain/etc/proxy-server.conf-sample?h=stable/liberty
##编辑 /etc/swift/proxy-server.conf 配置文件
[DEFAULT]
...
bind_port = 8080
user = swift
swift_dir = /etc/swift
[pipeline:main]
pipeline = catch_errors gatekeeper healthcheck proxy-logging cache container_sync bulk ratelimit authtoken keystoneauth container-quotas account-quotas slo dlo versioned_writes proxy-logging proxy-server
[app:proxy-server]
use = egg:swift#proxy
...
account_autocreate = true
[filter:keystoneauth]
use = egg:swift#keystoneauth
...
operator_roles = admin,user
#注释掉[filter:authtoken] 其他所有的选项只使用下面
[filter:authtoken]
paste.filter_factory = keystonemiddleware.auth_token:filter_factory
...
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = swift
password = swift
delay_auth_decision = true
[filter:cache]
use = egg:swift#memcache
...
memcache_servers = 127.0.0.1:11211
## Object storage 节点配置
## 安装配置软件
[root@object1 ~]# yum install xfsprogs rsync -y
[root@object1 ~]# fdisk -l
磁盘 /dev/sda:599.6 GB, 599550590976 字节,1170997248 个扇区
Units = 扇区 of 1 * 512 = 512 bytes
扇区大小(逻辑/物理):512 字节 / 512 字节
I/O 大小(最小/最佳):512 字节 / 512 字节
磁盘标签类型:dos
磁盘标识符:0x000bbc86
设备 Boot Start End Blocks Id System
/dev/sda1 * 2048 204802047 102400000 83 Linux
/dev/sda2 204802048 221186047 8192000 82 Linux swap / Solaris
/dev/sda3 221186048 611327999 195070976 83 Linux
/dev/sda4 611328000 1170997247 279834624 83 Linux
[root@object1 ~]# mkfs.xfs -f /dev/sda3
[root@object1 ~]# mkfs.xfs -f /dev/sda4
[root@object1 ~]# mkdir -p /opt/node/sda3
[root@object1 ~]# mkdir -p /opt/node/sda4
## 编辑 /etc/fstab 让系统启动自动加载
/dev/sda3 /opt/node/sda3 xfs noatime,nodiratime,nobarrier,logbufs=8 0 2
/dev/sda4 /opt/node/sda4 xfs noatime,nodiratime,nobarrier,logbufs=8 0 2
## 挂载硬盘
[root@object1 ~]# mount -a
[root@object1 ~]# df -h
文件系统 容量 已用 可用 已用% 挂载点
/dev/sda1 98G 1.2G 97G 2% /
devtmpfs 7.8G 0 7.8G 0% /dev
tmpfs 7.8G 0 7.8G 0% /dev/shm
tmpfs 7.8G 8.4M 7.8G 1% /run
tmpfs 7.8G 0 7.8G 0% /sys/fs/cgroup
/dev/sda3 186G 33M 186G 1% /opt/node/sda3
/dev/sda4 267G 33M 267G 1% /opt/node/sda4
## 编辑 /etc/rsyncd.conf 配置文件
[root@object1 ~]# vim /etc/rsyncd.conf
--------------------------------------------------------
uid = swift
gid = swift
log file = /var/log/rsyncd.log
pid file = /var/run/rsyncd.pid
address = 10.6.17.14
[account]
max connections = 2
path = /opt/node/
read only = false
lock file = /var/lock/account.lock
[container]
max connections = 2
path = /opt/node/
read only = false
lock file = /var/lock/container.lock
[object]
max connections = 2
path = /opt/node/
read only = false
lock file = /var/lock/object.lock
--------------------------------------------------------
## 启动服务
[root@object1 ~]# systemctl enable rsyncd.service
[root@object1 ~]# systemctl start rsyncd.service
## 安装配置软件
[root@object1 ~]# yum install openstack-swift-account openstack-swift-container \
> openstack-swift-object
## 下载配置文件
[root@object1 ~]# curl -o /etc/swift/account-server.conf https://git.openstack.org/cgit/openstack/swift/plain/etc/account-server.conf-sample?h=stable/liberty
[root@object1 ~]# curl -o /etc/swift/container-server.conf https://git.openstack.org/cgit/openstack/swift/plain/etc/container-server.conf-sample?h=stable/liberty
[root@object1 ~]# curl -o /etc/swift/object-server.conf https://git.openstack.org/cgit/openstack/swift/plain/etc/object-server.conf-sample?h=stable/liberty
## 编辑/etc/swift/account-server.conf 配置文件
[DEFAULT]
...
bind_ip = 10.6.17.14
bind_port = 6002
user = swift
swift_dir = /etc/swift
devices = /opt/node
mount_check = true
[pipeline:main]
pipeline = healthcheck recon account-server
[filter:recon]
use = egg:swift#recon
...
recon_cache_path = /var/cache/swift
## 编辑/etc/swift/container-server.conf 配置文件
[DEFAULT]
...
bind_ip = 10.6.17.14
bind_port = 6001
user = swift
swift_dir = /etc/swift
devices = /opt/node
mount_check = true
[pipeline:main]
pipeline = healthcheck recon container-server
[filter:recon]
use = egg:swift#recon
...
recon_cache_path = /var/cache/swift
## 编辑/etc/swift/object-server.conf 配置文件
[DEFAULT]
...
bind_ip = 10.6.17.14
bind_port = 6000
user = swift
swift_dir = /etc/swift
devices = /opt/node
mount_check = true
[pipeline:main]
pipeline = healthcheck recon object-server
[filter:recon]
use = egg:swift#recon
...
recon_cache_path = /var/cache/swift
recon_lock_path = /var/lock
## 授权目录
[root@object1 ~]# chown -R swift:swift /opt/node
## 创建 swift 缓存 目录 cache/swift
[root@object1 ~]# mkdir -p /var/cache/swift
[root@object1 ~]# chown -R swift:swift /var/cache/swift
## 在controller 上创建 ring 帐号
## 创建 account.builder 文件
[root@controller ~]# cd /etc/swift/
[root@controller swift]# swift-ring-builder account.builder create 10 2 1
## 添加 object 节点
[root@controller swift]# swift-ring-builder account.builder add \
> --region 1 --zone 1 --ip 10.6.17.14 --port 6002 --device sda3 --weight 100
输出:
Device d0r1z1-10.6.17.14:6002R10.6.17.14:6002/sda3_"" with 100.0 weight got id 0
[root@controller swift]# swift-ring-builder account.builder add \
> --region 1 --zone 2 --ip 10.6.17.14 --port 6002 --device sda4 --weight 100
输出:
Device d1r1z2-10.6.17.14:6002R10.6.17.14:6002/sda4_"" with 100.0 weight got id 1
## 验证ring
[root@controller swift]# swift-ring-builder account.builder
account.builder, build version 2
1024 partitions, 3.000000 replicas, 1 regions, 2 zones, 2 devices, 100.00 balance, 0.00 dispersion
The minimum number of hours before a partition can be reassigned is 1
The overload factor is 0.00% (0.000000)
Devices: id region zone ip address port replication ip replication port name weight partitions balance meta
0 1 1 10.6.17.14 6002 10.6.17.14 6002 sda3 100.00 0 -100.00
1 1 2 10.6.17.14 6002 10.6.17.14 6002 sda4 100.00 0 -100.00
# 平衡 ring
[root@controller swift]# swift-ring-builder account.builder rebalance
.......
Reassigned 1024 (100.00%) partitions. Balance is now 0.00. Dispersion is now 0.00
## 创建容器 ring
[root@controller ~]# cd /etc/swift/
[root@controller swift]# swift-ring-builder container.builder create 10 2 1
## 创建存储节点ring
[root@controller swift]# swift-ring-builder container.builder add \
> --region 1 --zone 1 --ip 10.6.17.14 --port 6001 --device sda3 --weight 100
输出:
Device d0r1z1-10.6.17.14:6001R10.6.17.14:6001/sda3_"" with 100.0 weight got id 0
[root@controller swift]# swift-ring-builder container.builder add \
> --region 1 --zone 2 --ip 10.6.17.14 --port 6001 --device sda4 --weight 100
输出:
Device d1r1z2-10.6.17.14:6001R10.6.17.14:6001/sda4_"" with 100.0 weight got id 1
## 验证容器 ring
[root@controller swift]# swift-ring-builder container.builder
container.builder, build version 2
1024 partitions, 3.000000 replicas, 1 regions, 2 zones, 2 devices, 100.00 balance, 0.00 dispersion
The minimum number of hours before a partition can be reassigned is 1
The overload factor is 0.00% (0.000000)
Devices: id region zone ip address port replication ip replication port name weight partitions balance meta
0 1 1 10.6.17.14 6001 10.6.17.14 6001 sda3 100.00 0 -100.00
1 1 2 10.6.17.14 6001 10.6.17.14 6001 sda4 100.00 0 -100.00
# 平衡 容器 ring
[root@controller swift]# swift-ring-builder container.builder rebalance
....
Reassigned 1024 (100.00%) partitions. Balance is now 0.00. Dispersion is now 0.00
## 创建 object ring
[root@controller ~]# cd /etc/swift/
[root@controller swift]# swift-ring-builder object.builder create 10 2 1
[root@controller swift]# swift-ring-builder object.builder add \
> --region 1 --zone 1 --ip 10.6.17.14 --port 6001 --device sda3 --weight 100
输出:
Device d0r1z1-10.6.17.14:6001R10.6.17.14:6001/sda3_"" with 100.0 weight got id 0
[root@controller swift]# swift-ring-builder object.builder add \
> --region 1 --zone 2 --ip 10.6.17.14 --port 6001 --device sda4 --weight 100
输出:
Device d1r1z2-10.6.17.14:6001R10.6.17.14:6001/sda4_"" with 100.0 weight got id 1
## 验证 object ring
[root@controller swift]# swift-ring-builder object.builder
object.builder, build version 2
1024 partitions, 3.000000 replicas, 1 regions, 2 zones, 2 devices, 100.00 balance, 0.00 dispersion
The minimum number of hours before a partition can be reassigned is 1
The overload factor is 0.00% (0.000000)
Devices: id region zone ip address port replication ip replication port name weight partitions balance meta
0 1 1 10.6.17.14 6001 10.6.17.14 6001 sda3 100.00 0 -100.00
1 1 2 10.6.17.14 6001 10.6.17.14 6001 sda4 100.00 0 -100.00
## 平衡 object ring
[root@controller swift]# swift-ring-builder object.builder rebalance
...
Reassigned 1024 (100.00%) partitions. Balance is now 0.00. Dispersion is now 0.00
## 查看生成的文件
[root@controller swift]# ls
account.builder backups container-reconciler.conf container-server object.builder object.ring.gz proxy-server.conf
account.ring.gz container.builder container.ring.gz container-server.conf object-expirer.conf proxy-server swift.conf
## 复制 account.ring.gz container.ring.gz object.ring.gz 到 其他 运行 代理服务 的节点/etc/swift 目录中 ,然后启动 openstack-swift-proxy.service 服务
## Object Storage 下载 swift.conf 配置文件
[root@object1 swift]# curl -o /etc/swift/swift.conf \
> https://git.openstack.org/cgit/openstack/swift/plain/etc/swift.conf-sample?h=stable/liberty
## 编辑 /etc/swift/swift.conf 配置文件
[root@object1 swift]# vim /etc/swift/swift.conf
[swift-hash]
...
swift_hash_path_suffix = changeme
swift_hash_path_prefix = changeme
[storage-policy:0]
...
name = Policy-0
default = yes
## 授权 目录
[root@object1 swift]# chown -R root:swift /etc/swift
## 在controller 以及 其他代理节点中 启动 代理服务
[root@controller ~]# systemctl enable openstack-swift-proxy.service memcached.service
[root@controller ~]# systemctl start openstack-swift-proxy.service memcached.service
## 在 Object Storage 节点中 启动服务
[root@object1 ~]# systemctl enable openstack-swift-account.service openstack-swift-account-auditor.service \
> openstack-swift-account-reaper.service openstack-swift-account-replicator.service
[root@object1 ~]# systemctl start openstack-swift-account.service openstack-swift-account-auditor.service \
> openstack-swift-account-reaper.service openstack-swift-account-replicator.service
[root@object1 ~]# systemctl enable openstack-swift-container.service \
> openstack-swift-container-auditor.service openstack-swift-container-replicator.service \
> openstack-swift-container-updater.service
[root@object1 ~]# systemctl start openstack-swift-container.service \
> openstack-swift-container-auditor.service openstack-swift-container-replicator.service \
> openstack-swift-container-updater.service
[root@object1 ~]# systemctl enable openstack-swift-object.service openstack-swift-object-auditor.service \
> openstack-swift-object-replicator.service openstack-swift-object-updater.service
[root@object1 ~]# systemctl start openstack-swift-object.service openstack-swift-object-auditor.service \
> openstack-swift-object-replicator.service openstack-swift-object-updater.service
## 验证服务
[root@controller ~]# echo "export OS_AUTH_VERSION=3" \
> | tee -a admin-openrc.sh
[root@controller ~]# source admin-openrc.sh
## 查看 swift 状态
[root@controller ~]# swift stat
Account: AUTH_0d1a69b17e444cb3b4884529b6bdb372
Containers: 0
Objects: 0
Bytes: 0
X-Put-Timestamp: 1449222121.42343
X-Timestamp: 1449222121.42343
X-Trans-Id: tx20d0064672a74cd88723f-0056615fe9
Content-Type: text/plain; charset=utf-8
## 测试 上传 文件
[root@controller ~]# touch FILE
[root@controller ~]# swift upload container1 FILE