Newer
Older
### NOTE: This Guide based on the package ceph-deploy is deprecated
## mds
* ds-507
* ds-304
## mon
* ds-507
* qn-cnfslhc
* ds-303
* ds-304
* cs-001
### Installation
## ssh keys exchange
Choose an admin node to use for the installation process.
Distribute the ssh key to all the hosts in the cluster.
## Install the deploy script
Add the ceph repository
```
yum install ceph-deploy
yum install ntp ntpdate ntp-doc
```
## Purge cluster
```
ceph-deploy purge qn-cnfslhc ds-001 ds-002 ds-303 ds-304 ds-507
ceph-deploy purgedata qn-cnfslhc ds-001 ds-002 ds-303 ds-304 ds-507
ceph-deploy forgetkeys
```
Create first monitor nodes:
```
ceph-deploy new qn-cnfslhc
```
This will create the following files:
```
ceph.conf
ceph.mon.keyring
```
Add the public network to the configuration file:
```
public_network = 131.154.128.0/22
```
ceph-deploy install node1 node2 node3 --release nautilus
and the admin keys to the nodes of your cluster:
copy the configuration file and admin keys gathered
above to all your Ceph nodes to enable you use the
ceph CLI without having to specify the monitor address
and ceph.client.admin.keyring each time you execute a command.
```
Then deploy the manager node
```
ceph-deploy -v mgr create qn-cnfslhc
```
If you have a dirty installation you may receive errors like:
```
[qn-cnfslhc][ERROR ] [errno 1] error connecting to the cluster
[qn-cnfslhc][ERROR ] exit code from command was: 1
[ceph_deploy.mgr][ERROR ] could not create mgr
[ceph_deploy][ERROR ] GenericError: Failed to create 1 MGRs
```
This means that you must remove the old keys from `/var/lib/ceph`
```
rm -rf /var/lib/ceph/bootstrap-mgr/
```
Check that the installation is fine:
```
sudo ceph -s
```
```
yum install ceph-mgr-dashboard # for nautilus
ceph mgr module enable dashboard
ceph config set mgr mgr/dashboard/qn-cnfslhc/server_addr 131.154.130.69
ceph config set mgr mgr/dashboard/qn-cnfslhc/server_port 5000
#ceph dashboard set-login-credentials admin <password>
ceph dashboard ac-user-set-password admin admin
## Remove monitor node
```
ceph-deploy -v mon destroy <id>
### Metadata
## Add metadata server
```
ceph-deploy mds create ds-507
```
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
### OSD
## Disk preparation
```
ceph-deploy disk zap ds-507 /dev/nvme0n1
```
Prepare data disks
```
lsblk
dmsetup remove
gdisk /dev/sdbi (x z y y)
ceph--c666c0d8--e77d--4d3e--931e--c7041572f747-osd--block--3414fd14--e0bf--4adf--bf5d--3c0412821d11
ceph-deploy disk zap cs-001 /dev/sdap
```
prepare journal partitions on ssd
```
vgcreate ceph-db-0 /dev/sdbj1
for i in $(seq 40 59); do echo "lvcreate -L 23GB -n db-$i ceph-db-0";done
lvcreate -L 23GB -n db-40 ceph-db-0
lvcreate -L 23GB -n db-41 ceph-db-0
lvcreate -L 23GB -n db-42 ceph-db-0
lvcreate -L 23GB -n db-43 ceph-db-0
lvcreate -L 23GB -n db-44 ceph-db-0
lvcreate -L 23GB -n db-45 ceph-db-0
lvcreate -L 23GB -n db-46 ceph-db-0
lvcreate -L 23GB -n db-47 ceph-db-0
lvcreate -L 23GB -n db-48 ceph-db-0
lvcreate -L 23GB -n db-49 ceph-db-0
lvcreate -L 23GB -n db-50 ceph-db-0
lvcreate -L 23GB -n db-51 ceph-db-0
lvcreate -L 23GB -n db-52 ceph-db-0
lvcreate -L 23GB -n db-53 ceph-db-0
lvcreate -L 23GB -n db-54 ceph-db-0
lvcreate -L 23GB -n db-55 ceph-db-0
lvcreate -L 23GB -n db-56 ceph-db-0
lvcreate -L 23GB -n db-57 ceph-db-0
lvcreate -L 23GB -n db-58 ceph-db-0
lvcreate -L 23GB -n db-59 ceph-db-0
for i in $(seq 40 59); do lvcreate -L 13GB -n wal-$i ceph-db-0;done
for i in $(seq 40 59); do lvresize -L 10G /dev/ceph-db-0/db-$i -y;done
[Official docs][https://docs.ceph.com/docs/master/install/install-ceph-gateway/]
ceph-deploy --overwrite-conf rgw create ds-517
Il gateway è già contattabile.
http://client-node:7480