Skip to content
Snippets Groups Projects

Current deployement

mds

  • ds-507
  • ds-304

mon

  • ds-507
  • qn-cnfslhc
  • ds-303
  • ds-304
  • cs-001

Installation

ssh keys exchange

Choose an admin node to use for the installation process. Distribute the ssh key to all the hosts in the cluster.

Install the deploy script

Add the ceph repository

yum install ceph-deploy
yum install ntp ntpdate ntp-doc

Purge cluster

ceph-deploy purge qn-cnfslhc ds-001 ds-002 ds-303 ds-304 ds-507
ceph-deploy purgedata qn-cnfslhc ds-001 ds-002 ds-303 ds-304 ds-507
ceph-deploy forgetkeys

Create first monitor nodes:

ceph-deploy new qn-cnfslhc

This will create the following files:

ceph.conf
ceph.mon.keyring

Add the public network to the configuration file:

public_network = 131.154.128.0/22

Install the nodes:

ceph-deploy install node1 node2 node3 --release nautilus

Deploy the initial monitoring node:

ceph-deploy mon create-initial

and the admin keys to the nodes of your cluster: copy the configuration file and admin keys gathered above to all your Ceph nodes to enable you use the ceph CLI without having to specify the monitor address and ceph.client.admin.keyring each time you execute a command.

ceph-deploy admin qn-cnfslhc ds-001 ...

Then deploy the manager node

ceph-deploy -v mgr create qn-cnfslhc

If you have a dirty installation you may receive errors like:

[qn-cnfslhc][ERROR ] [errno 1] error connecting to the cluster
[qn-cnfslhc][ERROR ] exit code from command was: 1
[ceph_deploy.mgr][ERROR ] could not create mgr
[ceph_deploy][ERROR ] GenericError: Failed to create 1 MGRs

This means that you must remove the old keys from /var/lib/ceph

rm -rf /var/lib/ceph/bootstrap-mgr/

Check that the installation is fine:

sudo ceph -s

Enable dashboard

The dashboard runs on a host with ceph-mgr active:

yum install ceph-mgr-dashboard # for nautilus
ceph mgr module enable dashboard
ceph config set mgr mgr/dashboard/qn-cnfslhc/server_addr 131.154.130.69
ceph config set mgr mgr/dashboard/qn-cnfslhc/server_port 5000
ceph dashboard set-login-credentials admin <password>
ceph config set mgr mgr/dashboard/ssl false

Monitors

Add monitor node

ceph-deploy -v mon create <id>
ceph-deploy -v  admin <id>

Remove monitor node

ceph-deploy -v mon destroy <id>

Metadata

Add metadata server

ceph-deploy mds create ds-507

OSD

Disk preparation

ceph-deploy disk zap ds-507 /dev/nvme0n1

Prepare data disks

lsblk
dmsetup remove
gdisk /dev/sdbi (x z y y)
ceph--c666c0d8--e77d--4d3e--931e--c7041572f747-osd--block--3414fd14--e0bf--4adf--bf5d--3c0412821d11

ceph-deploy disk zap cs-001 /dev/sdap

prepare journal partitions on ssd

vgcreate ceph-db-0 /dev/sdbj1

for i in $(seq 40 59); do echo "lvcreate -L 23GB -n db-$i ceph-db-0";done
lvcreate -L 23GB -n db-40 ceph-db-0
lvcreate -L 23GB -n db-41 ceph-db-0
lvcreate -L 23GB -n db-42 ceph-db-0
lvcreate -L 23GB -n db-43 ceph-db-0
lvcreate -L 23GB -n db-44 ceph-db-0
lvcreate -L 23GB -n db-45 ceph-db-0
lvcreate -L 23GB -n db-46 ceph-db-0
lvcreate -L 23GB -n db-47 ceph-db-0
lvcreate -L 23GB -n db-48 ceph-db-0
lvcreate -L 23GB -n db-49 ceph-db-0
lvcreate -L 23GB -n db-50 ceph-db-0
lvcreate -L 23GB -n db-51 ceph-db-0
lvcreate -L 23GB -n db-52 ceph-db-0
lvcreate -L 23GB -n db-53 ceph-db-0
lvcreate -L 23GB -n db-54 ceph-db-0
lvcreate -L 23GB -n db-55 ceph-db-0
lvcreate -L 23GB -n db-56 ceph-db-0
lvcreate -L 23GB -n db-57 ceph-db-0
lvcreate -L 23GB -n db-58 ceph-db-0
lvcreate -L 23GB -n db-59 ceph-db-0

for i in $(seq 40 59); do lvcreate -L 13GB -n wal-$i ceph-db-0;done

for i in $(seq 40 59); do lvresize -L 10G /dev/ceph-db-0/db-$i -y;done

Client installation

Rados gateway

[Official docs][https://docs.ceph.com/docs/master/install/install-ceph-gateway/]

ceph-deploy install --rgw ds-517
ceph-deploy --overwrite-conf rgw create ds-517

Il gateway è già contattabile.
http://client-node:7480