Thanks for responding :-)

I am not a Ceph expert either.

Steps to reproduce:
Short version: On a running ceph-cluster with at least one blacklisted osd: 
install ceph-iscsi.

Long version:
Installing Ceph is complicated, so here is a way to do it with ansible. I know 
this is a lot, but believe me: this is the EASY way.

Normally you would install a ceph-cluster with at least 3 servers, but
below is described how to do it on one machine.

Fresh install of Ubuntu 20.04.1 LTS server. Can be in a VM, but make sure you 
have 3 exstra disks attached and at least 4GB RAM. Below I assume install is on 
sda and sdb, sdc and sdd are attached blank hds ( I have used 10GB for each).
Make sure install is up-to-date:
thw@ff-ceph-4:~$ sudo apt update
thw@ff-ceph-4:~$ sudo apt dist-upgrade
thw@ff-ceph-4:~$ sudo reboot

Here I have the default admin user thw.
Assuming hostname ff-ceph-4
make sure ff-ceph-4 can resolve to the external ip-address of the local 
machine, eg. by adding it to /etc/hosts

Make sure user thw can sudo any command without password by sudo visudo and 
adding the line:
thw    ALL=(ALL) NOPASSWD:ALL

thw@ff-ceph-4:~$ sudo apt install ansible git

Create user ansible with password ansible:
thw@ff-ceph-4:~$ sudo adduser ansible

thw@ff-ceph-4:~$ su - ansible

get Ceph ansible playbooks:
ansible@ff-ceph-4:~$ git clone https://github.com/ceph/ceph-ansible.git
ansible@ff-ceph-4:~$ cd ceph-ansible
ansible@ff-ceph-4:~$ git checkout stable-5.0

Copy the attached all.yml to group_vars dir in ceph-ansible. You can diff to 
all.yml.sample to see what I have changed. I advise you to do this. Make sure 
the monitor_interface: line lists the name of your network interface, and that 
public_network: is the network that interface is on.
Copy the attached inventory to current-dir (/home/ansible/ceph-ansible)
NOTE: I could only attach ONE file, so this will be attached in new comment 
below.

make sure user ansible kan login to thw account with ssh-key:
ansible@ff-ceph-4:~$ ssh-keygen
ansible@ff-ceph-4:~$ ssh-copy-id thw@ff-ceph-4
ansible@ff-ceph-4:~$ cp site.yml.sample site.yml
ansible@ff-ceph-4:~$ ansible-playbook -i inventory site.yml

This will take a few minutes to run.
If something goes wrong see if you can fix it, and re-run the ansible-playbook 
command.

At the end you should hopefully have a running ceph-"cluster".
Go back to thw user (or add ansible user to sudo-list) and run
thw@ff-ceph-4:~$ sudo ceph -s

Line 3 should read:
    health: HEALTH_OK
    
To reproduce the bug, you should have some blacklist entries. I have them on a 
new install - I don't know why. Check with:
thw@ff-ceph-4:~$ sudo ceph osd blacklist ls

If there are entries listed, fine.
If not, create an entry with:
thw@ff-ceph-4:~$ sudo ceph osd blacklist add  <ip-address-of-your host>

Now:
thw@ff-ceph-4:~$ sudo ceph osd pool create rbd
thw@ff-ceph-4:~$ sudo rbd pool init rbd
thw@ff-ceph-4:~$ sudo apt install ceph-iscsi

You should now see the exceptions in journalctl


** Attachment added: "variables for ansible playbook"
   
https://bugs.launchpad.net/ubuntu/+source/ceph-iscsi/+bug/1909399/+attachment/5449256/+files/all.yml

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1909399

Title:
  Exception during removal of OSD blacklist entries

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ceph-iscsi/+bug/1909399/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

Reply via email to