Bigstack

Bigstack

  • Docs

›Maintenance

Introduction

  • Overview

Installation

  • Control Converged
  • Control
  • Compute node (HCI)
  • Storage node (SDS)
  • NIC configuration

    • Network bonding
    • VLAN
    • VLAN on bonding

    High Availability

    • Network offload
    • Control node
    • Compute node
    • Storage node

Quick Start

  • Get Started

Compute

  • flavors
  • images
  • images(CLI)
  • instances
  • Instance Snapshot

Volumes

  • Volume
  • Snapshots
  • Backup
  • Transfer

Network

  • Internet
  • Private Network(LAN)
  • Router
  • Security Group
  • Floating IP
  • Load Balancers

    • Create a Load Balancer
    • Delete a Load Balancer

DNS

  • Create Public Record
  • Create Private Record

Object Store

  • Object Storage
  • Object Storage (S3 API)

File Store

  • Tenant Share

Identity

  • Create a project
  • Modify Quota
  • Create a User
  • Create a group

Storage

  • Add a disk
  • Remove a disk
  • Prepare disk
  • Remove osd
  • Cache
  • Restful API

Disaster Recovery

  • Start scheduler
  • Full Backup
  • Incremental Backup
  • Cleanup
  • Schedule job
  • Increase quota

Kubernetes

    Instance (VM)

    • Create Template
    • Deploy Cluster
    • Storage Class
    • Workload with L4 LoadBalancer

    Compute Node(HCI)

    • Deploy Cluster
    • Storage Class

Monitoring

  • Host
  • Storage
  • Network
  • Instance
  • Top Hosts
  • Top Instances

Logs

  • Alert
  • Notification

    • Email
    • Slack
  • ELK Stack

Maintenance

  • Startup
  • Shutdown
  • Add/Remove a Harddisk
  • Repair Services
  • Fixpack
  • Update CubeOS
  • Support file
  • Export Instance
  • Create CubeOS snapshot
  • Restore CubeOS with snapshot
  • Remove Compute node
  • Remove Storage node

Download

  • Cloud Image

Operating System

  • CentOS 6 - root disk
  • MS Windows

    • Retrieve password
    • Extend partition

Storage node Removal Guide

Remove Storage node from cluster

Make sure your storage data replication either 2 or 3.

Connect to controller

$ ssh [email protected]
Warning: Permanently added '192.168.1.x' (ECDSA) to the list of known hosts.
Password:

Check the storage status

control01> storage
control01:storage> status
  cluster:
    id:     c6e64c49-09cf-463b-9d1c-b6645b4b3b85
    health: HEALTH_OK

  services:
    mon: 3 daemons, quorum control01,control02,control03
    mgr: control01(active), standbys: control02, control03
    mds: cephfs-1/1/1 up  {0=control01=up:active}, 2 up:standby
    osd: 24 osds: 24 up, 24 in
    rgw: 3 daemons active

  data:
    pools:   23 pools, 1837 pgs
    objects: 10.50k objects, 12.7GiB
    usage:   31.3GiB used, 3.74TiB / 3.77TiB avail
    pgs:     1837 active+clean

  io:
    client:   15.5KiB/s rd, 0B/s wr, 15op/s rd, 10op/s wr

+----+-----------+-------+-------+--------+---------+--------+---------+-----------+
| id |    host   |  used | avail | wr ops | wr data | rd ops | rd data |   state   |
+----+-----------+-------+-------+--------+---------+--------+---------+-----------+
| 0  | control01 | 2063M |  117G |    0   |     0   |    2   |    61   | exists,up |
| 1  | control01 | 2020M |  117G |    0   |     0   |    1   |    36   | exists,up |
| 2  | control01 | 1089M |  135G |    0   |     0   |    0   |     0   | exists,up |
| 3  | control01 | 1081M |  135G |    0   |     0   |    0   |     0   | exists,up |
| 4  | control02 | 1656M |  117G |    0   |     0   |    0   |     0   | exists,up |
| 5  | control02 | 2073M |  116G |    0   |     0   |    0   |     0   | exists,up |
| 6  | control02 | 1089M |  135G |    0   |     0   |    0   |     0   | exists,up |
| 7  | control02 | 1089M |  135G |    0   |     0   |    4   |     0   | exists,up |
| 8  | control03 | 1781M |  117G |    0   |     0   |    0   |     0   | exists,up |
| 9  | control03 | 1961M |  117G |    0   |     0   |    7   |   157   | exists,up |
| 10 | control03 | 1089M |  135G |    0   |     0   |    0   |     0   | exists,up |
| 11 | control03 | 1089M |  135G |    0   |     0   |    0   |     0   | exists,up |
| 12 | compute01 | 1462M | 56.5G |    0   |     0   |    0   |     0   | exists,up |
| 13 | compute01 | 1400M | 56.6G |    0   |     0   |    0   |     0   | exists,up |
| 14 | compute01 | 1334M | 56.7G |    0   |     0   |    0   |     6   | exists,up |
| 15 | compute01 | 1426M | 56.6G |    0   |     0   |    0   |     0   | exists,up |
| 16 | compute01 | 1101M |  464G |    0   |     0   |    0   |    19   | exists,up |
| 17 | compute01 | 1089M |  464G |    0   |     0   |    0   |     0   | exists,up |
| 18 | storage01 | 1040M | 57.0G |    0   |     0   |    0   |     0   | exists,up |
| 19 | storage01 | 1040M | 57.0G |    0   |     0   |    0   |     0   | exists,up |
| 20 | storage01 | 1040M | 57.0G |    0   |     0   |    0   |     0   | exists,up |
| 21 | storage01 | 1048M | 57.0G |    0   |     0   |    0   |     0   | exists,up |
| 22 | storage01 | 1081M |  464G |    0   |     0   |    0   |     0   | exists,up |
| 23 | storage01 | 1105M |  464G |    0   |     0   |    0   |     0   | exists,up |
+----+-----------+-------+-------+--------+---------+--------+---------+-----------+

Remove node

Removing storage01 from the cluster

control01:cluster> remove_node
1: compute01
2: storage01
Enter index: 2
this command is only applicable for compute or storage nodes
make sure its running instances have been properly terminated or migrated
shutdown the target host before proceeding
Enter 'YES' to confirm: YES
control01:cluster>

Check the storage status

storage01 node has been removed from the cluster, ceph is recovering the lost replication data automatically

control01> storage
control01:storage> status
  cluster:
    id:     c6e64c49-09cf-463b-9d1c-b6645b4b3b85
    health: HEALTH_WARN
            Reduced data availability: 2 pgs inactive
            Degraded data redundancy: 139/21222 objects degraded (0.655%), 10 pgs degraded

  services:
    mon: 3 daemons, quorum control01,control02,control03
    mgr: control01(active), standbys: control02, control03
    mds: cephfs-1/1/1 up  {0=control01=up:active}, 2 up:standby
    osd: 18 osds: 18 up, 18 in; 510 remapped pgs
    rgw: 3 daemons active

  data:
    pools:   23 pools, 1837 pgs
    objects: 10.50k objects, 12.7GiB
    usage:   25.4GiB used, 2.61TiB / 2.63TiB avail
    pgs:     10.670% pgs unknown
             0.435% pgs not active
             139/21222 objects degraded (0.655%)
             1406 active+clean
             214  active+clean+remapped
             196  unknown
             9    active+recovery_wait+degraded
             5    activating+remapped
             3    activating
             3    active+undersized
             1    active+undersized+degraded+remapped+backfilling

  io:
    client:   13.1KiB/s rd, 0B/s wr, 13op/s rd, 9op/s wr
    recovery: 1.11MiB/s, 4objects/s

+----+-----------+-------+-------+--------+---------+--------+---------+-----------+
| id |    host   |  used | avail | wr ops | wr data | rd ops | rd data |   state   |
+----+-----------+-------+-------+--------+---------+--------+---------+-----------+
| 0  | control01 | 2060M |  117G |    0   |     0   |    0   |     0   | exists,up |
| 1  | control01 | 2025M |  117G |    0   |     0   |    0   |     0   | exists,up |
| 2  | control01 | 1093M |  135G |    0   |     0   |    0   |     0   | exists,up |
| 3  | control01 | 1086M |  135G |   10   |     0   |    5   |    69   | exists,up |
| 4  | control02 | 1668M |  117G |    0   |     0   |    0   |     0   | exists,up |
| 5  | control02 | 2086M |  116G |    0   |     0   |    0   |     0   | exists,up |
| 6  | control02 | 1093M |  135G |    0   |     0   |    0   |     0   | exists,up |
| 7  | control02 | 1094M |  135G |    0   |     0   |    4   |     0   | exists,up |
| 8  | control03 | 1785M |  117G |    0   |     0   |    0   |     0   | exists,up |
| 9  | control03 | 1957M |  117G |    0   |     0   |    0   |     0   | exists,up |
| 10 | control03 | 1093M |  135G |    0   |     0   |    0   |     0   | exists,up |
| 11 | control03 | 1094M |  135G |    0   |     0   |    0   |     0   | exists,up |
| 12 | compute01 | 1463M | 56.5G |    0   |     0   |    0   |     0   | exists,up |
| 13 | compute01 | 1402M | 56.6G |    0   |     0   |    0   |     0   | exists,up |
| 14 | compute01 | 1336M | 56.7G |    0   |     0   |    0   |     0   | exists,up |
| 15 | compute01 | 1427M | 56.6G |    0   |     0   |    0   |     0   | exists,up |
| 16 | compute01 | 1106M |  464G |    0   |     0   |    0   |     0   | exists,up |
| 17 | compute01 | 1094M |  464G |    0   |     0   |    0   |     0   | exists,up |
+----+-----------+-------+-------+--------+---------+--------+---------+-----------+

Check repair services

control1> cluster
control1:cluster> check_repair
          Service  Status  Report
      ClusterLink      ok  [ link(v) ]
  ClusterSettings      ok  [ etcd(v) ]
        HaCluster      ok  [ hacluster(v) ]
         MsgQueue      ok  [ rabbitmq(v) ]
           IaasDb      ok  [ mysql(v) ]
        VirtualIp      ok  [ vip(v) haproxy_ha(v) ]
          Storage      ok  [ ceph(v) ceph_mon(v) ceph_mgr(v) ceph_mds(v) ceph_osd(v) ceph_rgw(v) ]
       ApiService      ok  [ haproxy(v) httpd(v) lmi(v) memcache(v) ]
          Compute      ok  [ nova(v) ]
          Network      ok  [ neutron(v) ]
            Image      ok  [ glance(v) ]
        BlockStor      ok  [ cinder(v) ]
         FileStor      ok  [ manila(v) ]
       ObjectStor      ok  [ swift(v) ]
    Orchestration      ok  [ heat(v) ]
            LBaaS      NG  [ octavia(8) ]
            LBaaS      ok  [ octavia(f) ]
           DNSaaS      ok  [ designate(v) ]
           K8SaaS      ok  [ k3s(v) rancher(v) ]
       InstanceHa      ok  [ masakari(v) ]
 DisasterRecovery      ok  [ freezer(v) ]
         DataPipe      ok  [ zookeeper(v) kafka(v) ]
          Metrics      ok  [ monasca(v) telegraf(v) grafana(v) ]
     LogAnalytics      ok  [ filebeat(v) auditbeat(v) logstash(v) es(v) kibana(v) ]
    Notifications      ok  [ influxdb(v) kapacitor(v) ]
Last updated on 12/8/2020
← Remove Compute nodeCloud Image →
  • Remove Storage node from cluster
    • Connect to controller
    • Check the storage status
    • Remove node
    • Check the storage status
    • Check repair services
Copyright © 2021 Bigstack Co., Ltd.