ユーザ用ツール

サイト用ツール


サイドバー



最近の更新



Tag Cloud

タグ数量
9
10
1
3
1
2
17
7
1
1
4
13
4
1
2
309
1
9
2
7
21
5
1
31
3
3
1
6
1
1
46
1
3
3
1
1
1
1
1
1
33
46
7
9
10
8
10
3
21
2
9
2
2
6
9
1
1
6
12
4
3
3
2
1
8
2
4
1
6
30
2
12
5
3
1
1
2
3
2
1
3
4
3
3
4
2
2
1
1
2
1
2
1
2
1
27
2
1
3
2
2
1
1
2
3
17
5
9
1
1
2
16
2
6
1
2
12
1
3
1
1
4
11
1
6
4
2
7
1
3
3
13
1
4
1
1
10
3
1
17
3
1
5
1
2
1
1
2
9
2
2
3
2
1
3
1
3
1
2
2
2
1
2
6
1
4
3
5
1
3
1
3
3
2
1
1
30
1
3
3
1
1
1
5
5
1
36
2
4
2
1
2
2
3
1
1
1
4
1
2
1
3
1
1
1
2
5
2
5
2
1
1
1
5
3
2
4
4
1
2
2
25
1
1
3
4
1
1
2
1
1
2
1
1
1
1
2
1
1
1
1
1
19
7
1
5
1
1
3
1
2
1
1
2
1
1
1
1
1
1
2
1
1
2
1
1
1
1
1
1
1
4
2
1
1
2
2
2
1
1
2
1
1
12
1
1
1
1
1
1
1
1
1
1
1
1
1
06_virtualization:05_container:18_kubernetes_glusterfs

18 Kubernetes GlusterFS

hostname IP
g-master172.16.0.103
g-work01172.16.0.93
g-work02172.16.0.153
g-work03172.16.0.166

1.GlusterFS Install

GlusterFSはシンプルに下記で作成
03 Ubuntu GlusterFS

2.EndPoint設定

glusterfs-endpoint.yaml

apiVersion: v1
kind: Endpoints
metadata:
  name: glusterfs
  labels:
    storage.k8s.io/name: glusterfs
    storage.k8s.io/part-of: kubernetes-complete-reference
    storage.k8s.io/created-by: ssbostan
subsets:
  - addresses:
      - ip: 172.16.0.93
        hostname: g-work01
      - ip: 172.16.0.153
        hostname: g-work02
      - ip: 172.16.0.166
        hostname: g-work03
    ports:
      - port: 1

作成

kubectl create -f glusterfs-endpoint.yaml

3.Serviceの設定

glusterfs-service.yaml

kind: Service
apiVersion: v1
metadata:
  name: glusterfs
spec:
  ports:
  - port: 1

作成

kubectl create -f glusterfs-service.yaml

4.GlusterFSでボリューム用意

k8s-volumeを作成している事を前提に説明。

# gluster volume status
Status of volume: k8s-volume
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick g-work01:/gluster/volume              49152     0          Y       33456
Brick g-work02:/gluster/volume              49152     0          Y       37407
Brick g-work03:/gluster/volume              49152     0          Y       31930
Self-heal Daemon on localhost               N/A       N/A        Y       33477
Self-heal Daemon on g-work03                N/A       N/A        Y       31951
Self-heal Daemon on g-work02                N/A       N/A        Y       37428
 
Task Status of Volume k8s-volume

PV用のディレクトリを用意しておく

# mount.glusterfs localhost:k8s-volume /mnt/

# df /mnt
Filesystem           1K-blocks    Used Available Use% Mounted on
localhost:k8s-volume  20961280 1968712  18992568  10% /mnt

# mkdir /mnt/pv01

5.PV

pv.yaml

apiVersion: v1
kind: PersistentVolume
metadata:
  name: k8s-volume-pv01
  labels:
    name: k8s-volume-pv01
spec:
  accessModes:
  - ReadWriteMany
  capacity:
    storage: 1Gi
  glusterfs:
    endpoints: glusterfs
    path: k8s-volume/pv01
    readOnly: false

作成

kubectl create -f pv.yaml

6.PVC

pvc.yaml

apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  creationTimestamp: null
  labels:
    io.kompose.service: k8s-volume-pvc01
  name: k8s-volume-pvc01
spec:
  accessModes:
  - ReadWriteMany
  resources:
    requests:
      storage: 1Gi
  selector:
     matchLabels:
       name: k8s-volume-pv01
status: {}

作成

kubectl create -f pvc.yaml

7.確認

PVとPVCができている事を確認

# kubectl get pv
NAME              CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS   CLAIM                      STORAGECLASS   REASON   AGE
k8s-volume-pv01   1Gi        RWX            Retain           Bound    default/k8s-volume-pvc01                           83s

# kubectl get pvc
NAME               STATUS   VOLUME            CAPACITY   ACCESS MODES   STORAGECLASS   AGE
k8s-volume-pvc01   Bound    k8s-volume-pv01   1Gi        RWX                           61s

Podに割り当ててみる

alpine-test.yaml

apiVersion: v1
kind: Pod
metadata:
  name: alpine-test
spec:
  containers:
  - image: alpine
    name: alpine
    command: ["tail", "-f", "/dev/null"]
    volumeMounts:
    - name: data-disk
      mountPath: /data
  volumes:
  - name: data-disk
    persistentVolumeClaim:
      claimName: k8s-volume-pvc01
  terminationGracePeriodSeconds: 0

Pod作成

kubectl create -f alpine-test.yaml

Pod内で確認

# kubectl exec -it alpine-test -- sh
/ # df /data/
Filesystem           1K-blocks      Used Available Use% Mounted on
172.16.0.93:k8s-volume/pv01
                      20961280   1968712  18992568   9% /data

8.障害テスト

1台落としてみる

g-work01 172.16.0.93を落としてみた。

# kubectl get node -o wide
NAME       STATUS     ROLES                  AGE     VERSION   INTERNAL-IP    
g-master   Ready      control-plane,master   5h41m   v1.23.5   172.16.0.103   
g-work01   NotReady   <none>                 5h39m   v1.23.5   172.16.0.93    
g-work02   Ready      <none>                 5h39m   v1.23.5   172.16.0.153   
g-work03   Ready      <none>                 4h5m    v1.23.5   172.16.0.166   

他のworkerからはPingは通らない

落としているので、当然Pingは通らない

# ping 172.16.0.93 -c 3
PING 172.16.0.93 (172.16.0.93) 56(84) bytes of data.
From 172.16.0.103 icmp_seq=1 Destination Host Unreachable
From 172.16.0.103 icmp_seq=2 Destination Host Unreachable
From 172.16.0.103 icmp_seq=3 Destination Host Unreachable

--- 172.16.0.93 ping statistics ---
3 packets transmitted, 0 received, +3 errors, 100% packet loss, time 2025ms

GlusterFSのクラスタからも外れている

# gluster vol status
Status of volume: k8s-volume
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick g-work02:/gluster/volume              49152     0          Y       37407
Brick g-work03:/gluster/volume              49152     0          Y       31930
Self-heal Daemon on localhost               N/A       N/A        Y       37428
Self-heal Daemon on g-work03                N/A       N/A        Y       31951
 
Task Status of Volume k8s-volume
------------------------------------------------------------------------------
There are no active volume tasks

それでも大丈夫

対象のIPにはPingできないのにちゃんとディスクは利用可能

/ # df /data
Filesystem           1K-blocks      Used Available Use% Mounted on
172.16.0.93:k8s-volume/pv01
                      20961280   3537200  17424080  17% /data


/data # dd if=/dev/zero of=TEST bs=1M count=1024
1024+0 records in
1024+0 records out

落としたノード起動後

ちゃんと落ちている間に更新されたファイルが同期されている

# ll /gluster/volume/pv01/
total 3847048
drwxr-xr-x 2 root root         88 Apr 19 08:05 ./
drwxr-xr-x 8 root root         90 Apr 19 07:31 ../
-rw-r--r-- 2 root root 1073741824 Apr 19 08:12 TEST
06_virtualization/05_container/18_kubernetes_glusterfs.txt · 最終更新: 2022/04/19 17:21 by matsui

Yesterday:860 Today:870 Total:237483