[Kubernetes] Install Kubernetes using Kubekey(v3.0.7) Artifact on VirtualBox
[Kubernetes] Install Kubernetes using Kubekey(v3.0.7) Artifact on VirtualBox
offline 설치 위한 artifact 참고
- version 참고
- kubernetes와 관련된 image는 https://github.com/kubesphere/ks-installer/releases에서 주요 release에만 포함되는 image-list.txt파일을 참고
- kubekey의 버전별로 kubernetes, kubesphere의 최신 지원 버전이 있음
- kubekey/version/components.json
- kubekey/cmd/kk/pkg/version/kubesphere/version_enum.go
- kubekey/cmd/kk/pkg/version/kubernetes/version_enum.go
- default 버전에 대한 설정은 kubekey/cmd/kk/apis/kubekey/v1alpha2/default.go 파일에 있다
- https://github.com/kubesphere/kubekey/blob/v3.0.13/docs/manifest_and_artifact.md
- https://github.com/kubesphere/ks-installer/releases/download/v3.4.1/images-list.txt
- https://kubesphere.io/docs/v3.4/installing-on-linux/introduction/air-gapped-installation
- https://github.com/kubesphere/kubekey/blob/v3.0.13/docs/manifest-example.md
script 다운로드
1
curl -sfL https://get-kk.kubesphere.io | VERSION=v3.0.7 sh -
artifact-3.0.7.yaml 작성
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
apiVersion: kubekey.kubesphere.io/v1alpha2
kind: Manifest
metadata:
name: artifact-v3.0.7
spec:
arches:
- amd64
operatingSystems:
- arch: amd64
type: linux
id: ubuntu
version: "20.04"
osImage: Ubuntu 20.04.4 LTS
repository:
iso:
localPath: ""
url: "https://github.com/kubesphere/kubekey/releases/download/v3.0.7/ubuntu-20.04-debs-amd64.iso"
kubernetesDistributions:
- type: kubernetes
version: v1.24.9
components:
helm:
version: v3.9.0
cni:
version: v0.9.1
etcd:
version: v3.4.13
calicoctl:
version: v3.23.2
## For now, if your cluster container runtime is containerd, KubeKey will add a docker 20.10.8 container runtime in the below list.
## The reason is KubeKey creates a cluster with containerd by installing a docker first and making kubelet connect the socket file of containerd which docker contained.
containerRuntimes:
- type: docker
version: 20.10.8
- type: containerd
version: 1.6.4
crictl:
version: v1.24.0
docker-registry:
version: "2"
harbor:
version: v2.5.3
docker-compose:
version: v2.2.2
images:
- docker.io/kubesphere/kube-apiserver:v1.27.2
- docker.io/kubesphere/kube-apiserver:v1.26.5
- docker.io/kubesphere/kube-apiserver:v1.25.10
- docker.io/kubesphere/kube-apiserver:v1.24.9
- docker.io/kubesphere/kube-controller-manager:v1.27.2
- docker.io/kubesphere/kube-controller-manager:v1.26.5
- docker.io/kubesphere/kube-controller-manager:v1.25.10
- docker.io/kubesphere/kube-controller-manager:v1.24.9
- docker.io/kubesphere/kube-scheduler:v1.27.2
- docker.io/kubesphere/kube-scheduler:v1.26.5
- docker.io/kubesphere/kube-scheduler:v1.25.10
- docker.io/kubesphere/kube-scheduler:v1.24.9
- docker.io/kubesphere/kube-proxy:v1.27.2
- docker.io/kubesphere/kube-proxy:v1.26.5
- docker.io/kubesphere/kube-proxy:v1.25.10
- docker.io/kubesphere/kube-proxy:v1.24.9
- docker.io/kubesphere/pause:3.8
- docker.io/kubesphere/pause:3.7
- docker.io/kubesphere/pause:3.6
- docker.io/kubesphere/pause:3.5
- docker.io/kubesphere/pause:3.4.1
- docker.io/coredns/coredns:1.8.6
- docker.io/coredns/coredns:1.8.0
- docker.io/calico/cni:v3.23.2
- docker.io/calico/kube-controllers:v3.23.2
- docker.io/calico/node:v3.23.2
- docker.io/calico/pod2daemon-flexvol:v3.23.2
- docker.io/calico/typha:v3.23.2
- docker.io/kubesphere/flannel:v0.12.0
- docker.io/openebs/provisioner-localpv:3.3.0
- docker.io/openebs/linux-utils:3.3.0
- docker.io/library/haproxy:2.3
- docker.io/kubesphere/nfs-subdir-external-provisioner:v4.0.2
- docker.io/kubesphere/k8s-dns-node-cache:1.15.12
# https://github.com/kubesphere/ks-installer/releases/download/v3.3.2/images-list.txt
##kubesphere-images
- docker.io/kubesphere/ks-installer:v3.4.1
- docker.io/kubesphere/ks-installer:v3.3.2
- docker.io/kubesphere/ks-apiserver:v3.4.1
- docker.io/kubesphere/ks-apiserver:v3.3.2
- docker.io/kubesphere/ks-console:v3.4.1
- docker.io/kubesphere/ks-console:v3.3.2
- docker.io/kubesphere/ks-controller-manager:v3.4.1
- docker.io/kubesphere/ks-controller-manager:v3.3.2
- docker.io/kubesphere/kubectl:v1.22.0
- docker.io/kubesphere/kubectl:v1.20.0
- docker.io/kubesphere/kubefed:v0.8.1
- docker.io/kubesphere/tower:v0.2.1
- docker.io/kubesphere/tower:v0.2.0
- docker.io/minio/minio:RELEASE.2019-08-07T01-59-21Z
- docker.io/minio/mc:RELEASE.2019-08-07T23-14-43Z
- docker.io/csiplugin/snapshot-controller:v4.0.0
- docker.io/kubesphere/nginx-ingress-controller:v1.3.1
- docker.io/kubesphere/nginx-ingress-controller:v1.1.0
- docker.io/mirrorgooglecontainers/defaultbackend-amd64:1.4
- docker.io/kubesphere/metrics-server:v0.4.2
- docker.io/library/redis:5.0.14-alpine
- docker.io/library/haproxy:2.0.25-alpine
- docker.io/library/alpine:3.14
- docker.io/osixia/openldap:1.3.0
- docker.io/kubesphere/netshoot:v1.0
##kubeedge-images
- docker.io/kubeedge/cloudcore:v1.13.0
- docker.io/kubeedge/cloudcore:v1.9.2
- docker.io/kubesphere/iptables-manager:v1.13.0
- docker.io/kubeedge/iptables-manager:v1.9.2
- docker.io/kubesphere/edgeservice:v0.3.0
- docker.io/kubesphere/edgeservice:v0.2.0
##gatekeeper-images
- docker.io/openpolicyagent/gatekeeper:v3.5.2
##openpitrix-images
- docker.io/kubesphere/openpitrix-jobs:v3.3.2
##kubesphere-devops-images
- docker.io/kubesphere/devops-apiserver:ks-v3.4.1
- docker.io/kubesphere/devops-apiserver:ks-v3.3.2
- docker.io/kubesphere/devops-controller:ks-v3.4.1
- docker.io/kubesphere/devops-controller:ks-v3.3.2
- docker.io/kubesphere/devops-tools:ks-v3.4.1
- docker.io/kubesphere/devops-tools:ks-v3.3.2
- docker.io/kubesphere/ks-jenkins:v3.4.0-2.319.3-1
- docker.io/kubesphere/ks-jenkins:v3.3.0-2.319.1
- docker.io/jenkins/inbound-agent:4.10-2
- docker.io/kubesphere/builder-base:v3.2.2
- docker.io/kubesphere/builder-nodejs:v3.2.0
- docker.io/kubesphere/builder-maven:v3.2.1-jdk11
- docker.io/kubesphere/builder-maven:v3.2.0
- docker.io/kubesphere/builder-python:v3.2.0
- docker.io/kubesphere/builder-go:v3.2.2-1.18
- docker.io/kubesphere/builder-go:v3.2.2-1.17
- docker.io/kubesphere/builder-go:v3.2.2-1.16
- docker.io/kubesphere/builder-go:v3.2.0
- docker.io/kubesphere/builder-base:v3.2.2-podman
- docker.io/kubesphere/builder-nodejs:v3.2.0-podman
- docker.io/kubesphere/builder-maven:v3.2.1-jdk11-podman
- docker.io/kubesphere/builder-maven:v3.2.0-podman
- docker.io/kubesphere/builder-python:v3.2.0-podman
- docker.io/kubesphere/builder-go:v3.2.0-podman
- docker.io/kubesphere/builder-go:v3.2.2-1.18-podman
- docker.io/kubesphere/builder-go:v3.2.2-1.17-podman
- docker.io/kubesphere/builder-go:v3.2.2-1.16-podman
- docker.io/kubesphere/s2ioperator:v3.2.1
- docker.io/kubesphere/s2irun:v3.2.0
- docker.io/kubesphere/s2i-binary:v3.2.0
- docker.io/kubesphere/tomcat85-java11-centos7:v3.2.0
- docker.io/kubesphere/tomcat85-java11-runtime:v3.2.0
- docker.io/kubesphere/tomcat85-java8-centos7:v3.2.0
- docker.io/kubesphere/tomcat85-java8-runtime:v3.2.0
- docker.io/kubesphere/java-11-centos7:v3.2.0
- docker.io/kubesphere/java-11-runtime:v3.2.0
- docker.io/kubesphere/java-8-centos7:v3.2.0
- docker.io/kubesphere/java-8-runtime:v3.2.0
- docker.io/kubesphere/nodejs-8-centos7:v3.2.0
- docker.io/kubesphere/nodejs-6-centos7:v3.2.0
- docker.io/kubesphere/nodejs-4-centos7:v3.2.0
- docker.io/kubesphere/python-36-centos7:v3.2.0
- docker.io/kubesphere/python-35-centos7:v3.2.0
- docker.io/kubesphere/python-34-centos7:v3.2.0
- docker.io/kubesphere/python-27-centos7:v3.2.0
- quay.io/argoproj/argocd:v2.3.3
- quay.io/argoproj/argocd-applicationset:v0.4.1
- ghcr.io/dexidp/dex:v2.30.2
- docker.io/library/redis:6.2.6-alpine
##kubesphere-monitoring-images
- docker.io/jimmidyson/configmap-reload:v0.7.1
- docker.io/jimmidyson/configmap-reload:v0.5.0
- docker.io/prom/prometheus:v2.39.1
- docker.io/prom/prometheus:v2.34.0
- docker.io/kubesphere/prometheus-config-reloader:v0.55.1
- docker.io/kubesphere/prometheus-operator:v0.55.1
- docker.io/kubesphere/kube-rbac-proxy:v0.11.0
- docker.io/kubesphere/kube-state-metrics:v2.6.0
- docker.io/kubesphere/kube-state-metrics:v2.5.0
- docker.io/prom/node-exporter:v1.3.1
- docker.io/prom/alertmanager:v0.23.0
- docker.io/thanosio/thanos:v0.31.0
- docker.io/thanosio/thanos:v0.25.2
- docker.io/grafana/grafana:8.3.3
- docker.io/kubesphere/kube-rbac-proxy:v0.11.0
- docker.io/kubesphere/kube-rbac-proxy:v0.8.0
- docker.io/kubesphere/notification-manager-operator:v2.3.0
- docker.io/kubesphere/notification-manager-operator:v1.4.0
- docker.io/kubesphere/notification-manager:v2.3.0
- docker.io/kubesphere/notification-manager:v1.4.0
- docker.io/kubesphere/notification-tenant-sidecar:v3.2.0
##kubesphere-logging-images
- docker.io/kubesphere/elasticsearch-curator:v5.7.6
- docker.io/kubesphere/opensearch-curator:v0.0.5
- docker.io/kubesphere/elasticsearch-oss:6.8.22
- docker.io/opensearchproject/opensearch:2.6.0
- docker.io/opensearchproject/opensearch-dashboards:2.6.0
- docker.io/kubesphere/fluentbit-operator:v0.14.0
- docker.io/kubesphere/fluentbit-operator:v0.13.0
- docker.io/library/docker:19.03
- docker.io/kubesphere/fluent-bit:v1.9.4
- docker.io/kubesphere/fluent-bit:v1.8.11
- docker.io/kubesphere/log-sidecar-injector:v1.2.0
- docker.io/elastic/filebeat:6.7.0
- docker.io/kubesphere/kube-events-operator:v0.6.0
- docker.io/kubesphere/kube-events-operator:v0.4.0
- docker.io/kubesphere/kube-events-exporter:v0.6.0
- docker.io/kubesphere/kube-events-exporter:v0.4.0
- docker.io/kubesphere/kube-events-ruler:v0.6.0
- docker.io/kubesphere/kube-events-ruler:v0.4.0
- docker.io/kubesphere/kube-auditing-operator:v0.2.0
- docker.io/kubesphere/kube-auditing-webhook:v0.2.0
##istio-images
- docker.io/istio/pilot:1.14.6
- docker.io/istio/pilot:1.11.1
- docker.io/istio/proxyv2:1.14.6
- docker.io/istio/proxyv2:1.11.1
- docker.io/jaegertracing/jaeger-operator:1.29
- docker.io/jaegertracing/jaeger-operator:1.27
- docker.io/jaegertracing/jaeger-agent:1.29
- docker.io/jaegertracing/jaeger-agent:1.27
- docker.io/jaegertracing/jaeger-collector:1.29
- docker.io/jaegertracing/jaeger-collector:1.27
- docker.io/jaegertracing/jaeger-query:1.29
- docker.io/jaegertracing/jaeger-query:1.27
- docker.io/jaegertracing/jaeger-es-index-cleaner:1.29
- docker.io/jaegertracing/jaeger-es-index-cleaner:1.27
- docker.io/kubesphere/kiali-operator:v1.50.1
- docker.io/kubesphere/kiali-operator:v1.38.1
- docker.io/kubesphere/kiali:v1.50
- docker.io/kubesphere/kiali:v1.38
# ##example-images
# - docker.io/library/busybox:1.31.1
# - docker.io/library/nginx:1.14-alpine
# - docker.io/joosthofman/wget:1.0
# - docker.io/nginxdemos/hello:plain-text
# - docker.io/library/wordpress:4.8-apache
# - docker.io/mirrorgooglecontainers/hpa-example:latest
# - docker.io/fluent/fluentd:v1.4.2-2.0
# - docker.io/library/perl:latest
# - docker.io/kubesphere/examples-bookinfo-productpage-v1:1.16.2
# - docker.io/kubesphere/examples-bookinfo-reviews-v1:1.16.2
# - docker.io/kubesphere/examples-bookinfo-reviews-v2:1.16.2
# - docker.io/kubesphere/examples-bookinfo-details-v1:1.16.2
# - docker.io/kubesphere/examples-bookinfo-ratings-v1:1.16.3
# ##weave-scope-images
# - docker.io/weaveworks/scope:1.13.0
registry:
auths:
"docker.io":
username: "username"
password: "password"
Export Artifact
1
sudo ./kk artifact export -m artifact-3.0.7.yaml -o artifact-3.0.7.tar.gz
Cluster 설치를 위한 config 파일 생성 및 작성
config 파일 생성
1
sudo ./kk create config --with-kubesphere v3.3.2 --with-kubernetes v1.24.9 -f config-sample.yaml
config 파일 작성
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227
apiVersion: kubekey.kubesphere.io/v1alpha2 kind: Cluster metadata: name: sample spec: hosts: - {name: manage-master, address: 192.168.10.100, internalAddress: 192.168.10.100, user: root, password: vagrant} - {name: manage-worker-1, address: 192.168.10.110, internalAddress: 192.168.10.110, user: root, password: vagrant} - {name: manage-worker-2, address: 192.168.10.120, internalAddress: 192.168.10.120, user: root, password: vagrant} roleGroups: etcd: - manage-master control-plane: - manage-master worker: - manage-worker-1 - manage-worker-2 registry: - manage-worker-1 controlPlaneEndpoint: ## Internal loadbalancer for apiservers # internalLoadbalancer: haproxy #domain: lb.kubesphere.local domain: 192.168.10.100 address: "" port: 6443 kubernetes: version: v1.24.9 clusterName: cluster.local autoRenewCerts: true containerManager: containerd etcd: type: kubekey network: plugin: calico kubePodsCIDR: 10.233.64.0/18 kubeServiceCIDR: 10.233.0.0/18 ## multus support. https://github.com/k8snetworkplumbingwg/multus-cni multusCNI: enabled: false registry: type: harbor auths: "dockerhub.kubekey.local": username: admin password: Harbor12345 privateRegistry: "dockerhub.kubekey.local" namespaceOverride: "kubesphereio" registryMirrors: [] insecureRegistries: [] addons: [] --- apiVersion: installer.kubesphere.io/v1alpha1 kind: ClusterConfiguration metadata: name: ks-installer namespace: kubesphere-system labels: version: v3.3.2 spec: persistence: storageClass: "" authentication: jwtSecret: "" zone: "" local_registry: "" namespace_override: "" # dev_tag: "" etcd: monitoring: false endpointIps: localhost port: 2379 tlsEnable: true common: core: console: enableMultiLogin: true port: 30880 type: NodePort # apiserver: # resources: {} # controllerManager: # resources: {} redis: enabled: false volumeSize: 2Gi openldap: enabled: false volumeSize: 2Gi minio: volumeSize: 20Gi monitoring: # type: external endpoint: http://prometheus-operated.kubesphere-monitoring-system.svc:9090 GPUMonitoring: enabled: false gpu: kinds: - resourceName: "nvidia.com/gpu" resourceType: "GPU" default: true es: # master: # volumeSize: 4Gi # replicas: 1 # resources: {} # data: # volumeSize: 20Gi # replicas: 1 # resources: {} logMaxAge: 7 elkPrefix: logstash basicAuth: enabled: false username: "" password: "" externalElasticsearchHost: "" externalElasticsearchPort: "" alerting: enabled: false # thanosruler: # replicas: 1 # resources: {} auditing: enabled: false # operator: # resources: {} # webhook: # resources: {} devops: enabled: false # resources: {} jenkinsMemoryLim: 8Gi jenkinsMemoryReq: 4Gi jenkinsVolumeSize: 8Gi events: enabled: false # operator: # resources: {} # exporter: # resources: {} # ruler: # enabled: true # replicas: 2 # resources: {} logging: enabled: false logsidecar: enabled: true replicas: 2 # resources: {} metrics_server: enabled: false monitoring: storageClass: "" node_exporter: port: 9100 # resources: {} # kube_rbac_proxy: # resources: {} # kube_state_metrics: # resources: {} # prometheus: # replicas: 1 # volumeSize: 20Gi # resources: {} # operator: # resources: {} # alertmanager: # replicas: 1 # resources: {} # notification_manager: # resources: {} # operator: # resources: {} # proxy: # resources: {} gpu: nvidia_dcgm_exporter: enabled: false # resources: {} multicluster: clusterRole: none network: networkpolicy: enabled: false ippool: type: none topology: type: none openpitrix: store: enabled: false servicemesh: enabled: false istio: components: ingressGateways: - name: istio-ingressgateway enabled: false cni: enabled: false edgeruntime: enabled: false kubeedge: enabled: false cloudCore: cloudHub: advertiseAddress: - "" service: cloudhubNodePort: "30000" cloudhubQuicNodePort: "30001" cloudhubHttpsNodePort: "30002" cloudstreamNodePort: "30003" tunnelNodePort: "30004" # resources: {} # hostNetWork: false iptables-manager: enabled: true mode: "external" # resources: {} # edgeService: # resources: {} terminal: timeout: 600
registry 설치
1
sudo ./kk init registry -f config-sample.yaml -a artifact-3.0.7.tar.gz
[ERROR] ssh error
- 각 node 별로 ssh가 안될시 root passwd가 맞지 않아 발생함.
- vagrant에서 vm이 생성되면 root 비번을 설정해줘야 하는 듯
1sudo passwd root
Harbor 인증서 복사 및 업데이트 (harbor curl: (60) SSL certificate problem: unable to get local issuer certificate)
1
2
3
4
5
6
7
8
9
10
11
12
13
sudo cp /etc/docker/certs.d/dockerhub.kubekey.local/ca.crt /usr/local/share/ca-certificates/harbor-ca.crt
scp -i /home/vagrant/.ssh/id_rsa /usr/local/share/ca-certificates/harbor-ca.crt root@192.168.10.110:/usr/local/share/ca-certificates/harbor-ca.crt
scp -i /home/vagrant/.ssh/id_rsa /usr/local/share/ca-certificates/harbor-ca.crt root@192.168.10.120:/usr/local/share/ca-certificates/harbor-ca.crt
# 각 node 별로 아래 작업
sudo update-ca-certificates
# 인증서 적용 확인
ls -lrt /etc/ssl/certs
- harbor-ca.pem -> /usr/local/share/ca-certificates/harbor-ca.crt
- ca-certificates.crt
systemctl restart containerd
harbor 주소 : [harbor 설치한 ip]:80
Harbor 프로젝트 생성 및 수정
1
curl -O https://raw.githubusercontent.com/kubesphere/ks-installer/master/scripts/create_project_harbor.sh
Harbor 프로젝트 수정 및 url 수정(https://dockerhub.kubekey.local)
파일 편집
1
vi create_project_harbor.sh
url 수정(https://dockerhub.kubekey.local)
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52
#!/usr/bin/env bash # Copyright 2018 The KubeSphere Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. url="https://dockerhub.kubekey.local" #Change the value of url to https://dockerhub.kubekey.local. user="admin" passwd="Harbor12345" harbor_projects=(library kubesphereio kubesphere argoproj calico coredns openebs csiplugin minio mirrorgooglecontainers osixia prom thanosio jimmidyson grafana elastic istio jaegertracing jenkins weaveworks openpitrix joosthofman nginxdemos fluent kubeedge openpolicyagent ) for project in "${harbor_projects[@]}"; do echo "creating $project" curl -u "${user}:${passwd}" -X POST -H "Content-Type: application/json" "${url}/api/v2.0/projects" -d "{ \"project_name\": \"${project}\", \"public\": true}" -k #Add -k at the end of the curl command. done
파일 권한 변경
1
chmod +x create_project_harbor.sh
실행
1
./create_project_harbor.sh
image 별도로 push 방법
1
sudo ./kk artifact image push -f config-sample.yaml -a artifact-3.0.7.tar.gz
[ERROR] Harbor에 image push 할 때 Unauthorized 에러 발생 때
- 다시 로그인
1 docker login [your.host.com]:port -u username -p password
Cluster 설치
1
sudo ./kk create cluster -f config-sample.yaml -a artifact-3.0.7.tar.gz
Install operating system packages
1 sudo ./kk create cluster -f config-sample.yaml -a artifact-3.0.7.tar.gz --with-packages
--skip-push-images
를 추가하면 harbor에 image를 push하는 과정으로 생략할 수 있다.
1 sudo ./kk create cluster -f config-sample.yaml -a artifact-3.0.7.tar.gz --skip-push-images
kubekey command 참고
Cluster 설치하면서 log 확인
1
kubectl logs -n kubesphere-system $(kubectl get pod -n kubesphere-system -l 'app in (ks-install, ks-installer)' -o jsonpath='{.items[0].metadata.name}') -f
Kubernetes 일반 유저 일 때
1
2
3
mkdir -p $HOME/.kube
sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
sudo chown $(id -u):$(id -g) $HOME/.kube/config
만약 일반 계정에서 아래와 sudo 명령어 없이 kubectl 명령어 사용시 아래와 같은 오류가 발생하면
- [ERROR] error loading config file
/etc/kubernetes/admin.conf
: open /etc/kubernetes/admin.conf: permission denied
- 아래 명령어를 입력하면 sudo 없이 사용 가능하다.
1 export KUBECONFIG=$HOME/.kube/config
[ERROR] error making pod data directories: mkdir /var/lib/kubelet/pods/86cfe394-ba32-4a9f-ad65-1fb21f98a4ba: read-only file system
1 2 3 chown -R kubelet:kubelet /var/lib/kubelet/pods chmod 750 /var/lib/kubelet/pods systemctl restart kubelet
This post is licensed under CC BY 4.0 by the author.