On this page, you can find an explanation of several ways how to access to the Kubernetes Cluster created in the Cloud Console.
In this article we will assume, that we have already created the following resources, that refer to the Project named Test-Project that was created in the Organization named Test-Organization:
For more information about creating and configuring these resources, see the following articles:
To get the access from the Ubuntu Virtual Machine to the created Kubernetes cluster using CLI, follow the next steps:
Login to your Ubuntu Virtual Machine from which you want to get access to the Kubernetes Cluster API;
for this we use SSH protocol - to find additional information about it, see the article - Access Linux VMssh -i ~/.ssh/id_rsa ubuntu@185.226.42.187
Update Ubuntu package sources by running the following command:sudo apt update
Install Kubernetes Python Client by running the next command:sudo apt install python3-pip
Install openstack cli tool by running two next commands one by one: sudo pip3 install python-openstackclient
sudo pip3 install python-magnumclient
Place RC File of the created CLI User to your Virtual Machine:vi openrc
Check that there were indicated the correct OS_USERNAME and OS_PROJECT_ID and press Esc:wq, then Enter to save the changes:
#!/usr/bin/env bash
export OS_AUTH_URL=https://upper-austria.ventuscloud.eu:443/v3
export OS_PROJECT_ID=e6370ed9ff7c4122b2b89cde7ff7d29f
export OS_PROJECT_NAME="e6370ed9ff7c4122b2b89cde7ff7d29f"
export OS_USER_DOMAIN_NAME="ventus"
if [ -z "$OS_USER_DOMAIN_NAME" ]; then unset OS_USER_DOMAIN_NAME; fi
export OS_PROJECT_DOMAIN_ID="e1780e7170674d5684076a726f683cfd"
if [ -z "$OS_PROJECT_DOMAIN_ID" ]; then unset OS_PROJECT_DOMAIN_ID; fi
unset OS_TENANT_ID
unset OS_TENANT_NAME
export OS_USERNAME="48529f99-8e10-4694-9412-70e1012de805:test-conn"
echo "Please enter your OpenStack Password for project $OS_PROJECT_NAME as user $OS_USERNAME: "
read -sr OS_PASSWORD_INPUT
export OS_PASSWORD=$OS_PASSWORD_INPUT
export OS_REGION_NAME="Upper-Austria"
if [ -z "$OS_REGION_NAME" ]; then unset OS_REGION_NAME; fi
export OS_INTERFACE=public
export OS_IDENTITY_API_VERSION=3
Execute openrc starting with dot:. openrc
Provide the password of the created CLI User and hit Enter - this password will be used to authenticate you in the Cloud Console.
Run the next command to get a list of all clusters created in the corresponding Project and to which your User has access:openstack coe cluster list
In our case the output will be next:
ubuntu@test-2:~$ openstack coe cluster list
+--------------------------------------+-----------+---------+------------+--------------+-----------------+---------------+
| uuid | name | keypair | node_count | master_count | status | health_status |
+--------------------------------------+-----------+---------+------------+--------------+-----------------+---------------+
| 63fe23ab-eaab-46af-9173-131db353c993 | test-cl-1 | mykey | 3 | 1 | UPDATE_COMPLETE | HEALTHY |
| 315c90c5-d74c-429d-b53f-fe0f68025a6e | test-cl-2 | mykey | 1 | 1 | CREATE_COMPLETE | HEALTHY |
+--------------------------------------+-----------+---------+------------+--------------+-----------------+---------------+
Run the following command to get the kubeconfig file for the Cluster named test-cl-2, that you will be accessing:mkdir ~/test-cl-2
openstack coe cluster config --dir ~/test-cl-2 test-cl-2
Export path to created config for as KUBECONFIG env variable:export KUBECONFIG=/home/ubuntu/test-cl-2/config
Install kubectl by running the next command:sudo snap install kubectl --classic
Run the next commands to test that you have access to the selected Cluster and all pods are running:kubectl get nodes
kubectl get pods --all-namespaces
If everything is fine, the output should be close to the following:
ubuntu@test-2:~$ kubectl get nodes
NAME STATUS ROLES AGE VERSION
test-cl-2-ad3t5fyciosn-master-0 Ready master 3d v1.21.1
test-cl-2-ad3t5fyciosn-node-0 Ready <none> 3d v1.21.1
ubuntu@test-2:~$ kubectl get pods --all-namespaces
NAMESPACE NAME READY STATUS RESTARTS AGE
kube-system coredns-d58bdb66c-77tpn 1/1 Running 0 3d
kube-system coredns-d58bdb66c-gxlcq 1/1 Running 0 3d
kube-system dashboard-metrics-scraper-5594697f48-2fdzj 1/1 Running 0 3d
kube-system draino-784c9b8-ztm86 1/1 Running 0 3d
kube-system k8s-keystone-auth-kcrv2 1/1 Running 0 3d
kube-system kube-dns-autoscaler-f57cd985f-lxp2g 1/1 Running 0 3d
kube-system kube-flannel-ds-9rhgc 1/1 Running 0 3d
kube-system kube-flannel-ds-gwvs8 1/1 Running 0 3d
kube-system kubernetes-dashboard-545f6f795d-k9ngg 1/1 Running 0 3d
kube-system magnum-metrics-server-5694f476c-lzjmw 1/1 Running 0 3d
kube-system npd-pd5jq 1/1 Running 0 3d
kube-system openstack-autoscaler-manager-65dd4f65c5-plvr2 1/1 Running 0 3d
kube-system openstack-cinder-csi-controllerplugin-0 5/5 Running 0 3d
kube-system openstack-cinder-csi-nodeplugin-px2bj 2/2 Running 0 3d
kube-system openstack-cloud-controller-manager-mfswf 1/1 Running 1 3d
To get the access from the CentOS Virtual Machine to the created Kubernetes cluster using CLI, follow the next steps:
Login to your CentOS Virtual Machine from which you want to get access to the Kubernetes Cluster API;
for this we use SSH protocol - to find additional information about it, see the article - Access Linux VMssh -i ~/.ssh/id_rsa centos@185.226.43.21
Enable the OpenStack repository;
on CentOS, the extras repository provides the RPM that enables the OpenStack repository. CentOS includes the extras repository by default, so you can simply install the package to enable the OpenStack repository. For CentOS 8, you will also need to enable the PowerTools repository:sudo yum install centos-release-openstack-ussuri
sudo yum config-manager --set-enabled PowerTools
Update CentOS package sources by running the following command:sudo yum upgrade
Install the appropriate OpenStack client for your CentOS version:sudo yum install python3-openstackclient
sudo yum install python3-magnumclient
Place RC File of the created CLI User to your Virtual Machine but first install the nano text editor, if not yet installed:sudo yum install -y nano
nano openrc
Сheck that there were indicated the correct OS_USERNAME and OS_PROJECT_ID and press Ctrl+S, then Ctrl+X to save the changes:
#!/usr/bin/env bash
export OS_AUTH_URL=https://upper-austria.ventuscloud.eu:443/v3
export OS_PROJECT_ID=e6370ed9ff7c4122b2b89cde7ff7d29f
export OS_PROJECT_NAME="e6370ed9ff7c4122b2b89cde7ff7d29f"
export OS_USER_DOMAIN_NAME="ventus"
if [ -z "$OS_USER_DOMAIN_NAME" ]; then unset OS_USER_DOMAIN_NAME; fi
export OS_PROJECT_DOMAIN_ID="e1780e7170674d5684076a726f683cfd"
if [ -z "$OS_PROJECT_DOMAIN_ID" ]; then unset OS_PROJECT_DOMAIN_ID; fi
unset OS_TENANT_ID
unset OS_TENANT_NAME
export OS_USERNAME="48529f99-8e10-4694-9412-70e1012de805:test-conn"
echo "Please enter your OpenStack Password for project $OS_PROJECT_NAME as user $OS_USERNAME: "
read -sr OS_PASSWORD_INPUT
export OS_PASSWORD=$OS_PASSWORD_INPUT
export OS_REGION_NAME="Upper-Austria"
if [ -z "$OS_REGION_NAME" ]; then unset OS_REGION_NAME; fi
export OS_INTERFACE=public
export OS_IDENTITY_API_VERSION=3
Execute openrc starting with dot:. openrc
Provide the password of the created CLI User and hit Enter - this password will be used to authenticate you in the Cloud Console.
Run the next command to get a list of all clusters created in the corresponding Project and to which your User has access:openstack coe cluster list
In our case the output will be next:
[centos@test-3 ~]$ openstack coe cluster list
+--------------------------------------+-----------+---------+------------+--------------+-----------------+---------------+
| uuid | name | keypair | node_count | master_count | status | health_status |
+--------------------------------------+-----------+---------+------------+--------------+-----------------+---------------+
| 63fe23ab-eaab-46af-9173-131db353c993 | test-cl-1 | mykey | 3 | 1 | UPDATE_COMPLETE | HEALTHY |
| 315c90c5-d74c-429d-b53f-fe0f68025a6e | test-cl-2 | mykey | 1 | 1 | CREATE_COMPLETE | HEALTHY |
+--------------------------------------+-----------+---------+------------+--------------+-----------------+---------------+
Run the following command to get the kubeconfig file for the Cluster named test-cl-2, that you will be accessing:mkdir ~/test-cl-2
openstack coe cluster config --dir ~/test-cl-2 test-cl-2
Export path to created config for as KUBECONFIG env variable:export KUBECONFIG=/home/centos/test-cl-2/config
Install the latest release of kubectl, make the kubectl binary executable and move the binary into your PATH by running the next commands:curl -LO https://storage.googleapis.com/kubernetes-release/release/'curl -s https://storage.googleapis.com/kubernetes-release/release/stable.txt'/bin/linux/amd64/kubectl
chmod +x ./kubectl
sudo mv ./kubectl /usr/local/bin/kubectl
Run the next commands to test that you have access to the selected Cluster and all pods are running:kubectl get nodes
kubectl get pods -A
If everything is fine, the output should be close to the following:
[centos@test-3 ~]$ kubectl get nodes
NAME STATUS ROLES AGE VERSION
test-cl-2-ad3t5fyciosn-master-0 Ready master 3d v1.21.1
test-cl-2-ad3t5fyciosn-node-0 Ready <none> 3d v1.21.1
[centos@test-3 ~]$ kubectl get pods -A
NAMESPACE NAME READY STATUS RESTARTS AGE
kube-system coredns-d58bdb66c-77tpn 1/1 Running 0 3d
kube-system coredns-d58bdb66c-gxlcq 1/1 Running 0 3d
kube-system dashboard-metrics-scraper-5594697f48-2fdzj 1/1 Running 0 3d
kube-system draino-784c9b8-ztm86 1/1 Running 0 3d
kube-system k8s-keystone-auth-kcrv2 1/1 Running 0 3d
kube-system kube-dns-autoscaler-f57cd985f-lxp2g 1/1 Running 0 3d
kube-system kube-flannel-ds-9rhgc 1/1 Running 0 3d
kube-system kube-flannel-ds-gwvs8 1/1 Running 0 3d
kube-system kubernetes-dashboard-545f6f795d-k9ngg 1/1 Running 0 3d
kube-system magnum-metrics-server-5694f476c-lzjmw 1/1 Running 0 3d
kube-system npd-pd5jq 1/1 Running 0 3d
kube-system openstack-autoscaler-manager-65dd4f65c5-plvr2 1/1 Running 0 3d
kube-system openstack-cinder-csi-controllerplugin-0 5/5 Running 0 3d
kube-system openstack-cinder-csi-nodeplugin-px2bj 2/2 Running 0 3d
kube-system openstack-cloud-controller-manager-mfswf 1/1 Running 1 3d
Since we created an SSH Key Pair (see Prerequisites of this article), the public key of which is deployed on our Cluster nodes, and the private key on our local system (for example, ~ / .ssh / id_rsa), we can connect to this Kubernetes Cluster remotely from our local server - via SSH to the Master Node of the selected Cluster, which IP is 185.226.41.220. For this, just use the following command:ssh -i ~/.ssh/id_rsa username@*10.111.22.333*
Username for Cluster nodes is core.
Replace username and 10.11.22.333 in the command with your data and specify the appropriate path to your private key. In our example, the command will look like this:ssh -i ~/.ssh/id_rsa core@185.226.41.220
After successfully connecting, you can test that you have access to the selected Cluster and all pods are running, just run the following commands:kubectl get nodes
kubectl get pods --all-namespaces
If everything is fine, the output should be close to the following:
[core@test-cl-2-ad3t5fyciosn-master-0 ~]$ kubectl get nodes
NAME STATUS ROLES AGE VERSION
test-cl-2-ad3t5fyciosn-master-0 Ready master 3d v1.21.1
test-cl-2-ad3t5fyciosn-node-0 Ready <none> 3d v1.21.1
[core@test-cl-2-ad3t5fyciosn-master-0 ~]$ kubectl get pods --all-namespaces
NAMESPACE NAME READY STATUS RESTARTS AGE
kube-system coredns-d58bdb66c-77tpn 1/1 Running 0 3d
kube-system coredns-d58bdb66c-gxlcq 1/1 Running 0 3d
kube-system dashboard-metrics-scraper-5594697f48-2fdzj 1/1 Running 0 3d
kube-system draino-784c9b8-ztm86 1/1 Running 0 3d
kube-system k8s-keystone-auth-kcrv2 1/1 Running 0 3d
kube-system kube-dns-autoscaler-f57cd985f-lxp2g 1/1 Running 0 3d
kube-system kube-flannel-ds-9rhgc 1/1 Running 0 3d
kube-system kube-flannel-ds-gwvs8 1/1 Running 0 3d
kube-system kubernetes-dashboard-545f6f795d-k9ngg 1/1 Running 0 3d
kube-system magnum-metrics-server-5694f476c-lzjmw 1/1 Running 0 3d
kube-system npd-pd5jq 1/1 Running 0 3d
kube-system openstack-autoscaler-manager-65dd4f65c5-plvr2 1/1 Running 0 3d
kube-system openstack-cinder-csi-controllerplugin-0 5/5 Running 0 3d
kube-system openstack-cinder-csi-nodeplugin-px2bj 2/2 Running 0 3d
kube-system openstack-cloud-controller-manager-mfswf 1/1 Running 1 3d