title | version | author |
---|---|---|
Getting Started Gaia-X Demonstrator @ PlusServer |
2021-05-06-001 |
Mathias Fechner |
- UI (Horizon): https://ui.gx-scs.sovereignit.cloud/
- auth_url (Keystone): https://api.gx-scs.sovereignit.cloud:5000
- keycloak_url: https://iam.gx-scs.sovereignit.cloud
- S3 endpoint (same as swift from catalog):https://api.gx-scs.sovereignit.cloud:8080
- Core Components (Keystone - auth, Cinder - block storage, Nova - compute, Glance - images, Neutron - networking)
- Heat (Orchestration)
- DNS (Designate)
- Loadbalancer (Octavia)
- Object Storage (Swift and S3 protocol via RadosGW, see below)
- Secrets Management (Barbican) and Cloud Container Engine (Magnum)
- A standard set of current Linux cloud images (Ubuntu, Debian, Fedora, CentOS, OpenSUSE,
Cirros, Clear Linux) with the standard cloud-init included for processing of user_data
to customize them. The images are kept up-to-date, so a
package_upgrade: true
should not lead to excessive boot time.
- Block-storage
- Object-storage (S3 and Swift via radosgw)
We have not yet aligned on a standard built-in solution, as none of the Open Source solutions fully uses and implements the k8s cluster API really well. We have done successful PoCs with Rancher, Kubermatic and SAP Gardener, and the automation for them can be found on https://github.com/SovereignCloudStack/
We have currently people working on the OpenStack k8s cluster API provider and on Gardener to integrate something out of the box, possibly leveraging Magnum to create a seed cluster on demand.
Continuous automated rollout in docker containers
- Base install with MaaS, Netbox, Linux, KVM on Bare Metal
- OSISM with kolla-ansible, ceph-ansible, infra tooling (database, msgQ, ...)
- Monitoring tools (netdata, prometheus, ...) for operations
- Lifecycle Management & Updates (patchman)
- Validation (CI) with smoke tests, refstack, openstack-health-monitor
- OVN 20.09.0
in OVS Networks VXLAN-based MTU was 1450, in OVN we are using Geneve where the MTU changed to 1442. in some Container deployments with SDN it is recommendable to use a MTU with maximum 1400
apt install python3-openstackclient
dnf install https://www.rdoproject.org/repos/rdo-release.el8.rpm;
dnf install python-openstackclient
yum install centos-release-openstack-victoria;
yum config-manager --set-enabled PowerTools;
yum install python-openstackclient
#zypper ar -f obs://Cloud:OpenStack:Victoria/ Cloud:OpenStack:Victoria
zypper install python-PROJECTclient
brew install openstackclient
pip install python-openstackclient
C:\>pip install python-PROJECTclient
# openstack
(openstack) server ...
(openstack) network ...
(openstack) volume ...
(openstack) image ...
Continue "..." with enter to show all options
Under the item API: you will find the
OpenRC file to source in to your environment or alternatively (better) you add the downloadable
clouds.yaml
(and secure.yaml
) to your ~/.config/openstack/
directory and set OS_CLOUD
or
pass --os-cloud=
.
Create AWS like credentials with openstack ec2 credentials create
.
If you use libs3, store the access field in S3_ACCESS_KEY_ID
and the secret field in
S3_SECRET_ACCESS_KEY
and set S3_HOSTNAME=gx-scs.okeanos.dev:8080
.
You will see the same buckets (containers) and objects whether you access your object store
via the swift or via the s3 protocol.
-
ansible os_auth, os_server
-
terraform openstack/provider OSISM
For your login you will need:
- Domain
- Username
- Password
in case of Keycloak you will need just Username and Password.
Contact us if you want to use keycloak to federate identities from another identity provider.
OpenStack SovereignCloudStack OSISM Ceph OpenvSwitch ansible Terraform cloud-init