Preparations
a) The current DCAEGEN2 boot strapping process assumes that the networking in the OpenStack is based on the following model:
a private network interconnecting the VMs; and an external network that provides "floating" IP addresses for the VMs.A router connects the two networks. Each VM is assigned two IP addresses, one allocated from the private network when the VM is launched. Then a floating IP is assigned to the VM from the external network. The UUID's of the private and external networks are needed for preparing the inputs.yaml file needed for running the bootstrap container.
b) Add a public key to openStack, note its name (we will use KEYNAME as example for below). Save the private key (we will use KEYPATH as its path example), make sure its permission is globally readable.
c) Load the flowing base VM images to OpenStack: a CentOS 7 base image and a Ubuntu 16.04 base image.
d) Obtain the resource IDs/UUIDs for resources needed by the inputs.yaml file, as explained below, from OpenStack.
On dev machine, set up a directory to hold environment-specific configuration files. Call its path CONFIGDIR.
Put the private key mentioned above into CONFIGDIR as a file named key
, and make it globally readable.
Create a file named inputs.yaml
in CONFIGDIR
1 centos7image_id: '7c8d7524-de1f-490b-8418-db294bfa2d65' 2 ubuntu1604image_id: '4b09c18b-d69e-4ba8-a1bd-562cab91ff20' 3 flavor_id: '4' 4 security_group: '55a11193-6559-4f6c-b2d2-0119a9817062' 5 public_net: 'admin_floating_228_net' 6 private_net: 'onap-f-net' 7 openstack: 8 username: 'MY_LOGIN' 9 password: 'MY_PASSWORD' 10 tenant_name: 'TENANT_NAME' 11 auth_url: 'KEYSTONE_AUTH_URL' 12 region: 'RegionOne' 13 keypair: 'KEYNME' 14 key_filename: '/opt/dcae/key' 15 location_prefix: 'onapr1' 16 location_domain: 'onapdevlab.onap.org' 17 codesource_url: 'https://nexus.onap.org/service/local/repositories/raw/content' 18 codesource_version: 'org.onap.dcaegen2.deployments/releases/scripts'
Here is a line-by-line explanation of the parameters
UUID of the OpenStack's CentOD 7 VM image
UUID of the OpenStack's Ubuntu 16.04 VM image
ID of the OpenStack's VM flavor to be used by DCAEGEN2 VMs
UUID of the OpenStack's security group to be used for DCAEGEN2 VMs
The name of the OpenStack network where public IP addresses are allocated from
The name of the OpenStack network where private IP addresses are allocated from
Group header for OpenStack Keystone parameters
User name
Password
Name of the OpenStack tenant/project where DCAEGEN2 VMs are deployed
penstack authentication API URL, for example 'https://horizon.playground.onap.org:5000/v2.0'
Name of the OpenStack region where DCAEGEN2 VMs are deployed, for example 'RegionOne'
Name of the public key uploaded to OpenStack in the Preparation step
Path to the private key within the container (!! Do not change!!)
Prefix (location code) of all DCAEGEN2 VMs
Domain name of the OpenStack tenant 'onapr1.playground.onap.org'
Location of the raw artifact repo hosting additional boot scripts called by DCAEGEN2 VMs' cloud-init, for example: 'https://nexus.onap.org/service/local/repositories/raw/content'
Path to the boot scripts within the raw artifact repo, for example: 'org.onap.dcaegen2.deployments/releases/scripts'
Create a file in CONFIGDIR called invinputs.yaml
. This contains environment-specific information for the inventory service. (TODO: examples only, not the correct values for the ONAP integration environment.)
1 docker_host_override: "platform_dockerhost" 2 asdc_address: "sdc.onap.org:8443" 3 asdc_uri: "https://sdc.onap.org:8443" 4 asdc_user: "ci" 5 asdc_password: !!str 123456 6 asdc_environment_name: "ONAP-AMDOCS" 7 postgres_user_inventory: "postgres" 8 postgres_password_inventory: "onap123" 9 service_change_handler_image: "nexus3.onap.org:10001/onap/org.onap.dcaegen2.platform.servicechange-handler:latest" 10 inventory_image: "nexus3.onap.org:10001/onap/org.onap.dcaegen2.platform.inventory-api:latest
Here is a line-by-line description of the parameters:
The service name for the platform docker host (should be the same in all environments)
The hostname and port of the SDC service
The URI of the SDC service
The SDC username
The SDC password
The SDC environment name
The postgres user name
The postgres password
The Docker image to be used for the service change handler (should be the same in all environments)
The Docker image to be used for the inventory service (should be the same in all environments)
Create a file in CONFIGDIR called phinputs.yaml
. This contains environment-specific information for the policy handler.
application_config: policy_handler : # parallelize the getConfig queries to policy-engine on each policy-update notification thread_pool_size : 4 # parallelize requests to policy-engine and keep them alive pool_connections : 20 # list of policyName prefixes (filters) that DCAE-Controller handles (=ignores any other policyName values) scope_prefixes : ["DCAE.Config_"] # retry to getConfig from policy-engine on policy-update notification policy_retry_count : 5 policy_retry_sleep : 5 # policy-engine config # These are the url of and the auth for the external system, namely the policy-engine (PDP). # We obtain that info manually from PDP folks at the moment. # In long run we should figure out a way of bringing that info into consul record # related to policy-engine itself. policy_engine : url : "https://policy-engine.onap.org:8081" path_pdp : "/pdp/" path_api : "/pdp/api/" headers : Accept : "application/json" "Content-Type" : "application/json" ClientAuth : "Basic bTAzOTQ5OnBvbGljeVIwY2sk" Authorization : "Basic dGVzdHBkcDphbHBoYTEyMw==" Environment : "TEST" target_entity : "policy_engine" # name of deployment-handler service in consul for policy-handler to direct the policy-updates to deploy_handler : "deployment_handler"
TODO: provide explanations
docker login -u docker -p docker nexus3.onap.org:10001 docker pull nexus3.onap.org:10001/onap/org.onap.dcaegen2.deployments.bootstrap:1.1-latest0 docker run -d --name boot -v CONFIGDIR:/opt/app/installer/config -e "LOCATION=dg2" nexus3.onap.org:10003/onap/org.onap.dcaegen2.deployments.bootstrap:1.1-latest
The container stays up even after the installation is complete. Using the docker exec command to get inside of the container, then run cfy commands to interact with the Cloudify Manager.
docker exec -it boot ./teardown