Step 1
|
Subscribe to Cisco Nexus Dashboard product in AWS Marketplace.
-
Log into your AWS account and navigate to the AWS Management Console
-
Navigate to .
-
Click Manage Subscriptions.
-
Click Discover products.
-
Search for Cisco Nexus Dashboard and click the result.
-
In the product page, click Continue to Subscribe.
-
Click Accept Terms.
It may take a couple of minutes for the subscription to be processed.
-
Finally click Continue to Configuration.
|
Step 2
|
Select software options and region.
-
From the Delivery Method dropdown, select Cisco Nexus Dashboard for Cloud .
-
From the Software Version dropdown, select the version you want to deploy.
-
From the Region dropdown, select the regions where the template will be deployed.
This must be the same region where you created your VPC.
-
Click Continue to Launch.
The product page appears, which shows a summary of your configuration and enables you to launch the cloud formation template.
|
Step 3
|
From the Choose Action, select Launch CloudFormation and click Launch.
The Create stack page appears.
|
Step 4
|
Create stack.
-
In the Prerequisite - Prepare template area, select Template is ready .
-
In the Specify Template area, select Amazon S3 URL for the template source.
The template will be populated automatically.
-
Click Next to continue.
The Specify stack details page appears.
|
Step 5
|
Specify stack details.
-
Provide the Stack name.
-
From the VPC identifier dropdown, select the VPC you created.
For example, vpc-038f83026b6a48e98(10.176.176.0/24) .
-
In the ND cluster Subnet block, provide the VPC subnet CIDR block.
Choose a subnet from the VPC CIDR that you defined. You can provide a smaller subnet or use the whole CIDR. The CIDR can be
a /24 or /25 subnet and will be segmented to be used across the availability zones.
For example, 10.176.176.0/24 .
-
From the Availability Zones dropdown, select one or more available zones.
We recommend you choose 3 availability zones. For regions that support only 2 availability zones, 2nd and 3rd nodes of the
cluster will launch in the second availability zone.
-
From the Number of Availability Zones dropdown, select the number of zones you added in the previous substep.
Ensure that the number matches the number of availability zones you selected in the previous substep.
-
Enable Data Interface EIP support.
This field enables external connectivity for the node. External connectivity is required for communication with Cisco ACI
fabrics outside AWS.
-
In the Password and Confirm Password fields, provide the password.
This password will be used for the Nexus Dashboard's rescue-user login, as well as the initial password for the GUI's admin user.
Note
|
You must provide the same password for all nodes or the cluster creation will fail.
|
-
From the SSH key pair dropdown, select the key pair you created.
-
In the Access control field, provide the external network allowed to access the cluster.
For example, 0.0.0.0/0 to be able to access the cluster from anywhere.
-
Click Next to continue.
|
Step 6
|
In the Advanced options screen, simply click Next.
|
Step 7
|
In the Review screen, verify template configuration and click Create stack.
|
Step 8
|
Wait for the deployment to complete, then start the VMs.
You can view the status of the instance deployment in the CloudFormation page, for example CREATE_IN_PROGRESS . You can click the refresh button in the top right corner of the page to update the status.
When the status changes to CREATE_COMPLETE , you can proceed to the next step.
|
Step 9
|
Note down all nodes' public IP addresses.
-
After all instances are deployed, navigate to the AWS console's page.
-
Note down which node is labeled as FirstMaster .
You will use this node's public IP address to complete cluster configuration.
-
Note down all nodes' public IP addresses.
You will provide this information to the GUI bootstrap wizard in the following steps.
|
Step 10
|
Enable password-based login on all nodes.
By default only PEM-based login is enabled for each node. To be able to SSH into the nodes using a password, as required by
the GUI setup wizard, you will need to explicitly enable password-based logins.
Note
|
You must enable password-based login on all nodes before proceeding to cluster bootstrap described in the following steps
or you will not be able to complete the cluster configuration.
|
-
SSH into one of the instances using its public IP address and the PEM file.
-
Enable password-based login.
On each node, run the following command:
# acs login-prompt enable
-
Repeat this step for the other two instances.
|
Step 11
|
Open your browser and navigate to https://<first-node-public-ip> to open the GUI.
Note
|
You must use the public IP address of the first node (FirstMaster ) or cluster configuration cannot be completed.
|
The rest of the configuration workflow takes place from the first node's GUI. You do not need to log in to or configure the
other two nodes directly.
Enter the password you provided for the first node and click Login
|
Step 12
|
Provide the Cluster Details.
In the Cluster Details screen of the Cluster Bringup wizard, provide the following information:
-
Provide the Cluster Name for this Nexus Dashboard cluster.
The cluster name must follow the RFC-1123 requirements.
-
(Optional) If you want to enable IPv6 functionality for the cluster, check the Enable IPv6 checkbox.
-
(Optional) If you want to enable NTP server authentication, click Add NTP Key.
In the additional fields, provide the following information:
-
NTP Key – a cryptographic key that is used to authenticate the NTP traffic between the Nexus Dashboard and the NTP server(s). You
will define the NTP servers in the following step, and multiple NTP servers can use the same NTP key.
-
Key ID – each NTP key must be assigned a unique key ID, which is used to identify the appropriate key to use when verifying the
NTP packet.
-
Auth Type – this release supports MD5 , SHA , and AES128CMAC authentication types.
-
Choose whether this key is Trusted. Untrusted keys cannot be used for NTP authentication.
After you've entered the information, click the checkmark icon to save it.
-
Click +Add NTP Host to add one or more NTP servers.
In the additional fields, provide the following information:
-
NTP Host – you must provide an IP address; fully qualified domain name (FQDN) are not supported.
-
Key ID – if you want to enable NTP authentication for this server, provide the key ID of the NTP key you defined in the previous
step.
-
Choose whether this NTP server is Preferred.
After you've entered the information, click the checkmark icon to save it.
Note
|
If the node into which you are logged in is configured with only an IPv4 address, but you have checked Enable IPv6 in a previous step and provided an IPv6 address for an NTP server, you will get the following validation error:
This is because the node does not have an IPv6 address yet (you will provide it in the next step) and is unable to connect
to an IPv6 address of the NTP server.
In this case, simply finish providing the other required information as described in the following steps and click Next to proceed to the next screen where you will provide IPv6 addresses for the nodes.
|
If you want to provide additional NTP servers, click +Add NTP Host again and repeat this substep.
-
Click +Add DNS Provider to add one or more DNS servers.
After you've entered the information, click the checkmark icon to save it.
-
Provide a Proxy Server.
For clusters that do not have direct connectivity to Cisco cloud, we recommend configuring a proxy server to establish the
connectivity. This allows you to mitigate risk from exposure to non-conformant hardware and software in your fabrics.
The proxy server must have the following URLs enabled:
dcappcenter.cisco.com
svc.intersight.com
svc.ucs-connect.com
svc-static1.intersight.com
svc-static1.ucs-connect.com
If you want to skip proxy configuration, mouse over the information (i) icon next to the field, then click Skip.
-
(Optional) If your proxy server required authentication, change Authentication required for Proxy to Yes and provide the login credentials.
-
(Optional) Expand the Advanced Settings category and change the settings if required.
Under advanced settings, you can configure the following:
-
Provide one or more search domains by clicking +Add DNS Search Domain.
After you've entered the information, click the checkmark icon to save it.
-
Provide custom App Network and Service Network.
The application overlay network defines the address space used by the application's services running in the Nexus Dashboard.
The field is pre-populated with the default 172.17.0.1/16 value.
The services network is an internal network used by the Nexus Dashboard and its processes. The field is pre-populated with
the default 100.80.0.0/16 value.
If you have checked the Enable IPv6 option earlier, you can also define the IPv6 subnets for the App and Service networks.
Application and Services networks are described in the Prerequisites and Guidelines section earlier in this document.
-
Click Next to continue.
Note
|
If your node has only an IPv4 management address but you have checked Enabled IPv6 and provided an IPv6 NTP server address, ensure that the NTP address is correct and click Confirm to proceed to the next screen where you will provide the nodes' IPv6 addresses.
|
|
Step 13
|
In the Node Details screen, provide the node's information.
-
Click the Edit button next to the first node.
-
Provide the node's Name.
The Management Network and Data Network information will be already populated from the VPC subnet you have configured before deploying the cluster.
The cluster creates six subnets from the given VPC CIDR, from which the data and management networks will be allocated for
the cluster's three nodes.
-
Leave IPv6 addresses and VLAN fields blank.
Cloud Nexus Dashboard clusters do not support these options.
-
Click Save to save the changes.
|
Step 14
|
Click Add Node to add the second node to the cluster.
The Node Details window opens.
-
Provide the node's Name.
-
In the Credentials section, provide the node's Public IP Address and the password you provided during template deployment, then click Verify.
The IP address and password are used to pull that node's Management Network and Data Network information, which will be populated in the fields below.
-
Click Save to save the changes.
|
Step 15
|
Repeat the previous step to add the 3rd node.
|
Step 16
|
In the Node Details page, click Next to continue.
|
Step 17
|
In the Confirmation screen, review and verify the configuration information and click Configure to create the cluster.
During the node bootstrap and cluster bring-up, the overall progress as well as each node's individual progress will be displayed
in the UI. If you do not see the bootstrap progress advance, manually refresh the page in your browser to update the status.
It may take up to 30 minutes for the cluster to form and all the services to start. When cluster configuration is complete,
the page will reload to the Nexus Dashboard GUI.
|
Step 18
|
Verify that the cluster is healthy.
It may take up to 30 minutes for the cluster to form and all the services to start.
After all three nodes are ready, you can log in to any one node via SSH as the rescue-user using the password you provided during node deployment and run the following command to verify cluster health:
-
Verify that the cluster is up and running.
You can check the current status of cluster deployment by logging in to any of the nodes and running the acs health command.
While the cluster is converging, you may see the following outputs: $ acs health
k8s install is in-progress
$ acs health
k8s services not in desired state - [...]
$ acs health
k8s: Etcd cluster is not ready When the cluster is up and running, the following output will be displayed:$ acs health
All components are healthy
-
Log in to the Nexus Dashboard GUI.
After the cluster becomes available, you can access it by browsing to any one of your nodes' management IP addresses. The
default password for the admin user is the same as the rescue-user password you chose for the first node of the Nexus Dashboard cluster.
|
Step 19
|
Update the nodes' security group with required ports.
This step describes how to update the Nexus Dashboard nodes' instances with the required port configuration for on-boarding
Cisco NDFC sites. If you do not plan on on-boarding any NDFC sites to your Nexus Dashboard cluster, you can skip this step.
Navigate to one of the nodes' data interface:
-
In the AWS console, navigate to Instances.
-
Select one of the Nexus Dashboard instances.
You will make changes to the default security group, so you only need to select one of the nodes.
-
Click the data interface (eth1 ).
-
Click the Interface ID.
The Network Interface properties page opens.
-
In the Network Interface page, click default in the Security groups column of the interface.
-
In the default security group's page, select the Inbound rules tab.
-
Click Edit inbound rules.
-
In the Edit inbound rules page, click Add rule to add a new inbound security rule, then provide the details to allow inbound communication on port 443.
Provide the following information for the new rule:
-
For Type, select Custom TCP .
-
For Port range, enter 443 .
-
For Source, provide the IP addresses of the NDFC controllers which you plan to onboard to your Nexus Dashboard.
-
Still in the Edit inbound rules page, click Add rule to add another inbound security rule, then provide the details to allow inbound communication on port 9092.
Provide the following information for the new rule:
-
For Type, select Custom TCP .
-
For Port range, enter 9092 .
-
For Source, provide the IP addresses of the NDFC controllers which you plan to onboard to your Nexus Dashboard.
|