Fix spaces in docs.
Change-Id: I22778fd5f89353850df775f60ac02c5e5d071686
Reviewed-on: https://asterix-gerrit.ics.uci.edu/1649
Sonar-Qube: Jenkins <jenkins@fulliautomatix.ics.uci.edu>
Tested-by: Jenkins <jenkins@fulliautomatix.ics.uci.edu>
BAD: Jenkins <jenkins@fulliautomatix.ics.uci.edu>
Integration-Tests: Jenkins <jenkins@fulliautomatix.ics.uci.edu>
Reviewed-by: Till Westmann <tillw@apache.org>
diff --git a/asterixdb/asterix-doc/src/main/installation/ansible.md b/asterixdb/asterix-doc/src/main/installation/ansible.md
index 056871f..5b2a6a5 100644
--- a/asterixdb/asterix-doc/src/main/installation/ansible.md
+++ b/asterixdb/asterix-doc/src/main/installation/ansible.md
@@ -29,36 +29,36 @@
CentOS
- $ sudo yum install python-pip
+ $ sudo yum install python-pip
Ubuntu
- $ sudo apt-get install python-pip
+ $ sudo apt-get install python-pip
macOS
- $ brew install pip
+ $ brew install pip
* Install Ansible, boto, and boto3 on your client machine:
- $ pip install ansible
- $ pip install boto
- $ pip install boto3
+ $ pip install ansible
+ $ pip install boto
+ $ pip install boto3
Note that you might need `sudo` depending on your system configuration.
**Make sure that the version of Ansible is no less than 2.2.1.0**:
- $ ansible --version
- ansible 2.2.1.0
+ $ ansible --version
+ ansible 2.2.1.0
* Download the AsterixDB distribution package, unzip it, and navigate to `opt/ansible/`
- $ cd opt/ansible
+ $ cd opt/ansible
The following files and directories are in the directory `opt/ansible`:
- README bin conf yaml
+ README bin conf yaml
`bin` contains scripts that deploy, start, stop and erase a multi-node AsterixDB cluster, according to
the configuration specified in files under `conf`, and `yaml` contains internal Ansible scripts that the shell
@@ -73,42 +73,33 @@
The following example configures a cluster with two slave nodes (172.0.1.11 and 172.0.1.12) and
one master node (172.0.1.10).
- [cc]
- 172.0.1.10
+ [cc]
+ 172.0.1.10
- [ncs]
- 172.0.1.11
- 172.0.1.12
+ [ncs]
+ 172.0.1.11
+ 172.0.1.12
**Configure passwordless ssh from your current client that runs the scripts to all nodes listed
in `conf/inventory` as well as `localhost`.**
If the ssh user account for target machines is different from your current username, please uncomment
and edit the following two lines:
- ;[all:vars]
- ;ansible_ssh_user=<fill with your ssh account username>
+ ;[all:vars]
+ ;ansible_ssh_user=<fill with your ssh account username>
If you want to specify advanced Ansible builtin variables, please refer to the
[Ansible documentation](http://docs.ansible.com/ansible/intro_inventory.html).
- * **Remote working directories**. Edit `conf/instance_settings.yml` to change the remote binary directories
- when necessary. By default, the binary directory will be under the home directory (as the value of
- Ansible builtin variable ansible_env.HOME) of the ssh user account on each node.
-
- # The name of the product being used.
- product: asterixdb
-
- # The parent directory for the working directory.
- basedir: "{{ ansible_env.HOME }}"
-
- # The working directory.
- binarydir: "{{ basedir }}/{{ product }}"
+ * **Remote working directories**. Edit `conf/instance_settings.yml` to change the remote binary directory
+ (the variable "binarydir") when necessary. By default, the binary directory will be under the home directory
+ (as the value of Ansible builtin variable ansible_env.HOME) of the ssh user account on each node.
## <a id="lifecycle">Cluster Lifecycle Management</a>
* Deploy the binary to all nodes:
- $ bin/deploy.sh
+ $ bin/deploy.sh
* Every time before starting the AsterixDB cluster, you can edit the instance configuration file
`conf/instance/cc.conf`, except that IP addresses/DNS names are generated and cannot
@@ -116,16 +107,16 @@
* Launch your AsterixDB cluster:
- $ bin/start.sh
+ $ bin/start.sh
Now you can use the multi-node AsterixDB cluster by opening the master node
listed in `conf/inventory` at port `19001` (which can be customized in `conf/instance/cc.conf`)
in your browser.
- * If you want to stop the the multi-node AsterixDB cluster, run the following script:
+ * If you want to stop the the multi-node AsterixDB cluster, run the following script:
- $ bin/stop.sh
+ $ bin/stop.sh
- * If you want to remove the binary on all nodes, run the following script:
+ * If you want to remove the binary on all nodes, run the following script:
- $ bin/erase.sh
+ $ bin/erase.sh
diff --git a/asterixdb/asterix-doc/src/main/installation/aws.md b/asterixdb/asterix-doc/src/main/installation/aws.md
index 8b6602b..920b93c 100644
--- a/asterixdb/asterix-doc/src/main/installation/aws.md
+++ b/asterixdb/asterix-doc/src/main/installation/aws.md
@@ -35,36 +35,36 @@
CentOS
- $ sudo yum install python-pip
+ $ sudo yum install python-pip
Ubuntu
- $ sudo apt-get install python-pip
+ $ sudo apt-get install python-pip
macOS
- $ brew install pip
+ $ brew install pip
* Install Ansible, boto, and boto3 on your client machine:
- $ pip install ansible
- $ pip install boto
- $ pip install boto3
+ $ pip install ansible
+ $ pip install boto
+ $ pip install boto3
Note that you might need `sudo` depending on your system configuration.
**Make sure that the version of Ansible is no less than 2.2.1.0**:
- $ ansible --version
- ansible 2.2.1.0
+ $ ansible --version
+ ansible 2.2.1.0
* Download the AsterixDB distribution package, unzip it, navigate to `opt/aws/`
- $ cd opt/aws
+ $ cd opt/aws
The following files and directories are in the directory `opt/aws`:
- README bin conf yaml
+ README bin conf yaml
`bin` contains scripts that start and terminate an AWS-based cluster instance, according to the configuration
specified in files under `conf`, and `yaml` contains internal Ansible scripts that the shell scripts in `bin` use.
@@ -86,85 +86,79 @@
* Configure your ssh setting by editing `~/.ssh/config` and adding the following entry:
- Host *.amazonaws.com
+ Host *.amazonaws.com
IdentityFile <path_of_private_key>
Note that \<path_of_private_key\> should be replaced by the path to the file that stores the private key for the
key pair that you uploaded to AWS and used in `conf/aws_settings`. For example:
- Host *.amazonaws.com
+ Host *.amazonaws.com
IdentityFile ~/.ssh/id_rsa
### <a id="config">Cluster Configuration</a>
- * **AWS settings**. Edit `conf/instance_settings.yml`. The meaning of each parameter is listed as follows:
+ * **AWS settings**. Edit `conf/instance_settings.yml`. The meaning of each parameter is listed as follows:
- # The OS image id for ec2 instances.
- image: ami-76fa4116
+ # The OS image id for ec2 instances.
+ image: ami-76fa4116
- # The data center region for ec2 instances.
- region: us-west-2
+ # The data center region for ec2 instances.
+ region: us-west-2
- # The tag for each ec2 machine. Use different tags for isolation.
- tag: scale_test
+ # The tag for each ec2 machine. Use different tags for isolation.
+ tag: scale_test
- # The name of a security group that appears in your AWS console.
- group: default
+ # The name of a security group that appears in your AWS console.
+ group: default
- # The name of a key pair that appears in your AWS console.
- keypair: <to be filled>
+ # The name of a key pair that appears in your AWS console.
+ keypair: <to be filled>
- # The AWS access key id for your IAM user.
- access_key_id: <to be filled>
+ # The AWS access key id for your IAM user.
+ access_key_id: <to be filled>
- # The AWS secret key for your IAM user.
- secret_access_key: <to be filled>
+ # The AWS secret key for your IAM user.
+ secret_access_key: <to be filled>
- # The AWS instance type. A full list of available types are listed at:
- # https://aws.amazon.com/ec2/instance-types/
- instance_type: t2.micro
+ # The AWS instance type. A full list of available types are listed at:
+ # https://aws.amazon.com/ec2/instance-types/
+ instance_type: t2.micro
- # The number of ec2 instances that construct a cluster.
- count: 3
+ # The number of ec2 instances that construct a cluster.
+ count: 3
- # The user name.
- user: ec2-user
+ # The user name.
+ user: ec2-user
- # Whether to reuse one slave machine to host the master process.
- cc_on_nc: false
+ # Whether to reuse one slave machine to host the master process.
+ cc_on_nc: false
**As described in [prerequisites](#Prerequisites), the following parameters must be customized:**
- # The tag for each ec2 machine. Use different tags for isolation.
- tag: scale_test
+ # The tag for each ec2 machine. Use different tags for isolation.
+ tag: scale_test
- # The name of a security group that appears in your AWS console.
- group: default
+ # The name of a security group that appears in your AWS console.
+ group: default
- # The name of a key pair that appears in your AWS console.
- keypair: <to be filled>
+ # The name of a key pair that appears in your AWS console.
+ keypair: <to be filled>
- # The AWS access key id for your IAM user.
- access_key_id: <to be filled>
+ # The AWS access key id for your IAM user.
+ access_key_id: <to be filled>
- # The AWS secrety key for your IAM user.
- secret_access_key: <to be filled>
+ # The AWS secrety key for your IAM user.
+ secret_access_key: <to be filled>
- * **Remote working directories**. Edit `conf/instance_settings.yml` to change the instance binary directories
- when necessary. By default, the binary directory will be under the home directory (as the value of
- Ansible builtin variable ansible_env.HOME) of the ssh user account on each node.
-
- # The parent directory for the working directory.
- basedir: "{{ ansible_env.HOME }}"
-
- # The working directory.
- binarydir: "{{ basedir }}/{{ product }}"
+ * **Remote working directories**. Edit `conf/instance_settings.yml` to change the remote binary directory
+ (the variable "binarydir") when necessary. By default, the binary directory will be under the home directory
+ (as the value of Ansible builtin variable ansible_env.HOME) of the ssh user account on each node.
### <a id="lifecycle">Cluster Lifecycle Management</a>
* Allocate AWS EC2 nodes (the number of nodes is specified in `conf/instance_settings.yml`)
and deploy the binary to all allocated EC2 nodes:
- bin/deploy.sh
+ bin/deploy.sh
* Before starting the AsterixDB cluster, you the instance configuration file `conf/instance/cc.conf`
can be modified with the exception of the IP addresses/DNS names which are are generated and cannot
@@ -172,7 +166,7 @@
* Launch your AsterixDB cluster on EC2:
- bin/start.sh
+ bin/start.sh
Now you can use the multi-node AsterixDB cluster on EC2 by by opening the master node
listed in `conf/instance/inventory` at port `19001` (which can be customized in `conf/instance/cc.conf`)
@@ -180,13 +174,13 @@
* If you want to stop the AWS-based AsterixDB cluster, run the following script:
- bin/stop.sh
+ bin/stop.sh
Note that this only stops AsterixDB but does not stop the EC2 nodes.
* If you want to terminate the EC2 nodes that run the AsterixDB cluster, run the following script:
- bin/terminate.sh
+ bin/terminate.sh
**Note that it will destroy everything in the AsterixDB cluster you installed and terminate all EC2 nodes
for the cluster.**