This is a quickstart guide for getting ASTERIX running in a distributed environment. This guide also introduces the ASTERIX installer (nicknamed Managix) and describes how it can be used to create/manage an ASTERIX instance. By following the simple steps described in this guide, you will get a running instance of ASTERIX. You shall be able to use ASTERIX from its Web interface and manage its lifecycle using Managix. This document assumes that you are running some version of Linux or MacOS X.
Prerequisite:
To know the version of Java installed on your system, execute the following:
$ java -version
If you have version as 1.7.0_x, similar to the output shown below, you are good to proceed.
java version "1.7.0_13" Java(TM) SE Runtime Environment (build 1.7.0_13-b20) Java HotSpot(TM) 64-Bit Server VM (build 23.7-b01, mixed mode)
If you need to upgrade or install java, please follow the instructions below.
For Linux: JDK 7 Linux Install JDK would be installed at a path under /usr/lib/jvm/jdk-version .
For Mac: JDK 7 Mac Install JDK would be installed at /Library/Java/JavaVirtualMachines/jdk-version/Contents/Home .
The java installation directory is referred as JAVA_HOME. Since we upgraded/installed Java, we need to ensure JAVA_HOME points to the installation directory of JDK 7. Modify your ~/.bash_profile (or ~/.bashrc) and define JAVA_HOME accordingly. After modifying, execute the following:
$ java -version
If the version information you obtain does not show 1.7, you need to update the PATH variable. To do so, execute the following:
$ echo "PATH=$JAVA_HOME/bin:$PATH" >> ~/.bash_profile (or ~/.bashrc) $ source ~/.bash_profile (or ~/.bashrc)
We also need to ensure that $JAVA_HOME/bin is in the PATH. $JAVA_HOME/bin should be included in the PATH value. We need to change the if $JAVA_HOME/bin is already in the PATH, we shall simply execute the following:
$ java
If you get the following message, you need to alter the PATH variable in your ~/.bash_profile or ~/.bashrc (whichever you use).
-bash: java: command not found
We assume a user Joe with a home directory as /home/joe. Please note that on Mac, the home directory for user Joe would be /Users/joe.
Ensure that JAVA_HOME variable is defined and points to the the java installation directory on your machine. To verify, execute the following.
$ echo $JAVA_HOME
If you do not see any output, JAVA_HOME is not defined. We need to add the following line to your profile located at /home/joe/.bash_profile or /home/joe/.bashrc, whichever you are using. If you do not any of these files, create a ~/.bash_profile.
export JAVA_HOME=<Path to Java installation directory>
After you have edited ~/.bash_profile (or ~/.bashrc), execute the following to make the changes effective in current shell.
$ source /home/joe/.bash_profile (or /home/joe/.bashrc)
Before proceeding, verify that JAVA_HOME is defined by executing the following.
$ echo $JAVA_HOME
If SSH is not enabled on your system, please follow the instruction below to enable/install it or else skip to the section Configuring Password-less SSH.
The Apple Mac OS X operating system has SSH installed by default but the SSH daemon is not enabled. This means you can’t login remotely or do remote copies until you enable it. To enable it, go to ‘System Preferences’. Under ‘Internet & Networking’ there is a ‘Sharing’ icon. Run that. In the list that appears, check the ‘Remote Login’ option. Also check the "All users" radio button for "Allow access for". This starts the SSH daemon immediately and you can remotely login using your username. The ‘Sharing’ window shows at the bottom the name and IP address to use. You can also find this out using ‘whoami’ and ‘ifconfig’ from the Terminal application.
sudo apt-get install openssh-server
Assumming that you have enabled SSH on your system, let us proceed.
For our single-machine setup of ASTERIX, we need to configure password-less SSH access to localhost. We assume that you are on the machine where you want to install ASTERIX. To verify if you already have password-less SSH configured, execute the following.
$ ssh 127.0.0.1
If you get an output similar to one shown below, type "yes" and press enter.
The authenticity of host '127.0.0.1 (127.0.0.1)' can't be established. RSA key fingerprint is aa:7b:51:90:74:39:c4:f6:28:a2:9d:47:c2:8d:33:31. Are you sure you want to continue connecting (yes/no)?
If you are not prompted for a password, that is if you get an output similar to one shown below, skip to the next section Configuring Managix.
$ ssh 127.0.0.1 Last login: Sat Mar 23 22:52:49 2013
You are here because you were prompted for a password. You need to configure password less SSH. Follow the instructions below.
$ ssh-keygen -t rsa -P "" Generating public/private rsa key pair. Enter file in which to save the key (/home/joe/.ssh/id_rsa): [We shall use the default value, so simply press enter]
If a key already exists, you should get an output similar to what is shown below. Press 'y' to overwrite the existing key.
/home/joe/.ssh/id_rsa already exists. Overwrite (y/n)?
You should see an output similar to one shown below.
The key fingerprint is: 4d:b0:30:14:45:cc:99:86:15:48:17:0b:39:a0:05:ca joe@joe-machine The key's randomart image is: +--[ RSA 2048]----+ | ..o+B@O= | |.. o ==*+ | |.E. oo . | | o | | S . | | | | | | | | | +-----------------+
Note: for Linux users, you may not get an image representation of the key, but this is not an error. Next, execute the following:
$ cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
We shall now retry SSH without password.
$ ssh 127.0.0.1
You may see an output similar to one shown below.
The authenticity of host '127.0.0.1 (127.0.0.1)' can't be established. RSA key fingerprint is aa:7b:51:90:74:39:c4:f6:28:a2:9d:47:c2:8d:33:31. Are you sure you want to continue connecting (yes/no)?
Type 'yes' and press the enter key. You should see an output similar to one shown below.
Warning: Permanently added '127.0.0.1' (RSA) to the list of known hosts. Last login: Thu Mar 28 12:27:10 2013
You should now be able to log in without being prompted for a password or a response.
ssh 127.0.0.1 Last login: Sat Mar 23 22:54:40 2013
Execute 'exit' to close the session.
$ exit logout Connection to 127.0.0.1 closed.
You will need the ASTERIX installer (a.k.a Managix). Download Managix from here; this includes the bits for Managix as well as ASTERIX.
Unzip the Managix zip bundle to an appropriate location. You may create a sub-directory: asterix-mgmt (short for asterix-management) under your home directory. We shall refer to this location as MANAGIX_HOME.
$ cd ~ /home/joe> $ mkdir asterix-mgmt /home/joe> $ cd asterix-mgmt /home/joe/asterix-mgmt> $ unzip <path to the Managix zip bundle> /home/joe/asterix-mgmt> $ export MANAGIX_HOME=`pwd` /home/joe/asterix-mgmt> $ export PATH=$PATH:$MANAGIX_HOME/bin
It is recommended that you add $MANAGIX_HOME/bin to your PATH variable in your bash profile . This can be done by executing the following.
currentDir=`pwd` echo "export MANAGIX_HOME=$currentDir" >> ~/.bash_profile echo "export PATH=$PATH:$MANAGIX_HOME/bin" >> ~/.bash_profile
Above, use ~/.bashrc instead of ~/.bash_profile if you are using ~/.bashrc .
To be able to create an ASTERIX instance and manage its lifecycle, the Managix requires you to configure a set of configuration files namely:
conf/managix-conf.xml
: A configuration XML file that contains configuration settings for Managix.$MANAGIX_HOME/clusters/local/local.xml
.Since we intend to run ASTERIX on a single node, Managix can auto-configure itself and populate the above mentioned configuration files. To auto-configure Managix, execute the following in the MANAGIX_HOME directory:
/home/joe/asterix-mgmt> $ managix configure
Let us do a sample run to validate the set of configuration files auto-generated by Managix.
/home/joe/asterix-mgmt> $ managix validate INFO: Environment [OK] INFO: Managix Configuration [OK] /home/joe/asterix-mgmt> $ managix validate -c $MANAGIX_HOME/clusters/local/local.xml INFO: Environment [OK] INFO: Cluster configuration [OK]
Now that we have configured Managix, we shall next create an ASTERIX instance. An ASTERIX instance is identified by a unique name and is created using the create
command. The usage description for the create
command can be obtained by executing the following.
$ managix help -cmd create Creates an ASTERIX instance with a specified name. Post creation, the instance is in ACTIVE state, indicating its availability for executing statements/queries. Usage arguments/options: -n Name of the ASTERIX instance. -c Path to the cluster configuration file
We shall now use the create command to create an ASTERIX instance by the name "my_asterix". In doing so, we shall use the cluster configuration file that was auto-generated by managix.
$ managix create -n my_asterix -c $MANAGIX_HOME/clusters/local/local.xml
A sample output of the above command is shown below:
INFO: Name:my_asterix Created:Thu Mar 07 11:14:13 PST 2013 Web-Url:http://127.0.0.1:19001 State:ACTIVE
The third line above shows the web-url http://127.0.0.1:19001 for ASTERIX's web-interface. The ASTERIX instance is in the 'ACTIVE' state indicating that you may access the web-interface by navigating to the web-url.
Type in the following "Hello World" query in the box:
let $message := 'Hello World!' return $message
Press the "Execute" button. If the query result shows on the output box, then Congratulations! You have successfully created an ASTERIX instance!
We assume that you have successfully completed the single-machine ASTERIX installation by following the instructions above in section [ASTERIX installation](#Section_1:_Single-Machine_ASTERIX_installation Single Machine). In this section, we shall cover advanced topics related to ASTERIX configuration. Before we proceed, it is imperative to go through some preliminary concepts related to ASTERIX runtime.
An ASTERIX runtime comprises of a ''master node'' and a set of ''worker nodes'', each identified by a unique id. The master node runs a ''Cluster Controller'' service (a.k.a. ''CC''), while each worker node runs a ''Node Controller'' service (a.k.a. ''NC''). Please note that a node in an ASTERIX cluster is a logical concept in the sense that multiple nodes may map to a single physical machine, which is the case for a single-machine ASTERIX installation. This association or mapping between an ASTERIX node and a physical machine is captured in a cluster configuration XML file. In addition, the XML file contains properties and parameters associated with each node.
As observed earlier, Managix can auto-configure itself for a single-machine setup. As part of auto-configuration, Managix generated the cluster XML file. Let us understand the components of the generated cluster XML file. If you have configured Managix (via the "configure" command), you can find a similar cluster XML file as $MANAGIX_HOME/clusters/local/local.xml. The following is a sample XML file generated on a Ubuntu (Linux) setup:
<?xml version="1.0" encoding="UTF-8" standalone="yes"?> <cluster xmlns="cluster"> <name>local</name> <java_home>/usr/lib/jvm/jdk1.7.0</java_home> <java_opts>-Xmx1048m</java_opts> <logdir>/home/joe/asterix-mgmt/clusters/local/working_dir/logs</logdir> <iodevices>/home/joe/asterix-mgmt/clusters/local/working_dir</iodevices> <store>storage</store> <workingDir> <dir>/home/joe/asterix-mgmt/clusters/local/working_dir</dir> <NFS>true</NFS> </workingDir> <master-node> <id>master</id> <client-ip>127.0.0.1</client-ip> <cluster-ip>127.0.0.1</cluster-ip> </master-node> <node> <id>node1</id> <cluster-ip>127.0.0.1</cluster-ip> </node> </cluster>
We shall next explain the components of the cluster configuration XML file.
The single-machine ASTERIX instance configuration that is auto-generated by Managix (using the "configure" command) involves a master node (CC) and a worker node (NC). Each node is assigned a unique id and provided with an ip address (called ''cluster-ip'') that maps a node to a physical machine. The following snippet from the above XML file captures the master/worker nodes in our ASTERIX installation.
<master-node> <id>master</id> <client-ip>127.0.0.1</client-ip> <cluster-ip>127.0.0.1</cluster-ip> </master-node> <node> <id>node1</id> <cluster-ip>127.0.0.1</cluster-ip> </node>
The following is a description of the different elements in the cluster configuration xml file.
The following is a list of properties associated with each worker node in an ASTERIX configuration.
All the above properties can be defined at the global level or a local level. In the former case, these properties apply to all the nodes in an ASTERIX configuration. In the latter case, these properties apply only to the node(s) under which they are defined. A property defined at the local level overrides the definition at the global level.
Next we explain the following setting in the file $MANAGIX_HOME/clusters/local/local.xml.
<workingDir> <dir>/Users/joe/asterix-mgmt/clusters/local/working_dir</dir> <NFS>true</NFS> </workingDir>
Managix associates a working directory with an ASTERIX instance and uses this directory for transferring binaries to each node. If there exists a directory that is readable by each node, Managix can use it to place binaries that can be accessed and used by all the nodes in the ASTERIX set up. A network file system (NFS) provides such a functionality for a cluster of physical machines such that a path on NFS is accessible from each machine in the cluster. In the single-machine set up described above, all nodes correspond to a single physical machine. Each path on the local file system is accessible to all the nodes in the ASTERIX setup and the boolean value for NFS above is thus set to true
.
Managix allows creation and management of multiple ASTERIX instances and uses Zookeeper as its back-end database to keep track of information related to each instance. We need to provide a set of one or more hosts that Managix can use to run a Zookeeper instance. Zookeeper runs as a daemon process on each of the specified hosts. At each host, Zookeeper stores data under the Zookeeper home directory specified as part of the configuration. The following is an example configuration $MANAGIX_HOME/conf/managix-conf.xml
that has Zookeeper running on the localhost (127.0.0.1) :
<?xml version="1.0" encoding="UTF-8" standalone="yes"?> <configuration xmlns="installer"> <zookeeper> <homeDir>/home/joe/asterix/.installer/zookeeper</homeDir> <clientPort>2900</clientPort> <servers> <server>127.0.0.1</server> </servers> </zookeeper> </configuration>
It is possible to have a single host for Zookeeper. A larger number of hosts would use Zookeeper's replication and fault-tolerance feature such that a failure of a host running Zookeeper would not result in loss of information about existing ASTERIX instances.
We assume that you have read the two sections above on single-machine ASTERIX setup. Next we explain how to install ASTERIX in a cluster of multiple machines. As an example, we assume we want to setup ASTERIX on a cluster of three machines, in which we use one machine (called machine A) as the master node and two other machines (called machine B and machine C) as the worker nodes, as shown in the following diagram:
Notice that each machine has a ''cluster-ip'' address, which is used by these machines for their intra-cluster communication. Meanwhile, the master machine also has a ''client-ip'' address, using which an end-user outside the cluster can communicate with this machine. The reason we differentiate between these two types of IP addresses is that we can have a cluster of machines using a private network. In this case they have internal ip addresses that cannot be used outside the network. In the case all the machines are on a public network, the "client-ip" and "cluster-ip" of the master machine can share the same address.
Next we describe how to set up ASTERIX in this cluster, assuming no Managix has been installed on these machines.
We first log into the master machine as the user "joe". On this machine, download Managix from here (save as above), then do the following steps similar to the single-machine case described above:
machineA> cd ~ machineA> mkdir asterix-mgmt machineA> cd asterix-mgmt machineA> unzip <path to the Managix zip bundle> machineA> export MANAGIX_HOME=`pwd` machineA> export PATH=$PATH:$MANAGIX_HOME/bin
We also need an ASTERIX configuration XML file for the cluster. We give the name to the cluster, say, "rainbow". We create a folder for the configuration of this cluster:
machineA> mkdir $MANAGIX_HOME/rainbow_cluster
For this cluster we create a configuration file $MANAGIX_HOME/rainbow_cluster/rainbow.xml
. The following is a sample file with explanation of the properties:
<cluster xmlns="cluster"> <!-- Name of the cluster --> <name>rainbow</name> <!-- username, which should be valid for all the three machines --> <username>joe</username> <!-- The working directory of Managix. It should be on a network file system (NFS) that can accessed by all the machine. Need to create it before running Managix. --> <workingDir> <dir>/home/joe/managix-workingDir</dir> <NFS>true</NFS> </workingDir> <!-- Directory for Asterix to store log information for each machine. Needs to be a local file system. Needs to create it before running Managix. --> <logdir>/mnt/joe/logs</logdir> <!-- Directory used by each worker node to store data files. Needs to be a local file system. Needs to create it before running Managix. --> <iodevices>/mnt/joe</iodevices> <store>storage</store> <!-- Java home for each machine with its JVM options --> <java_home>/usr/lib/jvm/jdk1.7.0</java_home> <java_opts>-Xmx1024m</java_opts> <!-- IP addresses of the master machine A --> <master-node> <id>master</id> <client-ip>128.195.52.177</client-ip> <cluster-ip>192.168.100.0</cluster-ip> </master-node> <!-- IP address(es) of machine B --> <node> <id>nodeB</id> <cluster-ip>192.168.100.1</cluster-ip> </node> <!-- IP address(es) of machine C --> <node> <id>nodeC</id> <cluster-ip>192.168.100.2</cluster-ip> </node> </cluster>
As stated before, each of the above properties can be defined at the cluster level, in which case it applies to all the nodes in the system. Each property can also be defined at a node level.
Once we have formed the cluster XML file, we can validate the configuration by doing the following:
managix validate -c $MANAGIX_HOME/rainbow_cluster/rainbow.xml
If the return message says "OK", it means that the XML configuration file is set properly.
The next steps of setting up SSH are similar to those in the single-machine setup case. We assume we have a common user account called "joe" on each machine in the cluster.
On the master machine, do the following:
machineA> ssh localhost
If you are prompted for a password, execute the following
machineA> ssh-keygen -t rsa -P "" machineA> cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
If $HOME is not on the NFS, copy the id_rsa.pub to the directory ~/.ssh (login with the same account) on each machine, and then do the following on each machine. (Notice that this step is not needed if the folder ".ssh" is on the NFS and can be accessed by all the nodes.)
cd ~/.ssh cat id_rsa.pub >> authorized_keys
Then run the following step again and type "Yes" if promoted:
machineA> ssh localhost
Managix is using a configuration XML file at $MANAGIX_HOME/conf/managix-conf.xml
to configure its own properties, such as its Zookeeper service. We can use the configure
command to auto-generate this configuration file:
machineA> managix configure
We use the validate command to validate managix configuration. To do so, execute the following.
machineA> managix validate INFO: Environment [OK] INFO: Managix Configuration [OK]
Note that the configure
command also generates a cluster configuration XML file at $MANAGIX_HOME/conf/clusters/local.xml. This file is not needed in the case of a cluster of machines.
Now that we have configured Managix, we shall next create an ASTERIX instance. An ASTERIX instance is identified by a unique name and is created using the create command. The usage description for the create command can be obtained by executing the following:
machineA> managix help -cmd create Creates an ASTERIX instance with a specified name. Post creation, the instance is in ACTIVE state, indicating its availability for executing statements/queries. Usage arguments/options: -n Name of the ASTERIX instance. -c Path to the cluster configuration file
We shall now use the create
command to create an ASTERIX instance called "rainbow_asterix". In doing so, we shall use the cluster configuration file that was auto-generated by Managix.
machineA> managix create -n rainbow_asterix -c $MANAGIX_HOME/clusters/rainbow/rainbow.xml
If the response message does not have warning, then Congratulations! You have successfully installed Asterix on this cluster of machines!
Please refer to the section Managing the Lifecycle of an ASTERIX Instance for a detailed description on the set of available commands/operations that let you manage the lifecycle of an ASTERIX instance. Note that the output of the commands varies with the cluster definition and may not apply to the cluster specification you built above.
Now that we have an ASTERIX instance running, let us use Managix to manage the instance's lifecycle. Managix provides the following set of commands/operations:
You may obtain the above listing by simply executing 'managix' :
$ managix
We already talked about create and validate commands. We shall next explain the rest of the commands listed above. We also provide sample output messages of these commands assuming we are running an ASTERIX instance on a single machine.
The describe
command provides information about an ASTERIX instance. The usage can be looked up by executing the following:
$ managix help -cmd describe Provides information about an ASTERIX instance. The following options are available: [-n] Name of the ASTERIX instance. [-admin] Provides a detailed description
The brackets indicate optional flags.
The output of the describe
command when used without the admin
flag contains minimal information and is similar to the output of the create command. Let us try running the describe command in "admin" mode.
$ managix describe -n my_asterix -admin INFO: Name:my_asterix Created:Thu Mar 07 19:07:00 PST 2013 Web-Url:http://127.0.0.1:19001 State:ACTIVE Master node:master:127.0.0.1 node1:127.0.0.1 Asterix version:0.0.5 Asterix Configuration output_dir = /tmp/asterix_output/ Metadata Node:node1 Processes NC at 127.0.0.1 [ 22195 ] CC at 127.0.0.1 [ 22161 ]
As seen above, the instance 'my_asterix' is configured such that all processes running at the localhost (127.0.0.1). The process id for each process (JVM) is shown next to it.
The stop
command can be used for shutting down an ASTERIX instance. After that, the instance is unavailable for executing queries. The usage can be looked up by executing the following:
$ managix help -cmd stop Shuts an ASTERIX instance that is in ACTIVE state. After executing the stop command, the ASTERIX instance transits to the INACTIVE state, indicating that it is no longer available for executing queries. Available arguments/options -n name of the ASTERIX instance.
To stop the ASTERIX instance.
$ managix stop -n my_asterix INFO: Stopped Asterix instance: my_asterix $ managix describe -n my_asterix INFO: Name: my_asterix Created:Thu Mar 07 19:07:00 PST 2013 Web-Url:http://127.0.0.1:19001 State:INACTIVE (Fri Mar 08 09:49:00 PST 2013)
The start
command starts an ASTERIX instance that is in the INACTIVE state. The usage can be looked up by executing the following:
$ managix help -cmd start Starts an ASTERIX instance that is in INACTIVE state. After executing the start command, the ASTERIX instance transits to the ACTIVE state, indicating that it is now available for executing statements/queries. Available arguments/options -n name of the ASTERIX instance.
Let us now start the ASTERIX instance.
$ managix start -n my_asterix INFO: Name:my_asterix Created:Thu Mar 07 19:07:00 PST 2013 Web-Url:http://127.0.0.1:19001 State:ACTIVE (Fri Mar 08 09:49:00 PST 2013)
In an undesirable event of data loss either due to a disk/system failure or accidental execution of a DDL statement (drop dataverse/dataset), you may need to recover the lost data. The backup command allows you to take a backup of the data stored with an ASTERIX instance. The backup can be taken on the local file system or on an HDFS instance. In either case, the snapshots are stored under a backup directory. You need to make sure the backup directory has appropriate read/write permissions. Configuring settings for backup can be found inside the Managix's configuration file located at $MANAGIX_HOME/conf/managix-conf.xml
.
Configuring backup on the local file system
We need to provide path to a backup directory on the local file system. The backup directory can be configured be editing the Managix configuration XML, found at $MANAGIX_HOME/conf/managix-conf.xml
.
<backup> <backupDir>Provide path to the backup directory here</backupDir> </backup>
Prior to taking a backup of an ASTERIX instance, it is required for the instance to be in the INACTIVE state. We do so by using the stop
command, as shown below:
$ managix stop -n my_asterix INFO: Stopped Asterix instance: my_asterix
We can now take the backup by executing the following:
$ managix backup -n my_asterix INFO: my_asterix backed up 0_Fri Mar 08 16:16:34 PST 2013 (LOCAL)
Configuring backup on an HDFS instance
To configure a backups to be taken on an HDFS instance, we need to provide required information about the running HDFS instance. This information includes the HDFS version and the HDFS url. Simply edit the Managix configuration file and provide the required information.
<backup> <backupDir>Provide path to the backup directory here</backupDir> <hdfs> <version>0.20.2</version> <url></url> </hdfs> </backup>
A sample output when a backup is taken on an HDFS is shown below:
$ managix backup -n my_asterix INFO: my_asterix backed up 1_Fri Mar 08 17:10:38 PST 2013 (HDFS)
Each time we take a backup, we are provided with a unique id (a monotonically increasing value starting with 0). This id is required when we need to restore from a previously taken backup. Information about all available backup snapshots can be obtained by using the describe
command in the admin mode, as shown below:
$ managix describe -n my_asterix -admin INFO: Name:my_asterix Created:Fri Mar 08 15:11:12 PST 2013 Web-Url:http://127.0.0.1:19001 State:INACTIVE (Fri Mar 08 16:14:20 PST 2013) Master node:master:127.0.0.1 node1:127.0.0.1 Backup:0 created at Fri Mar 08 16:16:34 PST 2013 (LOCAL) Backup:1 created at Fri Mar 08 17:10:38 PST 2013 (HDFS) Asterix version:0.0.5 Asterix Configuration Metadata Node:node1 Processes
The above output shows the available backup identified by it's id (0). We shall next describe the method for restoring an ASTERIX instance from a backup snapshot.
The restore
command allows you to restore an ASTERIX instance's data from a previously taken backup. The usage description can be obtained as follows:
$ managix help -cmd restore Restores an ASTERIX instance's data from a previously taken backup. Available arguments/options -n name of the ASTERIX instance -b id of the backup snapshot
The following command restores our ASTERIX instance from the backup snapshot identified by the id (0). Prior to restoring an instance from a backup, it is required that the instance is in the INACTIVE state.
$ managix restore -n my_asterix -b 0 INFO: Asterix instance: my_asterix has been restored from backup
You can start the ASTERIX instance by using the start command.
As the name suggests, the delete
command permanently removes an ASTERIX instance by cleaning up all associated data/artifacts. The usage can be looked up by executing the following:
$ managix help -cmd delete Permanently deletes an ASTERIX instance. The instance must be in the INACTIVE state. Available arguments/options -n name of the ASTERIX instance. $ managix delete -n my_asterix INFO: Asterix instance my_asterix deleted.
Managix uses Zookeeper service for storing all information about created ASTERIX instances. The Zookeeper service runs in the background and can be shut down using the shutdown
command.
$ managix shutdown
The help
command provides a usage description of a Managix command.
$ managix help -cmd <command name>
As an example, for looking up the help for the configure
command, execute the following
$ managix help -cmd configure Auto-generates the ASTERIX installer configruation settings and ASTERIX cluster configuration settings for a single node setup.
Question What is meant by the "UNUSABLE" state in the lifecycle of an ASTERIX instance ?
Answer When Managix fails to start a required process (CC/NC), the instance transits to an UNUSABLE state. The reason for the failure needs to be looked up in the logs. Before we attempt to start the instance again, any processes that got launched as part of failed attempt must be stopped. No other operation except "stop" is supported in the UNUSABLE state.
Get rid of the started processes:-
$MANAGIX_HOME/bin/managix stop -n my_asterix
Any processes associated with the instance are killed and the instance moves to the INACTIVE state. You may now delete the instance by executing the following
$MANAGIX_HOME/bin/managix delete -n <name of your ASTERIX instance>
Note that above would remove all traces of the instance including the logs and thus the reason for the failed attempt.
OR
make a subsequent attempt to start the instance if you realized a mistake in the cluster configuration XML and have corrected it. To start the instance, we execute the following.
$MANAGIX_HOME/bin/managix start -n <name of your ASTERIX instance>