The Vertica Forum recently got a makeover! Let us know what you think by filling out this short, anonymous survey.
Please take this survey to help us learn more about how you use third party tools. Your input is greatly appreciated!
HELP HELP!!!,Database with 1 or 2 nodes cannot be k-safe and it may lose data if it crashes

Answers
@shang, your installation doesn't meet the minimum requirements. Please check https://www.vertica.com/docs/9.2.x/HTML/Content/Authoring/AdministratorsGuide/ConfiguringTheDB/PhysicalSchema/DesigningForK-Safety.htm
@shang would it be possible to share the admintools.conf from /opt/vertica/conf and admintools.log from /opt/vertica/log.
Once we have both the log we can check and provide more information about the create db failure.
@Sankarmn thanks for your replay.
However, I try to install a cluster with three nodes that have the same problem~.
and the install log as bellow~
47.114.142.96,47.114.137.249,47.114.48.164 , this is my three Aliyun ECS server address,
[root@iZbp1hduhk14uh847bf7ncZ ~]# rpm -Uvh vertica-10.0.0-0.x86_64.RHEL6.rpm
Preparing... ########################################### [100%]
1:vertica ########################################### [100%]
Vertica Analytic Database v10.0.0-0 successfully installed on host iZbp1hduhk14uh847bf7ncZ
To complete your NEW installation and configure the cluster, run:
/opt/vertica/sbin/install_vertica
To complete your Vertica UPGRADE, run:
/opt/vertica/sbin/update_vertica
Important
Before upgrading Vertica, you must backup your database. After you restart your
database after upgrading, you cannot revert to a previous Vertica software version.
View the latest Vertica documentation at https://www.vertica.com/documentation/vertica/
[root@iZbp1hduhk14uh847bf7ncZ ~]# /opt/vertica/sbin/install_vertica --hosts 47.114.142.96,47.114.137.249,47.114.48.164 --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --hosts (-s) to addresses...
Error: A cluster exists but does not match the provided --hosts
47.114.48.164 in --hosts but not in cluster.
47.114.137.249 in --hosts but not in cluster.
47.114.142.96 in --hosts but not in cluster.
127.0.0.1 in cluster but not in --hosts.
Hint: omit --hosts for existing clusters. To change a cluster use --add-hosts or --remove-hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
[root@iZbp1hduhk14uh847bf7ncZ ~]# cat /opt/vertica/config/admintools.conf
[Configuration]
format = 3
install_opts = --hosts '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
v_test1_node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Database:test1]
restartpolicy = ksafe
port = 5433
path = /home/dbadmin/test1
nodes = v_test1_node0001
is_eon_mode = False
depot_base_dir = None
depot_size = None
communal_storage_url = None
num_shards = None
is_first_start_after_revive = False
branch_name =
thanks again!
@Sankarmn thanks for your reply.
However,this problem still unsolve when I change to three nodes with a cluster.
and the log as bellow~
note: this 47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx is my three cloud server address,
[root@iZbZ ~]# rpm -Uvh vertica-10.0.0-0.x86_64.RHEL6.rpm
Preparing... ########################################### [100%]
1:vertica ########################################### [100%]
Vertica Analytic Database v10.0.0-0 successfully installed on host iZbp1hduhk14uh847bf7ncZ
To complete your NEW installation and configure the cluster, run:
/opt/vertica/sbin/install_vertica
To complete your Vertica UPGRADE, run:
/opt/vertica/sbin/update_vertica
Important
Before upgrading Vertica, you must backup your database. After you restart your
database after upgrading, you cannot revert to a previous Vertica software version.
View the latest Vertica documentation at https://www.vertica.com/documentation/vertica/
[root@iZbZ ~]# /opt/vertica/sbin/install_vertica --hosts 47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --hosts (-s) to addresses...
Error: A cluster exists but does not match the provided --hosts
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxx.xxx in --hosts but not in cluster.
127.0.0.1 in cluster but not in --hosts.
Hint: omit --hosts for existing clusters. To change a cluster use --add-hosts or --remove-hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
[root@iZbZ ~]# cat /opt/vertica/config/admintools.conf
[Configuration]
format = 3
install_opts = --hosts '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
v_test1_node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Database:test1]
restartpolicy = ksafe
port = 5433
path = /home/dbadmin/test1
nodes = v_test1_node0001
is_eon_mode = False
depot_base_dir = None
depot_size = None
communal_storage_url = None
num_shards = None
is_first_start_after_revive = False
branch_name =
thanks again~
@Nimmi_gupta
appreciate it for your reply
bellow is my admintools.log and Vertica installing log
[dbadmin@ncZ ~]$ admintools -t create_db -x auth_params.conf --communal-storage-location=oss://xianan-bucket --depot-path=/home/dbadmin/depot/ --shard-count=4 -s 47.114.xxx.xxx,47.114.xxx.xxx -d DBTest2
Info: no password specified, using none
Default depot size in use
Database with 1 or 2 nodes cannot be k-safe and it may lose data if it crashes
Error: These nodes/hosts do not appear to be part of this cluster:
47.114.xxx.xxx, 47.114.xxx.xxx
Hint: Valid inputs are IP addresses, node names, or hostnames.
Hostnames must resolve to a single address within the target family.
Hostnames and addresses must be in the IPv4 address family.
[dbadmin@ncZ ~]$ cat /opt/vertica/config/admintools.conf
[Configuration]
format = 3
install_opts = --hosts '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
v_test1_node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Database:test1]
restartpolicy = ksafe
port = 5433
path = /home/dbadmin/test1
nodes = v_test1_node0001
is_eon_mode = False
depot_base_dir = None
depot_size = None
communal_storage_url = None
num_shards = None
is_first_start_after_revive = False
branch_name =
thanks again for your reply~
It looks like you have already created a single-node cluster with "test1" database.
I think this is the quickest solution:
Add new hosts with -A / --add-hosts option (replace -s with -A) then create the new database with admintools -t create_db
Alternately, you may need to remove the test1 database so you can remove the 127.0.0.1 node definition attached to test1 by running "admintools -t drop_db"
Then you can reset the nodes with install_vertica -s --clean where --clean will remove the old node list including 127.0.0.1 and replace with new list from -s option.
thanks for your reply.
now I have already dropped the 'test1' database. However, another error still unsolved and the log as below;
==================console log============================
[root@iZ~]# /opt/vertica/sbin/install_vertica -A 47.114.xxx.xxx
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --add-hosts (-A) to addresses...
Error: Existing single-node localhost (loopback) cluster cannot be expanded
Hint: Move cluster to external address first. See online documentation.
Installation FAILED with errors.
Installation stopped before any changes were made.
========admintools.conf=============
[root@iZb ~]# cat /opt/vertica/config/admintools.conf
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Configuration]
format = 3
install_opts = -s '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
thanks for your help again~
Hi @Bryan_H
thanks for your reply.
I have already deleted the database, but I want to know what's mean "nstall_vertica -s --clean where --clean will remove the old node list including 127.0.0.1 and replace with new list from -s option." cause I got another error when I operator as you say. And console output and log as below.
[root@iZbp ~]# /opt/vertica/sbin/install_vertica --clean
Vertica Analytic Database 10.0.0-0 Installation Tool
Error: No machines will be included in the cluster!
Hint: provide --hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
[root@iZbZ ~]# /opt/vertica/sbin/install_vertica -s 47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --hosts (-s) to addresses...
Error: A cluster exists but does not match the provided --hosts
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxxx.xxx in --hosts but not in cluster.
127.0.0.1 in cluster but not in --hosts.
Hint: omit --hosts for existing clusters. To change a cluster use --add-hosts or --remove-hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
Thanks for your support again and let me know if you need additional information.
You need to specify -s parameter if you are using --clean parameter
If this is a cloud/virtualized environment, try this:
If this is not a cloud/virtualized environment, remove the --point-to-point flag.
thanks for your reply. However, I got this error too as below.
(note this operator all in cloud server)
[root@iZcZ ~]# /opt/vertica/sbin/install_vertica -s '47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx' --clean --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm --point-to-point
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --hosts (-s) to addresses...
Error: cannot find which cluster host is the local host.
Hint: Is this node in the cluster? Did its IP address change?
Installation FAILED with errors.
Installation stopped before any changes were made.
[root@iZbZ ~]# cat /opt/vertica/config/admintools.conf
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Configuration]
format = 3
install_opts = -s '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
appreciate it for your help~
It looks like you created a cluster once before with localhost 127.0.0.1.
Try renaming /opt/vertica/config/admintools.conf to something else. Do this on all 3 nodes:
Then, on any one node, try again:
Hi ~
thanks for everyone~
now I have already installed a Vertica cluster with three hosts' success.
However, I got another error when I want to create an EON mode database.
Can Vertica EON mode support other Object Storage Service which not AWS or Auxier?
[dbadmin@iZcZ ~]$ admintools -t create_db -x auth_params.conf -s '172.16.xxx.xxx,172.16.xxx.xxx,172.16.xxx.xxx' -d VMart3 -p 123 --depot-path=/home/dbadmin/depot --shard-count=6 --communal-storage-location=oss://xxxx-bucket -D /home/dbadmin/data/ -c /home/dbadmin/catalog/ --depot-size 5G
Distributing changes to cluster.
Creating database VMart3
Bootstrap on host 172.16.xxx.66 return code -6 stdout '' stderr ''
Error: Bootstrap on host 172.16.xxx.66 return code -6 stdout '' stderr ''
thanks for friends help!
Hi @LenoyJ ~
thanks for your reply, I have already solved Vertica install problems.
unfortunate I got new error now when I try to create an EON mode database like pre message, would appreciate it if you can give me some suggestion about it.
thanks again
Best wish~
I doubt AliCloud's OSS (oss://) would work as a supported communal storage location for Eon yet (as of 10.0). Supported Eon mode locations in the cloud today are AWS & GCP. On-prem Eon mode deployments includes Pure Storage, MinIO & HDFS. Though you could theoretically deploy MinIO and HDFS on the cloud.
Enterprise Mode, on the other hand, does not have a requirement of a communal storage and should work on any commodity/virtualized hardware that meets our minimum requirements. It looks like you are trying out Vertica and deploying a few nodes, maybe try Enterprise instead for now? Like so:
Hi @LenoyJ
Thanks for your reply.
unfortunately, my work is testing AliCloud's OSS can or cannot work as communal storage for EON, I can't change to use Enterprise Mode.
so can you give me some advice about this problem please?
thanks again.
Does Alicloud OSS offer a connection using S3 protocol? If so, then you may be able to connect using equivalent s3:// URL.
Please see Alicloud documentation for S3 compatibility information: https://www.alibabacloud.com/help/doc-detail/64919.html?spm=a2c5t.11065259.1996646101.searchclickresult.2c434c0d8ByRNe
Unfortunately, we don't officially support other protocols.
thanks for your reply.
However, I got an error when I change to S3 as communal storage to create an EON mode database.
the log as below.
appreciate it if you can give me some suggestion about this error.
[dbadmin@iZbp1hduhk14uh847bf7ncZ ~]$ admintools -t create_db -x auth_params.conf -s '172.16.xxx.a,172.16.xxx.b,172.16.xxx.c' -d VMart1 -p 123 --depot-path=/home/dbadmin/depot --shard-count=6 --communal-storage-location=s3://vertica-test-1 -D /home/dbadmin/data/ -c /home/dbadmin/catalog/ --depot-size 5G
Distributing changes to cluster.
Creating database VMart1
Starting bootstrap node v_vmart1_node0002 (172.16.xxx.b)
Starting nodes:
v_vmart1_node0002 (172.16.xxx.b)
Starting Vertica on all nodes. Please wait, databases with a large catalog may take a while to initialize.
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (UP)
Creating database nodes
Creating node v_vmart1_node0001 (host 172.16.xxx.a)
Creating node v_vmart1_node0003 (host 172.16.xxx.c)
Generating new configuration information
Stopping single node db before adding additional nodes.
Database shutdown complete
Starting all nodes
Start hosts = ['172.16.xxx.xx', '172.16.xxx.xx', '172.16.xxx.xx']
Starting nodes:
v_vmart1_node0002 (172.16.xxx.xx)
v_vmart1_node0001 (172.16.xxx.xx)
v_vmart1_node0003 (172.16.xxx.xx)
Starting Vertica on all nodes. Please wait, databases with a large catalog may take a while to initialize.
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
It is suggested that you continue waiting.
Do you want to continue waiting? (yes/no) [yes]
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
It is suggested that you continue waiting.
Do you want to continue waiting? (yes/no) [yes]
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
It is suggested that you continue waiting.
Do you want to continue waiting? (yes/no) [yes]
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
It is suggested that you continue waiting.
Do you want to continue waiting? (yes/no) [yes]
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
ERROR: Not all nodes came up, but not all down. Run scrutinize.
Unable to establish vsql script connection: Unable to connect to 'VMart1'
Unable to establish client-server connection: Unable to connect to 'VMart1'
Unable to create depot storage locations (if Eon) without a client-server connection.
Unable to rebalance shards (if Eon) without a client-server connection.
Unable to set K-safety value without a client-server connection.
Unable to install default extension packages without a vsql script connection
Unable to sync database catalog (if Eon) without a client-server connection.
Database creation SQL tasks included one or more failures (see above).
Database VMart1 created successfully, some nodes may have had startup problems.
thanks again for your attention~