@shang would it be possible to share the admintools.conf from /opt/vertica/conf and admintools.log from /opt/vertica/log.
Once we have both the log we can check and provide more information about the create db failure.
@Sankarmn thanks for your replay.
However, I try to install a cluster with three nodes that have the same problem~.
and the install log as bellow~
47.114.142.96,47.114.137.249,47.114.48.164 , this is my three Aliyun ECS server address,
Mapping hostnames in --hosts (-s) to addresses...
Error: A cluster exists but does not match the provided --hosts
47.114.48.164 in --hosts but not in cluster.
47.114.137.249 in --hosts but not in cluster.
47.114.142.96 in --hosts but not in cluster.
127.0.0.1 in cluster but not in --hosts.
Hint: omit --hosts for existing clusters. To change a cluster use --add-hosts or --remove-hosts.
Installation FAILED with errors.
Mapping hostnames in --hosts (-s) to addresses...
Error: A cluster exists but does not match the provided --hosts
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxx.xxx in --hosts but not in cluster.
127.0.0.1 in cluster but not in --hosts.
Hint: omit --hosts for existing clusters. To change a cluster use --add-hosts or --remove-hosts.
Installation FAILED with errors.
@Nimmi_gupta said: @shang would it be possible to share the admintools.conf from /opt/vertica/conf and admintools.log from /opt/vertica/log.
Once we have both the log we can check and provide more information about the create db failure.
@Nimmi_gupta
appreciate it for your reply
bellow is my admintools.log and Vertica installing log
[dbadmin@ncZ ~]$ admintools -t create_db -x auth_params.conf --communal-storage-location=oss://xianan-bucket --depot-path=/home/dbadmin/depot/ --shard-count=4 -s 47.114.xxx.xxx,47.114.xxx.xxx -d DBTest2
Info: no password specified, using none
Default depot size in use
Database with 1 or 2 nodes cannot be k-safe and it may lose data if it crashes
Error: These nodes/hosts do not appear to be part of this cluster:
47.114.xxx.xxx, 47.114.xxx.xxx
Hint: Valid inputs are IP addresses, node names, or hostnames.
Hostnames must resolve to a single address within the target family.
Hostnames and addresses must be in the IPv4 address family.
[dbadmin@ncZ ~]$ cat /opt/vertica/config/admintools.conf
[Configuration]
format = 3
install_opts = --hosts '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
It looks like you have already created a single-node cluster with "test1" database.
I think this is the quickest solution:
Add new hosts with -A / --add-hosts option (replace -s with -A) then create the new database with admintools -t create_db
Alternately, you may need to remove the test1 database so you can remove the 127.0.0.1 node definition attached to test1 by running "admintools -t drop_db"
Then you can reset the nodes with install_vertica -s --clean where --clean will remove the old node list including 127.0.0.1 and replace with new list from -s option.
@Bryan_H said:
It looks like you have already created a single-node cluster with "test1" database.
I think this is the quickest solution:
Add new hosts with -A / --add-hosts option (replace -s with -A) then create the new database with admintools -t create_db
thanks for your reply.
now I have already dropped the 'test1' database. However, another error still unsolved and the log as below;
==================console log============================
[root@iZ~]# /opt/vertica/sbin/install_vertica -A 47.114.xxx.xxx
Vertica Analytic Database 10.0.0-0 Installation Tool
Validating options...
Mapping hostnames in --add-hosts (-A) to addresses...
Error: Existing single-node localhost (loopback) cluster cannot be expanded
Hint: Move cluster to external address first. See online documentation.
Installation FAILED with errors.
Installation stopped before any changes were made.
@Bryan_H said:
Alternately, you may need to remove the test1 database so you can remove the 127.0.0.1 node definition attached to test1 by running "admintools -t drop_db"
Then you can reset the nodes with install_vertica -s --clean where --clean will remove the old node list including 127.0.0.1 and replace with new list from -s option.
Hi @Bryan_H
thanks for your reply.
I have already deleted the database, but I want to know what's mean "nstall_vertica -s --clean where --clean will remove the old node list including 127.0.0.1 and replace with new list from -s option." cause I got another error when I operator as you say. And console output and log as below.
Error: No machines will be included in the cluster!
Hint: provide --hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
[root@iZbZ ~]# /opt/vertica/sbin/install_vertica -s 47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm
Vertica Analytic Database 10.0.0-0 Installation Tool
Validating options...
Mapping hostnames in --hosts (-s) to addresses...
Error: A cluster exists but does not match the provided --hosts
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxxx.xxx in --hosts but not in cluster.
127.0.0.1 in cluster but not in --hosts.
Hint: omit --hosts for existing clusters. To change a cluster use --add-hosts or --remove-hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
Thanks for your support again and let me know if you need additional information.
If this is not a cloud/virtualized environment, remove the --point-to-point flag.
thanks for your reply. However, I got this error too as below.
(note this operator all in cloud server)
[root@iZcZ ~]# /opt/vertica/sbin/install_vertica -s '47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx' --clean --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm --point-to-point
Vertica Analytic Database 10.0.0-0 Installation Tool
Validating options...
Mapping hostnames in --hosts (-s) to addresses...
Error: cannot find which cluster host is the local host.
Hint: Is this node in the cluster? Did its IP address change?
Installation FAILED with errors.
Installation stopped before any changes were made.
It looks like you created a cluster once before with localhost 127.0.0.1.
Try renaming /opt/vertica/config/admintools.conf to something else. Do this on all 3 nodes:
Hi ~
thanks for everyone~
now I have already installed a Vertica cluster with three hosts' success.
However, I got another error when I want to create an EON mode database.
Can Vertica EON mode support other Object Storage Service which not AWS or Auxier?
@LenoyJ said:
It looks like you created a cluster once before with localhost 127.0.0.1.
Try renaming /opt/vertica/config/admintools.conf to something else. Do this on all 3 nodes:
Hi @LenoyJ ~
thanks for your reply, I have already solved Vertica install problems.
unfortunate I got new error now when I try to create an EON mode database like pre message, would appreciate it if you can give me some suggestion about it.
@shang said:
Can Vertica EON mode support other Object Storage Service which not AWS or Auxier?
I doubt AliCloud's OSS (oss://) would work as a supported communal storage location for Eon yet (as of 10.0). Supported Eon mode locations in the cloud today are AWS & GCP. On-prem Eon mode deployments includes Pure Storage, MinIO & HDFS. Though you could theoretically deploy MinIO and HDFS on the cloud.
Enterprise Mode, on the other hand, does not have a requirement of a communal storage and should work on any commodity/virtualized hardware that meets our minimum requirements. It looks like you are trying out Vertica and deploying a few nodes, maybe try Enterprise instead for now? Like so:
@shang said:
Can Vertica EON mode support other Object Storage Service which not AWS or Auxier?
I doubt AliCloud's OSS (oss://) would work as a supported communal storage location for Eon yet (as of 10.0). Supported Eon mode locations in the cloud today are AWS & GCP. On-prem Eon mode deployments includes Pure Storage, MinIO & HDFS. Though you could theoretically deploy MinIO and HDFS on the cloud.
Enterprise Mode, on the other hand, does not have a requirement of a communal storage and should work on any commodity/virtualized hardware that meets our minimum requirements. It looks like you are trying out Vertica and deploying a few nodes, maybe try Enterprise instead for now? Like so:
Hi @LenoyJ
Thanks for your reply.
unfortunately, my work is testing AliCloud's OSS can or cannot work as communal storage for EON, I can't change to use Enterprise Mode.
so can you give me some advice about this problem please?
thanks again.
Answers
@shang, your installation doesn't meet the minimum requirements. Please check https://www.vertica.com/docs/9.2.x/HTML/Content/Authoring/AdministratorsGuide/ConfiguringTheDB/PhysicalSchema/DesigningForK-Safety.htm
@shang would it be possible to share the admintools.conf from /opt/vertica/conf and admintools.log from /opt/vertica/log.
Once we have both the log we can check and provide more information about the create db failure.
@Sankarmn thanks for your replay.
However, I try to install a cluster with three nodes that have the same problem~.
and the install log as bellow~
47.114.142.96,47.114.137.249,47.114.48.164 , this is my three Aliyun ECS server address,
[root@iZbp1hduhk14uh847bf7ncZ ~]# rpm -Uvh vertica-10.0.0-0.x86_64.RHEL6.rpm
Preparing... ########################################### [100%]
1:vertica ########################################### [100%]
Vertica Analytic Database v10.0.0-0 successfully installed on host iZbp1hduhk14uh847bf7ncZ
To complete your NEW installation and configure the cluster, run:
/opt/vertica/sbin/install_vertica
To complete your Vertica UPGRADE, run:
/opt/vertica/sbin/update_vertica
Important
Before upgrading Vertica, you must backup your database. After you restart your
database after upgrading, you cannot revert to a previous Vertica software version.
View the latest Vertica documentation at https://www.vertica.com/documentation/vertica/
[root@iZbp1hduhk14uh847bf7ncZ ~]# /opt/vertica/sbin/install_vertica --hosts 47.114.142.96,47.114.137.249,47.114.48.164 --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --hosts (-s) to addresses...
Error: A cluster exists but does not match the provided --hosts
47.114.48.164 in --hosts but not in cluster.
47.114.137.249 in --hosts but not in cluster.
47.114.142.96 in --hosts but not in cluster.
127.0.0.1 in cluster but not in --hosts.
Hint: omit --hosts for existing clusters. To change a cluster use --add-hosts or --remove-hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
[root@iZbp1hduhk14uh847bf7ncZ ~]# cat /opt/vertica/config/admintools.conf
[Configuration]
format = 3
install_opts = --hosts '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
v_test1_node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Database:test1]
restartpolicy = ksafe
port = 5433
path = /home/dbadmin/test1
nodes = v_test1_node0001
is_eon_mode = False
depot_base_dir = None
depot_size = None
communal_storage_url = None
num_shards = None
is_first_start_after_revive = False
branch_name =
thanks again!
@Sankarmn thanks for your reply.
However,this problem still unsolve when I change to three nodes with a cluster.
and the log as bellow~
note: this 47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx is my three cloud server address,
[root@iZbZ ~]# rpm -Uvh vertica-10.0.0-0.x86_64.RHEL6.rpm
Preparing... ########################################### [100%]
1:vertica ########################################### [100%]
Vertica Analytic Database v10.0.0-0 successfully installed on host iZbp1hduhk14uh847bf7ncZ
To complete your NEW installation and configure the cluster, run:
/opt/vertica/sbin/install_vertica
To complete your Vertica UPGRADE, run:
/opt/vertica/sbin/update_vertica
Important
Before upgrading Vertica, you must backup your database. After you restart your
database after upgrading, you cannot revert to a previous Vertica software version.
View the latest Vertica documentation at https://www.vertica.com/documentation/vertica/
[root@iZbZ ~]# /opt/vertica/sbin/install_vertica --hosts 47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --hosts (-s) to addresses...
Error: A cluster exists but does not match the provided --hosts
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxx.xxx in --hosts but not in cluster.
127.0.0.1 in cluster but not in --hosts.
Hint: omit --hosts for existing clusters. To change a cluster use --add-hosts or --remove-hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
[root@iZbZ ~]# cat /opt/vertica/config/admintools.conf
[Configuration]
format = 3
install_opts = --hosts '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
v_test1_node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Database:test1]
restartpolicy = ksafe
port = 5433
path = /home/dbadmin/test1
nodes = v_test1_node0001
is_eon_mode = False
depot_base_dir = None
depot_size = None
communal_storage_url = None
num_shards = None
is_first_start_after_revive = False
branch_name =
thanks again~
@Nimmi_gupta
appreciate it for your reply
bellow is my admintools.log and Vertica installing log
[dbadmin@ncZ ~]$ admintools -t create_db -x auth_params.conf --communal-storage-location=oss://xianan-bucket --depot-path=/home/dbadmin/depot/ --shard-count=4 -s 47.114.xxx.xxx,47.114.xxx.xxx -d DBTest2
Info: no password specified, using none
Default depot size in use
Database with 1 or 2 nodes cannot be k-safe and it may lose data if it crashes
Error: These nodes/hosts do not appear to be part of this cluster:
47.114.xxx.xxx, 47.114.xxx.xxx
Hint: Valid inputs are IP addresses, node names, or hostnames.
Hostnames must resolve to a single address within the target family.
Hostnames and addresses must be in the IPv4 address family.
[dbadmin@ncZ ~]$ cat /opt/vertica/config/admintools.conf
[Configuration]
format = 3
install_opts = --hosts '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
v_test1_node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Database:test1]
restartpolicy = ksafe
port = 5433
path = /home/dbadmin/test1
nodes = v_test1_node0001
is_eon_mode = False
depot_base_dir = None
depot_size = None
communal_storage_url = None
num_shards = None
is_first_start_after_revive = False
branch_name =
thanks again for your reply~
It looks like you have already created a single-node cluster with "test1" database.
I think this is the quickest solution:
Add new hosts with -A / --add-hosts option (replace -s with -A) then create the new database with admintools -t create_db
Alternately, you may need to remove the test1 database so you can remove the 127.0.0.1 node definition attached to test1 by running "admintools -t drop_db"
Then you can reset the nodes with install_vertica -s --clean where --clean will remove the old node list including 127.0.0.1 and replace with new list from -s option.
thanks for your reply.
now I have already dropped the 'test1' database. However, another error still unsolved and the log as below;
==================console log============================
[root@iZ~]# /opt/vertica/sbin/install_vertica -A 47.114.xxx.xxx
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --add-hosts (-A) to addresses...
Error: Existing single-node localhost (loopback) cluster cannot be expanded
Hint: Move cluster to external address first. See online documentation.
Installation FAILED with errors.
Installation stopped before any changes were made.
========admintools.conf=============
[root@iZb ~]# cat /opt/vertica/config/admintools.conf
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Configuration]
format = 3
install_opts = -s '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
thanks for your help again~
Hi @Bryan_H
thanks for your reply.
I have already deleted the database, but I want to know what's mean "nstall_vertica -s --clean where --clean will remove the old node list including 127.0.0.1 and replace with new list from -s option." cause I got another error when I operator as you say. And console output and log as below.
[root@iZbp ~]# /opt/vertica/sbin/install_vertica --clean
Vertica Analytic Database 10.0.0-0 Installation Tool
Error: No machines will be included in the cluster!
Hint: provide --hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
[root@iZbZ ~]# /opt/vertica/sbin/install_vertica -s 47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --hosts (-s) to addresses...
Error: A cluster exists but does not match the provided --hosts
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxx.xxx in --hosts but not in cluster.
47.114.xxxx.xxx in --hosts but not in cluster.
127.0.0.1 in cluster but not in --hosts.
Hint: omit --hosts for existing clusters. To change a cluster use --add-hosts or --remove-hosts.
Installation FAILED with errors.
Installation stopped before any changes were made.
Thanks for your support again and let me know if you need additional information.
You need to specify -s parameter if you are using --clean parameter
If this is a cloud/virtualized environment, try this:
If this is not a cloud/virtualized environment, remove the --point-to-point flag.
thanks for your reply. However, I got this error too as below.
(note this operator all in cloud server)
[root@iZcZ ~]# /opt/vertica/sbin/install_vertica -s '47.114.xxx.xxx,47.114.xxx.xxx,47.114.xxx.xxx' --clean --rpm vertica-10.0.0-0.x86_64.RHEL6.rpm --point-to-point
Vertica Analytic Database 10.0.0-0 Installation Tool
Mapping hostnames in --hosts (-s) to addresses...
Error: cannot find which cluster host is the local host.
Hint: Is this node in the cluster? Did its IP address change?
Installation FAILED with errors.
Installation stopped before any changes were made.
[root@iZbZ ~]# cat /opt/vertica/config/admintools.conf
[SSHConfig]
ssh_user =
ssh_ident =
ssh_options = -oConnectTimeout=30 -o TCPKeepAlive=no -o ServerAliveInterval=15 -o ServerAliveCountMax=2 -o StrictHostKeyChecking=no -o BatchMode=yes
[BootstrapParameters]
awsendpoint = null
awsregion = null
[Configuration]
format = 3
install_opts = -s '127.0.0.1' --clean
default_base = /home/dbadmin
controlmode = broadcast
controlsubnet = default
spreadlog = False
last_port = 5433
tmp_dir = /tmp
ipv6 = False
atdebug = False
atgui_default_license = False
unreachable_host_caching = True
aws_metadata_conn_timeout = 2
rebalance_shards_timeout = 36000
database_state_change_poll_timeout = 21600
wait_for_shutdown_timeout = 3600
pexpect_verbose_logging = False
sync_catalog_retries = 2000
admintools_config_version = 109
[Cluster]
hosts = 127.0.0.1
[Nodes]
node0001 = 127.0.0.1,/home/dbadmin,/home/dbadmin
appreciate it for your help~
It looks like you created a cluster once before with localhost 127.0.0.1.
Try renaming /opt/vertica/config/admintools.conf to something else. Do this on all 3 nodes:
Then, on any one node, try again:
Hi ~
thanks for everyone~
now I have already installed a Vertica cluster with three hosts' success.
However, I got another error when I want to create an EON mode database.
Can Vertica EON mode support other Object Storage Service which not AWS or Auxier?
[dbadmin@iZcZ ~]$ admintools -t create_db -x auth_params.conf -s '172.16.xxx.xxx,172.16.xxx.xxx,172.16.xxx.xxx' -d VMart3 -p 123 --depot-path=/home/dbadmin/depot --shard-count=6 --communal-storage-location=oss://xxxx-bucket -D /home/dbadmin/data/ -c /home/dbadmin/catalog/ --depot-size 5G
Distributing changes to cluster.
Creating database VMart3
Bootstrap on host 172.16.xxx.66 return code -6 stdout '' stderr ''
Error: Bootstrap on host 172.16.xxx.66 return code -6 stdout '' stderr ''
thanks for friends help!
Hi @LenoyJ ~
thanks for your reply, I have already solved Vertica install problems.
unfortunate I got new error now when I try to create an EON mode database like pre message, would appreciate it if you can give me some suggestion about it.
thanks again
Best wish~
I doubt AliCloud's OSS (oss://) would work as a supported communal storage location for Eon yet (as of 10.0). Supported Eon mode locations in the cloud today are AWS & GCP. On-prem Eon mode deployments includes Pure Storage, MinIO & HDFS. Though you could theoretically deploy MinIO and HDFS on the cloud.
Enterprise Mode, on the other hand, does not have a requirement of a communal storage and should work on any commodity/virtualized hardware that meets our minimum requirements. It looks like you are trying out Vertica and deploying a few nodes, maybe try Enterprise instead for now? Like so:
Hi @LenoyJ
Thanks for your reply.
unfortunately, my work is testing AliCloud's OSS can or cannot work as communal storage for EON, I can't change to use Enterprise Mode.
so can you give me some advice about this problem please?
thanks again.
Does Alicloud OSS offer a connection using S3 protocol? If so, then you may be able to connect using equivalent s3:// URL.
Please see Alicloud documentation for S3 compatibility information: https://www.alibabacloud.com/help/doc-detail/64919.html?spm=a2c5t.11065259.1996646101.searchclickresult.2c434c0d8ByRNe
Unfortunately, we don't officially support other protocols.
thanks for your reply.
However, I got an error when I change to S3 as communal storage to create an EON mode database.
the log as below.
appreciate it if you can give me some suggestion about this error.
[dbadmin@iZbp1hduhk14uh847bf7ncZ ~]$ admintools -t create_db -x auth_params.conf -s '172.16.xxx.a,172.16.xxx.b,172.16.xxx.c' -d VMart1 -p 123 --depot-path=/home/dbadmin/depot --shard-count=6 --communal-storage-location=s3://vertica-test-1 -D /home/dbadmin/data/ -c /home/dbadmin/catalog/ --depot-size 5G
Distributing changes to cluster.
Creating database VMart1
Starting bootstrap node v_vmart1_node0002 (172.16.xxx.b)
Starting nodes:
v_vmart1_node0002 (172.16.xxx.b)
Starting Vertica on all nodes. Please wait, databases with a large catalog may take a while to initialize.
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (DOWN)
Node Status: v_vmart1_node0002: (UP)
Creating database nodes
Creating node v_vmart1_node0001 (host 172.16.xxx.a)
Creating node v_vmart1_node0003 (host 172.16.xxx.c)
Generating new configuration information
Stopping single node db before adding additional nodes.
Database shutdown complete
Starting all nodes
Start hosts = ['172.16.xxx.xx', '172.16.xxx.xx', '172.16.xxx.xx']
Starting nodes:
v_vmart1_node0002 (172.16.xxx.xx)
v_vmart1_node0001 (172.16.xxx.xx)
v_vmart1_node0003 (172.16.xxx.xx)
Starting Vertica on all nodes. Please wait, databases with a large catalog may take a while to initialize.
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
It is suggested that you continue waiting.
Do you want to continue waiting? (yes/no) [yes]
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
It is suggested that you continue waiting.
Do you want to continue waiting? (yes/no) [yes]
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
It is suggested that you continue waiting.
Do you want to continue waiting? (yes/no) [yes]
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
It is suggested that you continue waiting.
Do you want to continue waiting? (yes/no) [yes]
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
Node Status: v_vmart1_node0001: (DOWN) v_vmart1_node0002: (DOWN) v_vmart1_node0003: (DOWN)
ERROR: Not all nodes came up, but not all down. Run scrutinize.
Unable to establish vsql script connection: Unable to connect to 'VMart1'
Unable to establish client-server connection: Unable to connect to 'VMart1'
Unable to create depot storage locations (if Eon) without a client-server connection.
Unable to rebalance shards (if Eon) without a client-server connection.
Unable to set K-safety value without a client-server connection.
Unable to install default extension packages without a vsql script connection
Unable to sync database catalog (if Eon) without a client-server connection.
Database creation SQL tasks included one or more failures (see above).
Database VMart1 created successfully, some nodes may have had startup problems.
thanks again for your attention~