install HCM failed : etcd: failed to get etcd API

Hello,

I execute the install of HCM 2019.11 (for lab env with one master only, and nfs share is located on master).

I got the following error :

[root@hcm-master-1 ITOM_Suite_Foundation_2019.11.00104]# ./install -m /usr/tmp/hcm-2019.11-metadata.tgz --nfs-server 10.123.123.123 --nfs-folder /var/vols/itom/core
2023-08-30T15:51:00+02:00 INFO : [NODE:hcm-master-1.xyz.fr] * initializing
2023-08-30T15:51:00+02:00 INFO : ITOM core platform installation pre-check............................. [ ALREADY DONE ]
2023-08-30T15:51:00+02:00 INFO : Create password for the administrator.
*********************************************************************
Passwords must meet the following requirements:
1) Minimum Length - 8 characters
2) Maximum length - 20 characters
3) Must contain all the following types of characters:
 . numbers (0-9)
 . lowercase letters (a-z)
 . uppercase letters (A-Z)
 . special characters, not including the space character (!,@,#,etc.)
*********************************************************************
Administrator password:   >
Confirm the password:     >
2023-08-30T15:51:28+02:00 INFO : Start to install the ITOM core platform on this server
2023-08-30T15:51:28+02:00 INFO : Configure the ITOM core platform required firewall settings .......... [ OK ]
2023-08-30T15:51:35+02:00 INFO : Copy the base installation files to /opt/kubernetes .................. [ ALREADY DONE ]
2023-08-30T15:51:35+02:00 INFO : [NODE:hcm-master-1.xyz.fr] * configuring docker
2023-08-30T15:51:35+02:00 INFO : Copy the docker installation files to /opt/kubernetes ................ [ ALREADY DONE ]
(...)
2023-08-30T15:51:38+02:00 INFO : Pack the installer bundle used for adding nodes through UI ........... [ ALREADY DONE ]
2023-08-30T15:51:38+02:00 INFO : Move the installer bundle used for adding nodes through UI ........... [ ALREADY DONE ]
2023-08-30T15:51:38+02:00 INFO : Check component's pods status ........................................
Installation stuck on this step because some pods are not "running" 1/1
[root@hcm-master-1 ~]# tail -200 /tmp/install.20230830155100.log
(...)
2023-08-30T16:08:15.671106645+02:00 DEBUG Component's pods ( cdf-apiserver itom-cdf-tiller itom-vault kubernetes-vault suite-db suite-installer-frontend) is not running. Wait for 180 seconds and recheck: 178
2023-08-30T16:08:20.674677036+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'cdf-apiserver-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:20.674677036+02:00 DEBUG exit_code: 0
2023-08-30T16:08:20.7729191+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'itom-cdf-tiller-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:20.7729191+02:00 DEBUG exit_code: 0
2023-08-30T16:08:20.870143982+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'itom-vault-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:20.870143982+02:00 DEBUG exit_code: 0
2023-08-30T16:08:20.98434796+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'kubernetes-vault-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:20.98434796+02:00 DEBUG exit_code: 0
2023-08-30T16:08:21.088022575+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'suite-db-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:21.088022575+02:00 DEBUG exit_code: 0
2023-08-30T16:08:21.178771976+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'suite-installer-frontend-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:21.178771976+02:00 DEBUG exit_code: 0
2023-08-30T16:08:21.271983341+02:00 DEBUG Component's pods ( cdf-apiserver itom-cdf-tiller itom-vault kubernetes-vault suite-db suite-installer-frontend) is not running. Wait for 180 seconds and recheck: 179
2023-08-30T16:08:26.275437325+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'cdf-apiserver-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:26.275437325+02:00 DEBUG exit_code: 0
2023-08-30T16:08:26.393642447+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'itom-cdf-tiller-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:26.393642447+02:00 DEBUG exit_code: 0
2023-08-30T16:08:26.485988564+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'itom-vault-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:26.485988564+02:00 DEBUG exit_code: 0
2023-08-30T16:08:26.57861773+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'kubernetes-vault-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:26.57861773+02:00 DEBUG exit_code: 0
2023-08-30T16:08:26.672582173+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'suite-db-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:26.672582173+02:00 DEBUG exit_code: 0
2023-08-30T16:08:26.76414748+02:00 DEBUG exec_cmd # kubectl get pods -n core|grep -P 'suite-installer-frontend-'|awk '/Running/{split($2,a,"/");if(a[1]==a[2]&&a[2]>0)print $1}'
2023-08-30T16:08:26.76414748+02:00 DEBUG exit_code: 0
2023-08-30T16:08:26.862001028+02:00 FATAL A timeout occurred while waiting for some component's pods ( cdf-apiserver itom-cdf-tiller itom-vault kubernetes-vault suite-db suite-installer-frontend) to running. Refer to /tmp/install.20230830155100.log file for detail information. If need, please contact system administrator or Micro Focus support.
[root@hcm-master-1 ITOM_Suite_Foundation_2019.11.00104]# kubectl get pods --all-namespaces -o wide
NAMESPACE     NAME                                 READY   STATUS      RESTARTS   AGE   IP              NODE                      NOMINATED NODE   READINESS GATES
core          itom-cdf-deployer-bqvf8              1/1     Running     0          24h   172.16.56.3     hcm-master-1.xyz.fr   <none>           <none>
core          itom-cdf-move-pack                   0/1     Completed   0          24h   172.16.56.5     hcm-master-1.xyz.fr   <none>           <none>
core          itom-vault-9c7587898-zsrds           0/1     Running     361        24h   172.16.56.4     hcm-master-1.xyz.fr   <none>           <none>
kube-system   apiserver-hcm-master-1.xyz.fr    1/1     Running     0          25h   10.254.100.40   hcm-master-1.xyz.fr   <none>           <none>
kube-system   controller-hcm-master-1.xyz.fr   1/1     Running     0          25h   10.254.100.40   hcm-master-1.xyz.fr   <none>           <none>
kube-system   coredns-cv8d2                        1/1     Running     0          25h   172.16.56.2     hcm-master-1.xyz.fr   <none>           <none>
kube-system   etcd-hcm-master-1.xyz.fr         1/1     Running     0          25h   10.254.100.40   hcm-master-1.xyz.fr   <none>           <none>
kube-system   kube-flannel-ds-amd64-v9hsk          1/1     Running     0          25h   10.254.100.40   hcm-master-1.xyz.fr   <none>           <none>
kube-system   scheduler-hcm-master-1.xyz.fr    1/1     Running     0          25h   10.254.100.40   hcm-master-1.xyz.fr   <none>           <none>
[root@hcm-master-1 ITOM_Suite_Foundation_2019.11.00104]# kubectl logs itom-vault-9c7587898-zsrds -n core
[s6-init] making user provided files available at /var/run/s6/etc...exited 0.
[s6-init] ensuring user provided files have correct perms...exited 0.
[fix-attrs.d] applying ownership & permissions fixes...
[fix-attrs.d] done.
[cont-init.d] executing container initialization scripts...
[cont-init.d] done.
[services.d] starting services
[services.d] done.
Trying to create a root certs and key if not present...
Cannot find ca.key in local path /ssl/. Try to find it in the k8s secret ric-root-cert
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed
100  7205    0  7205    0     0   242k      0 --:--:-- --:--:-- --:--:--  251k
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed
100  7205    0  7205    0     0   281k      0 --:--:-- --:--:-- --:--:--  281k
Found ca.key in the k8s secret ric-root-cert. Saving ca.key & ca.crt from the k8s secret ric-root-cert into local path /ssl/
No existing server.key or server.crt in local path /ssl/. Generate a new server.key & a new server.crt
Generating RSA private key, 4096 bit long modulus
.....................................++++
...............................................................................................................++++
e is 65537 (0x010001)
Signature ok
subject=CN = itom
Getting CA Private Key
http_code:000: unknown error
try to start server
Error initializing storage of type etcd: failed to get etcd API version: client: etcd cluster is unavailable or misconfigured; error #0: dial tcp: lookup hcm-master-1.xyz.fr on 172.1.2.3:53: no such host

http_code:000: unknown error
try to start server
Error initializing storage of type etcd: failed to get etcd API version: client: etcd cluster is unavailable or misconfigured; error #0: dial tcp: lookup hcm-master-1.xyz.fr on 172.1.2.3:53: no such host

http_code:000: unknown error
try to start server
Error initializing storage of type etcd: failed to get etcd API version: client: etcd cluster is unavailable or misconfigured; error #0: dial tcp: lookup hcm-master-1.xyz.fr on 172.1.2.3:53: no such host

http_code:000: unknown error
try to start server
Error initializing storage of type etcd: failed to get etcd API version: client: etcd cluster is unavailable or misconfigured; error #0: dial tcp: lookup hcm-master-1.xyz.fr on 172.1.2.3:53: no such host
If anybody has some clue to help me to resolve this Issue, I would be grateful.
Thanks and regards,
Jean-Philippe HAAG

Tags:

  • Hello Jean-Philippe,

    I'm sorry to hear that you are encountering problems.

    Is there a specific use case that you are installing HCM 2019.11 instead of HCM 2020.05?

    Based on the messages that you provided, it appears that there is a problem with the etcd pod configuration and Name Resolution.

    Are you using /etc/hosts files for your name resolution or DNS?

    Do all of your hosts (including your master) end with .fr for the Top-Level DNS domain?

    The 172.16.x.x address is also used for routing between the pods and between the hosts.

    Here's an example from an HCM 2020.05 install.properties file:

    # Parameter: POD_CIDR
    # Type : Optional
    # WARNING : ! Micro Focus strongly recommends to only change the default value if your network configuration requires you
    # to do so and if you have a good understanding of the Flannel network fabric configuration requirements.
    # Use : Specifies the private network address range for the Kubernetes pods. This network does not need to exist or match
    # with your network. The default value will work fine in most cases. The network addresses notation is the
    # Classless Inter-Domain Routing (CIDR) format. A CIDR notation is comprised of an IP address followed by forward
    # slash ('/') character followed by a a decimal number. The minimum useful network prefix is /24 and the maximum
    # network prefix is /8. If POD_CIDR is not specified, then the default value is "172.16.0.0/16".
    # For the default and allowable values see the table "Relationship between POD_CIDR and POD_CIDR_SUBNETLEN values".
    # Note : This address range must not overlap with the IP range assigned for Kubernetes services. See SERVICE_CIDR.
    # Format : String
    # Example : POD_CIDR="172.16.0.0/16"
    #POD_CIDR="172.16.0.0/16"

    ...
    # Parameter: SERVICE_CIDR
    # Type : Optional
    # WARNING : ! Micro Focus strongly recommends to only change the default value if your network configuration requires you
    # to do so and if you have a good understanding of the Flannel network fabric configuration requirements.
    # Use : Specifies the network address for the Kubernetes services. The network addresses notation is the
    # Classless Inter-Domain Routing (CIDR) format. A CIDR notation is comprised of an IP address followed by forward
    # slash ('/') character followed by a decimal number. The minimum useful network prefix is /27 and the maximum
    # network prefix is /12. If SERVICE_CIDR is not specified, then the default value is "172.17.17.0/24". E.g, if you
    # set SERVICE_CIDR to /27, Kubernetes can only allocated addresses for 32 services.
    # Note : This must not overlap with any IP ranges assigned to nodes for pods. See POD_CIDR.
    # Format : String
    # Example : SERVICE_CIDR="172.17.17.0/24"
    #SERVICE_CIDR="172.17.17.0/24"
    ...

    The IP address that you are listing for the master may have a conflict and could be the cause of the problem.

    You can edit your install.properties file to change the IP Address range used by these services if your host IP addresses conflict.

    I hope that this is helpful.

    Regards,

    Mark

  • Hello Mark,

    First, thank you for your help.

    Here are my answer : 

    - Is there a specific use case that you are installing HCM 2019.11 instead of HCM 2020.05?

    +No specific use case, just wanted to be on the same version of HCM as a previous customer (2019.11).

    - Are you using /etc/hosts files for your name resolution or DNS?

    +Yes, exactly.

    - Do all of your hosts (including your master) end with .fr for the Top-Level DNS domain?

    +Yes, we dont really use a domain, as machine are only accessible from a gateway, not directly, and domain is not used on these hcm servers.But I set the hostname to respect the pre-requisites on step prepare (if I let short hostname ie. hcm-master-1, then it failed the pre-requisites step): fhttps://docs.microfocus.com/doc/Hybrid_Cloud_Management/2019.11/ConfigHostNameResolution

    [root@hcm-master-1 ~]# cat /etc/hosts
    127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4
    ::1 localhost6 localhost6.localdomain6
    10.X.Y.40 hcm-master-1.xyz.fr hcm-master-1
    10.X.Y.41 hcm-worker-1.xyz.fr hcm-worker-1
    10.Y.Y.42 hcm-vertica-1.xyz.fr hcm-vertica-1
    10.X.Y.43 hcm-rdb-1.xyz.fr hcm-rdb-1

    - You can edit your install.properties file to change the IP Address range used by these services if your host IP addresses conflict.

    + I will do so, thanks.But should I launch uninstall after modifying install.properties or simply run again the install below ?

     ./install -m /usr/tmp/hcm-2019.11-metadata.tgz --nfs-server 10.123.123.123 --nfs-folder /var/vols/itom/core

    Best regards,

    Jean-Philippe

  • Just before changing routing, I try to disable firewalld, and launch again install,

    And logs are little bit different : 

    http_code:000: unknown error
    try to start server
    Error initializing storage of type etcd: failed to get etcd API version: client: etcd cluster is unavailable or misconfigured; error #0: dial tcp: lookup hcm-master-1.xy.fr on 172.1.2.3:53: server misbehaving
    while before : 

    http_code:000: unknown error
    try to start server
    Error initializing storage of type etcd: failed to get etcd API version: client: etcd cluster is unavailable or misconfigured; error #0: dial tcp: lookup hcm-master-1.xy.fr on 172.1.2.3:53: no such host

  • Verified Answer

    Hello Jean-Philippe,

    Thank you for your response and this additional information.

    I asked about the .fr because there are known issues with older versions of CDF when using .local for the FQDN.  I didn't think it would be an issue for .fr, but it was something to be aware of to possibly investigate further if needed.

    Thank you for showing that the IP addresses of the nodes are not using the 172.x.x.x address, so there should not be any conflict for that item.

    You are not using DNS.  This is very significant and there are specific configuration changes that need to be made within the install.properties file in order to not use DNS.  If you do not make these changes, then the install will fail, similar to what you are encountering.

    You need to set the KUBE_DNS_HOSTS parameter within the install.properties file.

    https://docs.microfocus.com/doc/Hybrid_Cloud_Management/2019.11/ConfigureInstallPropertiesOnPremise#KUBE_DNS_HOSTS

    The first line under this topic states:
    Configuration of this parameter is mandatory in a non-DNS environment.

    Since the errors that you are encountering are name resolution issues, this is most likely the cause of your issue.

    I would also run the 'ip route' command to see if the 172.x.x.x IP addresses show within the output.

    Unfortunately, HCM 2019.11 is no longer supported.  In fact, HCM 2020.05 is no longer supported as of July 31, 2023.

    Good luck with your efforts.

    Regards,

    Mark

  • Hello Mark,

    Thanks for your answer.

    Indeed, I'm not using DNS.I noticed on file install.properties that the comment for parameter KUBE_DNS_HOSTS was "Optional" (that's why I leaved blank).

    As of your recommendation, I changed its value to : KUBE_DNS_HOSTS="/etc/hosts"

    and 

    [root@hcm-master-1 ITOM_Suite_Foundation_2019.11.00104]# cat /etc/hosts
    127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4
    ::1 localhost6 localhost6.localdomain6
    10.X.Y.40 hcm-master-1.xyz.fr hcm-master-1
    10.X.Y.41 hcm-worker-1.xyz.fr hcm-worker-1
    10.X.Y42 hcm-vertica-1.xyz.fr hcm-vertica-1
    10.X.Y.43 hcm-rdb-1.xyz.fr hcm-rdb-1

    [root@hcm-master-1 ITOM_Suite_Foundation_2019.11.00104]# ip route
    default via 10.X.Y.1 dev eth0 proto static metric 100
    10.X.Y.0/24 dev eth0 proto kernel scope link src 10.X.Y.40 metric 100
    172.16.56.0/24 dev cni0 proto kernel scope link src 172.16.56.1
    172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1

    I launched again installation : 

    ./install -m /usr/tmp/hcm-2019.11-metadata.tgz --nfs-server 10.X.Y.40 --nfs-folder /var/vols/itom/core

    but I got the same errors as previously from the install log: 

    023-09-04T09:03:35+02:00 INFO : Pack the installer bundle used for adding nodes through UI ........... [ ALREADY DONE ]
    2023-09-04T09:03:35+02:00 INFO : Move the installer bundle used for adding nodes through UI ........... [ ALREADY DONE ]
    2023-09-04T09:03:35+02:00 INFO : Check component's pods status ........................................ [ FAILED ]
    2023-09-04T09:20:24+02:00 FATAL : A timeout occurred while waiting for some component's pods ( cdf-apiserver itom-cdf-tiller itom-vault kubernetes-vault suite-db suite-installer-frontend) to running. Refer to /tmp/install.20230904090310.log file for detail information. If need, please contact system administrator or Micro Focus support.

    [root@hcm-master-1 tmp]# kubectl logs itom-vault-9c7587898-knzbg -n core

    (...)

    http_code:000: unknown error
    try to start server
    Error initializing storage of type etcd: failed to get etcd API version: client: etcd cluster is unavailable or misconfigured; error #0: dial tcp: lookup hcm-master-1.coralio.fr on 172.17.17.78:53: no such host

    http_code:000: unknown error
    try to start server
    Error initializing storage of type etcd: failed to get etcd API version: client: etcd cluster is unavailable or misconfigured; error #0: dial tcp: lookup hcm-master-1.coralio.fr on 172.17.17.78:53: no such host

    root@hcm-master-1 tmp]# ifconfig -a
    cni0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500
    inet 172.16.56.1 netmask 255.255.255.0 broadcast 0.0.0.0
    inet6 fe80::3c58:c6ff:fe74:8a04 prefixlen 64 scopeid 0x20<link>
    ether 3e:58:c6:74:8a:04 txqueuelen 1000 (Ethernet)
    RX packets 402837 bytes 28385811 (27.0 MiB)
    RX errors 0 dropped 0 overruns 0 frame 0
    TX packets 401374 bytes 157863154 (150.5 MiB)
    TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0

    docker0: flags=4099<UP,BROADCAST,MULTICAST> mtu 1500
    inet 172.17.0.1 netmask 255.255.0.0 broadcast 172.17.255.255
    ether 02:42:5d:8e:e6:4e txqueuelen 0 (Ethernet)
    RX packets 0 bytes 0 (0.0 B)
    RX errors 0 dropped 0 overruns 0 frame 0
    TX packets 0 bytes 0 (0.0 B)
    TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0

    eth0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500
    inet 10.X.Y.40 netmask 255.255.255.0 broadcast 10.254.100.255
    inet6 xxxxxx prefixlen 64 scopeid 0x20<link>
    ether da:d3:6a:34:57:71 txqueuelen 1000 (Ethernet)
    RX packets 42636 bytes 117247503 (111.8 MiB)
    RX errors 0 dropped 0 overruns 0 frame 0
    TX packets 26944 bytes 1949720 (1.8 MiB)
    TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0

    lo: flags=73<UP,LOOPBACK,RUNNING> mtu 65536
    inet 127.0.0.1 netmask 255.0.0.0
    inet6 ::1 prefixlen 128 scopeid 0x10<host>
    loop txqueuelen 1000 (Boucle locale)
    RX packets 24014503 bytes 5422304171 (5.0 GiB)
    RX errors 0 dropped 0 overruns 0 frame 0
    TX packets 24014503 bytes 5422304171 (5.0 GiB)
    TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0

    veth4e21ec3c: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500
    inet6 xxxxxxx prefixlen 64 scopeid 0x20<link>
    ether 36:9a:8f:78:86:b3 txqueuelen 0 (Ethernet)
    RX packets 1 bytes 42 (42.0 B)
    RX errors 0 dropped 0 overruns 0 frame 0
    TX packets 20 bytes 1516 (1.4 KiB)
    TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0

    veth6064fb45: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500
    inet6 xxxxxxx prefixlen 64 scopeid 0x20<link>
    ether d6:d3:aa:1a:69:77 txqueuelen 0 (Ethernet)
    RX packets 685133 bytes 73134236 (69.7 MiB)
    RX errors 0 dropped 0 overruns 0 frame 0
    TX packets 688909 bytes 168532029 (160.7 MiB)
    TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0

    vethc42ed56b: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500
    inet6 fe80::702e:f0ff:fe19:b0b6 prefixlen 64 scopeid 0x20<link>
    ether 72:2e:f0:19:b0:b6 txqueuelen 0 (Ethernet)
    RX packets 1 bytes 42 (42.0 B)
    RX errors 0 dropped 0 overruns 0 frame 0
    TX packets 23 bytes 1734 (1.6 KiB)
    TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0

    vethde890c0b: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500
    inet6 fe80::8cf4:deff:fe9a:534b prefixlen 64 scopeid 0x20<link>
    ether 8e:f4:de:9a:53:4b txqueuelen 0 (Ethernet)
    RX packets 345172 bytes 35333462 (33.6 MiB)
    RX errors 0 dropped 0 overruns 0 frame 0
    TX packets 339957 bytes 63774998 (60.8 MiB)
    TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0

    I think changes on install.properties are not applied, ie do I need to do something special so that the changes on install.properties are applied ?

    Best regards,

    Jean-Philippe.

  • Hello,
    just this message to keep you inform that installation was complete successfully, after uninstall and install again after KUBE_DNS_HOSTS correctly set with a hosts file.
    Thanks Mark for your help.

    Best regards,

    Jean-Philippe

  • My server master where I launch installation is a centos 7.

    Now I need to connect to https://hcm-master-1.xxx.fr:3000
    My colleague set a reverse proxy from the gateway server to expose access to master server, by configuring haproxy.

    I can acces login page of portal on port 3000, but I cannot connect, I know password is right because if I set a bad password, I got a message username/password incorrect.

    It should be a traffic/flow problem.If anyone have an idea where I can found log, I would be grateful.

    Thanks.

    Jean-Philippe

  • From my browser in developper mode I get this response when I try connect

    {
        "timestamp": 1693902942432,
        "status": 500,
        "error": "Internal Server Error",
        "exception": "org.springframework.web.client.ResourceAccessException",
        "message": "I/O error on GET request for \"https://idm.core:443/idm-service/v3.0/tokens/xxx\": idm.core: Name or service not known; nested exception is java.net.UnknownHostException: idm.core: Name or service not known",
        "path": "/urest/v1.1/deployment/bcb8a404-7b5e-4fa4-a13d-4ae32f1f8ee4"
    }

  • Hello Jean-Philippe,

    I'm in a training class this week and do not have access to my lab setup.  However, this looks like an issue with accessing the IDM pod within the Core namespace.

    Look at the logs for the IDM pod as well as the state to see more details.  There is also the nginx pod that brings up the login page and directs to the IDM pod.

    Good luck.

    Regards,

    Mark

  • Thanks for you answer Mark.
    I searched logs from pod apiserver : # kubectl logs cdf-apiserver-5cff5497b9-5k7sw -c cdf-apiserver -n core

    (...)
    2023-09-06 07:49:01.214 ERROR 257 --- [nio-8080-exec-9] o.a.c.c.C.[.[.[/].[dispatcherServlet] : Servlet.service() for servlet [dispatcherServlet] in context with path [] threw exception

    org.springframework.web.client.ResourceAccessException: I/O error on GET request for "">idm.core:443/.../anonymous": idm.core; nested exception is java.net.UnknownHostException: idm.core


    (...)
    2023-09-06 07:49:13.471 ERROR 257 --- [nio-8080-exec-5] o.a.c.c.C.[.[.[/].[dispatcherServlet]    : Servlet.service() for servlet [dispatcherServlet] in context with path [] threw exception

    org.springframework.web.client.ResourceAccessException: I/O error on GET request for "https://idm.core:443/idm-service/v3.0/tokens/anonymous": idm.core: Name or service not known; nested exception is java.net.UnknownHostException: idm.core: Name or service not known

    (...)
    2023-09-06 07:49:21.195  WARN 257 --- [TaskScheduler-1] s.v.service.imp.InstallModeServiceImpl   : Failed to get base-configmap ,PathNotFoundException [No results for path: $['data']['EXTERNAL_ACCESS_HOST']]
    2023-09-06 07:49:21.237 DEBUG 257 --- [TaskScheduler-1] atchdog$$EnhancerBySpringCGLIB$$de427864 : getHostForSwaggerUI=hcm-master-1.xxx.fr