Troubleshooting issues

You can find information about the issues or problems that you might encounter while working with Rational® Test Automation Server. Details about issues, their causes and the resolutions that you can apply to fix the issues are described.

The troubleshooting issues are presented to you in the following tables based on where or when you might encounter these issues on Rational® Test Automation Server.
Table 1. Troubleshooting issues: installation

Problem

Description

Solution

When you are installing the server software and you encounter errors in the scripts that are running.

At times, scripts might not appear to be running due to any of the following reasons:
  • Slow connection speeds.
  • Insufficient CPU, memory, or disk resources.
  • A firewall that was configured incorrectly is already enabled.
You can complete any of the following tasks:
  • To identify the issue, you can perform a diagnostic check by running the following command:
    journalctl -u k3s
    This command displays the log that you can use to check for the problem.
  • Run the following command to see which pods are running and which pods are not running:
    kubectl get pods -A

    Run the following command to get details about a specific pod:

    kubectl describe pod -n <namespace> <pod name>
  • Follow the on-screen instructions to resolve the errors.
  • Some issues can be solved by re-running the following script:
    ubuntu-init.sh

The DNS is not working as expected.

The DNS configuration that is used by the cluster can be displayed by using the following command:

kubectl get cm -n kube-system coredns -ojsonpath="{.data.Corefile}"
The forward setting displays the nameservers that are used. For example, you might see the following in the corefile:
 .:53 {
   :
      forward . 1.1.1.1 8.8.8.8
   :
  }

A script (ubuntu-set-dns.sh) is supplied for managing these values.

For example, to set the DNS values for the values shown in the previous example:
sudo ubuntu-set-dns.sh --server 1.1.1.1 --server 8.8.8.8
Note: If you do not use sudo in the command, the script runs but the configuration might be lost if the cluster is restarted.
To learn more about the behavior of the script, run the following command:
ubuntu-set-dns.sh --help
Table 2. Troubleshooting issues: server administration

Problem

Description

Solution

Less recently patched versions of OpenShift can give errors whenever OpenShift performs internal checks.

For example, the following error is displayed:
ValidationError(Route.spec.to): missing required field "weight"

The errors are caused because OpenShift performs the internal checks that are invalid.

Apply the latest OpenShift patches when they become available.

If you do not want to apply the patches or cannot apply the patches, you can disable these checks in OpenShift by appending the following option to the helm command:

--disable-openapi-validation

If your user-realm role is changed when you are logged in a session, the changed role is not applied immediately or even after the browser is refreshed.

You must log out of the session and log in again for the changed role to take effect.

You see the following message displayed on Rational® Test Automation Server:

you can’t request to join a project that has no owners

You requested to join an orphaned project. The project is orphaned because the administrator deleted that user ID from the user registry.

That user registry can be either the local Keycloak user registry or an external user registry, for example LDAP that is federated with the local Keycloak registry.

That user ID can be deleted by the administrator in cases such as the user leaving the project or company.

Ask your administrator to take ownership of the project, and then add you as a project owner or member.
Table 3. Troubleshooting issues: resource monitoring

Problem

Description

Solution

You are not able to add a Prometheus server as a Resource Monitoring source.

The cause might be that you have not installed the Prometheus server at the time of server installation.

Verify that the Prometheus server was installed in Helm at the time of server installation. See Installing the server software on Ubuntu by using k3s. If not, consult your cluster administrator to get the Promethues server installed and configured.

Table 4. Troubleshooting issues: configuring test runs

Problem

Description

Solution

When you configure a run of a schedule that matches the following conditions:
  • The schedule has two user groups configured to run on static agents when the schedule was created in Rational® Performance Tester V10.1.
  • One of the user groups is disabled and the asset is committed to the remote repository.
Both the static agents are displayed as available for the test run in the Location tab of the Execute test asset dialog box when only one agent that is configured for the user group must be available.
The cause might be because of the following reasons:
  • The schedule was created in Rational® Performance Tester V10.1.
  • The user group that is disabled is not removed or deleted from the test resources.
  • The agent configured on the disabled user group is already added as an agent to the server project and is available for selection.
To resolve the problem, select from either of the following methods:
  • By using Rational® Performance Tester V10.1.1.
    Perform the following steps:
    1. Open the schedule in Rational® Performance Tester V10.1.1.
    2. Save the schedule and the project.
    3. Commit your test asset to the remote repository.
    4. Proceed to configure a run for the schedule on Rational® Test Automation Server V10.1.1.
  • By using Rational® Performance Tester V10.1.
    Perform the following steps:
    1. Select the disabled user group.
    2. Click Remove.
    3. Save the schedule and the project.
    4. Commit your test asset to the remote repository.
    5. Proceed to configure a run for the schedule on Rational® Test Automation Server V10.1.1.
Table 5. Troubleshooting issues: test or stub runs

Problem

Description

Solution

You encounter any of the following issues:
  • When many tests are run simultaneously on the default cluster location and you observe the following issues:
    • Out of memory errors.
    • Observe that the test runs are slow with a high CPU usage.
    • The Kubernetes pods are getting evicted.
  • When you run an AFT suite that contains multiple Web UI tests and you observe the following issues:
    • Error stating that the browser might not be installed or the browser version is unsupported.
    • Error stating multiple random time-outs or an internal error.
The issue is seen when any of the following events occur:
  • Many tests are run in parallel.
  • The memory that is used by the tests during the test run exceeds the allocated default memory of 1 GB.
  • The default memory of the container is not adequate for the test run.
  • Pods are evicted due to low node memory.

To resolve the problem, you can increase the resource allocation for test runs.

You can enter arguments in the Java Arguments field in the Advanced settings panel of the Execute test asset dialog box when configuring a test run.

Important: The memory settings that you configure for a test run is persisted for the test when ever you run it. You must use this setting judiciously. Configuring all tests for an increased memory limit might affect subsequent test runs or cause other memory issues when tests run simultaneously.
You can increase the resource allocation for test runs by using any of the following arguments:
For... Enter the argument... Result

Specifying the memory limit of the init container .

-Dexecution.init.resource.memory.limit

Change the memory limit of the init container from the default value of 1 Gi to 1024Mi.

Configuring a larger memory request for the init container to avoid pod eviction.

-Dexecution.init.resource.memory.request=1024Mi

Increases the initial memory request for the init container from the default value of 64Mi to 1024Mi.

Specifying the cpu request for the init container.

-Dexecution.init.resource.cpu.request

Increases the cpu limit of the init container from the default value of 50m to 60m.

Specifying a maximum heap size for the test run.

-Xmx4g

Increases the allotted 1 GB memory to 4 GB.

Specifying the memory limit of the container explicitly for the test run.

-Dexecution.resource.memory.limit=<custom_memory_value>Gi
Note: You must enter the value for the memory limit that you want in <custom_memory_value>.

Increases the allotted memory to any value you specify.

For example, if you want to set the memory limit to 5 GB, the argument can be:

-Dexecution.resource.memory.limit=5Gi.

The argument enables the memory to be set to 5 GB and overrides the default memory size.

Specifying the memory request for the container used by the test run.

-Dexecution.resource.memory.request

Increases the memory request from the default value of 64Mi to 1Gi.

Specifying the cpu request for the container used by the test run.

-Dexecution.resource.cpu.request

Increases the cpu limit of the container from the default value of 50m to 60m.

You are not able to run the Istio stubs from the Execution page.

The cause might be that you have not enabled the service virtualization via Istio at the time of server installation. The default configuration does not enable service virtualization via Istio.

Contact your cluster administrator or if you have the privileges, configure Helm as follows:

  • For Multi-tenant clusters

    If the cluster is shared and the product may only virtualize services running in specific namespaces then add the following parameter to the Helm install:

    --set execution.istio.enabled=true

    Then enable service virtualization in specific namespaces using this command:

    kubectl create rolebinding istio-virtualization-enabled -n {namespace} --clusterrole={my-ots}-execution-istio-test-system --serviceaccount=test-system:{my-ots}-execution
    Note: Uninstalling the chart does not clean up these manually created role bindings.
  • For Single-tenant clusters

    If the cluster is not shared and the product may virtualize any service running in the whole cluster then add the following parameters to the Helm install:

    --set execution.istio.enabled=true

    --set execution.istio.clusterRoleBinding.create=true

The cause might be that the fully qualified domain name is not specified in the Host field for the stub when it was created. Verify and ensure to add the fully qualified domain name of the server in the Host field when the physical transport for the stub is configured in Rational® Integration Tester.

When HTTP stubs on Rational® Test Automation Server and called via the HTTP proxy, the calls fail with HTTP 404 errors.

You must enable service virtualization through Istio at the time of installation of the server software. See Installation of the server software.

Table 6. Troubleshooting issues: test results and reports

Problem

Description

Solution

You are not able to view the Jaeger traces for the tests you ran.

The cause can be as follows:
  • Might be that you have not installed Jaeger at the time of server installation.
  • The Jaeger trace is not supported for the particular test you ran.
Check for any of the following solutions: